• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Infrared and Visible Image Fusion Based on Res2 Net-Transformer Automatic Encoding and Decoding

    2024-05-25 14:42:46ChunmingWuWukaiLiuandXinMa
    Computers Materials&Continua 2024年4期

    Chunming Wu ,Wukai Liu and Xin Ma

    1Key Laboratory of Modern Power System Simulation and Control&Renewable Energy Technology,School of Electrical Engineering,Northeast Electric Power University,Jilin,132012,China

    2School of Electrical Engineering,Northeast Electric Power University,Jilin,132012,China

    3School of Aeronautical Engineering,Jilin Institute of Chemical Technology,Jilin,132022,China

    ABSTRACT A novel image fusion network framework with an autonomous encoder and decoder is suggested to increase the visual impression of fused images by improving the quality of infrared and visible light picture fusion.The network comprises an encoder module,fusion layer,decoder module,and edge improvement module.The encoder module utilizes an enhanced Inception module for shallow feature extraction,then combines Res2 Net and Transformer to achieve deep-level co-extraction of local and global features from the original picture.An edge enhancement module (EEM) is created to extract significant edge features.A modal maximum difference fusion strategy is introduced to enhance the adaptive representation of information in various regions of the source image,thereby enhancing the contrast of the fused image.The encoder and the EEM module extract features,which are then combined in the fusion layer to create a fused picture using the decoder.Three datasets were chosen to test the algorithm proposed in this paper.The results of the experiments demonstrate that the network effectively preserves background and detail information in both infrared and visible images,yielding superior outcomes in subjective and objective evaluations.

    KEYWORDS Image fusion;Res2 Net-Transformer;infrared image;visible image

    1 Introduction

    Image fusion involves combining two or more pictures using a certain algorithm to create a new image.Utilizing the spatial and temporal correlation as well as information complementarity of many photos enhances the clarity and comprehensiveness of the scene description,making the fused image more appropriate for human visual perception and automated machine detection[1].Infrared sensors detect heat source targets by detecting thermal radiation and may operate in all situations and weather.They have an excellent ability to differentiate between background and target information and demonstrate considerable resistance to interference.Nevertheless,the pictures acquired frequently lack structural texture.Visible sensors can analyze intricate scene and texture characteristics by interpreting light reflection,mirroring the observation pattern of the human eye.The identification effect is moderate and sensitive to light,which might make it challenging to use in low-light conditions.Image fusion technology combines visible and infrared sensors to leverage their strengths and compensate for their weaknesses,resulting in images with enhanced detail and improved target perception.These images can be used to aid other related activities in other areas,such as medical diagnosis,smart driving,and security monitoring[2].

    Current methods for combining infrared and visible pictures may be generally divided into two categories: Conventional fusion techniques and deep learning-driven fusion techniques.Traditional fusion methods include image pyramid transform [3],wavelet transform based [4],contour wavelet[5],multi-scale geometric transform,spatial filtering,etc.[6–8].For example,in literature[9],weighted average and nuclear norm are set as fusion strategies;however,for modal features of infrared images,the target is highlighted by pixel brightness perception,while for visible light,the feature is highlighted by structural texture,and the most perfect feature may not be obtained.Also,in the literature[10],the visible and infrared images are decomposed separately using multiscale transform(MST),and colour fusion is performed manually.Conventional fusion techniques need manual configuration,and the resultant fused picture is dependent on manually established guidelines.Furthermore,conventional approaches have a drawback in feature extraction since they extract a limited number of characteristics.Deep learning has advanced image processing by effectively extracting intricate features from images,addressing the limitations of traditional methods in feature extraction.Consequently,fusion methods based on deep learning are increasingly utilized.

    Deep learning approaches may be categorized into four groups,with the first being a convolutional neural network [11–13],which extends the multi-scale processing of images.Optimization network approaches,like filters or feature extraction networks,convert the spatial domain into the basic domain and the detail domain.The average of the basic domain is combined with high-frequency texture features in the detail domain.The detailed feature map generated from the original pre-trained neural network is integrated.The basic image and detailed image are fused to create the final fused picture.In literature [14],Zhang et al.designed a feature coding block composed of convolutions and proposed a lightweight convolutional neural network containing only nine convolutional neural networks(CNNs)layers for encoding and decoding,which can carry out feature fusion more fully.In the literature[15],Tang et al.introduced a method for fusing infrared and visible light images using Convolutional Neural Networks (CNNs) and saliency detection.They utilized a two-channel CNN to generate the initial fused image.A saliency map is created by comparing the structural similarity between the infrared and visible source pictures with the original fused image.A weighted fusion rule is employed to create the fused picture,maintaining the intensity and texture homogeneity in the saliency zone.The second category is based on generative adversarial networks [16,17].In this type of article,image fusion is considered as a result of adversarial equilibrium,where the generator creates fused images that contain the glowing and texture information of the source image and keeps adding detailed information for generation and adversarial,and finally,the generator produces fused images that are indistinguishable by the discriminator.These end-to-end models avoid the manual design of fusion rules.In literature[18],Liu et al.proposed a comparison of visible infrared image conversion networks and introduced a class-activation graph-based attention module into the generator and discriminator for generating adversarial networks (GANs) to capture richer contextual information in images.For example,in literature[19],Li et al.utilized the coupled generative adversarial network in picture fusion to expedite the network’s convergence.The third category is an automatic encoding and decoding-based image fusion architecture where the network is taught during the training phase to extract features from the source picture.During the testing step,the feature maps are combined individually,and then the integrated picture is rebuilt using a decoder.Deep neural networks(DNNs)are commonly employed to extract features from input images,followed by the application of specific fusion algorithms to combine these characteristics and accomplish the image fusion job.In a study referenced as[20],it was shown that utilizing solely convolutional and pooling layers together with a loss function,without incorporating reference quality measurements,effectively combines the feature.Wang et al.retrieved superficial and profound image features by using two branches of the encoder in literature[21].The fusion layer uses residual blocks to combine two sets of features from the infrared and visible images inside the same branch,and the decoder creates the final fused image.The fourth category is fusion methods based on Transformer,which was first proposed in the literature[22]in 2017 and applied to natural language processing with good results.Since then,Transformer applications have gradually emerged in image processing.For example,in literature [23],a Transformer network was used for automatic detection in image processing and achieved good results,followed by image recovery in literature[24]in 2021.The approach involved utilizing a convolutional layer for extracting shallow features initially,followed by employing the Swin Transformer for deep feature extraction.In 2022,in literature[25],Wang et al.introduced the use of a Transformer in image fusion for infrared and visible light.They developed a fully attentional feature encoding backbone to capture longrange dependencies.This pure Transformer network has superior representational abilities compared to convolutional neural networks and effectively maintains the brightness of infrared images while preserving visible details.

    However,the aforementioned deep learning methods still have certain issues.For instance,in convolutional neural networks,the excessive use of convolutional layers can lead to the vanishing or exploding gradients of the network.While GANs train the generator and discriminator in an adversarial manner,there are potential problems such as training imbalance and mode collapse.Sometimes,the generator is able to generate samples of high quality,but the discriminator can still accurately distinguish them,causing training to become difficult.Standalone encoder-decoder networks or Transformer networks also have their limitations.This research suggests an automatic encoded and decoded infrared-visible image fusion network that combines Res2 Net and Transformer to overcome the limits of individual approaches and improve the quality of fused pictures.This paper’s contributions can be summarized as follows:

    (1) The introduction of the improved Inception module enables the extraction of multi-scale features on the one hand,and on the other hand,it also increases the robustness of the network.

    (2) An Edge Enhancement Module (EEM) is created to extract additional edge characteristics from the original photos.

    (3)A novel encoder consisting of Res2 Net and Transformer is built for simultaneous extraction of local and global information from infrared and visible images.

    (4) A modal maximum difference degree fusion technique is proposed to enhance the adaptive representation of information in various regions of the source picture during fusion,resulting in a fused image that aligns better with human visual perception.

    (5) The proposed method achieved favourable results in subjective evaluations and objective analyses compared to several typical methods across three different datasets.

    The remainder of this article is organized as follows.Section 2 presents the network fusion framework and the framework of each module,Section 3 provides the loss function and fusion technique,Section 4 covers the experimental findings and analysis,and Section 5 concludes the document.

    2 Proposed Methodology

    2.1 Overall Network Structure

    The network has three primary modules: Feature extraction,feature fusion,and feature reconstruction.The feature extraction module consists of three stages: Extracting features from infrared pictures,extracting features from visible light images,and extracting edge features from both types of photos.During each stage,both the infrared (IR) and visible source images are simultaneously processed.Shallow feature extraction is conducted using an enhanced Inception module,passing through each convolutional layer sequentially.Feature enhancement and a Transformer cascade of Res2 Net are utilized for feature extraction to generate the background and detail feature maps of IR and visible light,denoted asBI,DI,BVandDV,respectively.Furthermore,there are edge features EFacquired from the feature extraction module.The feature fusion layer combines the retrieved features and sends them to the three-layer convolution decoder.During the decoding stage,the feature maps from the Inception module and the first convolution are added to the last and penultimate convolution layers to preserve pixel intensity and gradient information from the source image.This process aims to prevent the loss of detailed information after multiple convolutions and enhance convergence speed,resulting in a fused image that combines features from both infrared and visible light images.Fig.1 displays the comprehensive structure of the fusion model introduced in this article.The next sections will provide a detailed explanation of the enhanced Inception module,Res2 Net Residual Swin Transformer (RRST) feature extractor,Transformer module,and edge improvement module integrated into the encoder for designing infrared and visible images.

    Figure 1: Infrared and visible image fusion network

    2.2 Improved Inception Module

    This research utilizes the Inception network to incorporate multi-scale information to enhance the fused picture characteristics and ensure the network’s resilience.The inception network is a significant achievement in the evolution of CNN classifiers.Before Inception,popular CNNs often added additional convolutional layers to increase the depth of the network in order to improve performance.The first iteration of the Inception network,Inception V1 or GoogLeNet,was created to develop a network with a strong local structure by executing numerous convolution and pooling operations simultaneously on the input image and combining all the results into a deep feature map.Utilizing various convolution procedures like 1×1,3×3,or 5×5 together with pooling operations may gather diverse information from the input picture.Running these operations simultaneously and merging the outcomes enhances the image representation.Subsequent to Inception V1,networks like Inception V2,Inception V3,Inception V4,and Inception-ResNet-V2[26–29]have been introduced.

    Fig.2 displays the enhanced Inception module structure utilized in this study,which maintains the benefits of feature maps,including intricate feature patterns at various sizes,employing diverse convolutional kernels,and capturing features at many scales.The network starts with a 1 × 1 convolution,followed by 3×3 and 5×5 convolutions to enhance network depth,enhance network nonlinearity,and decrease network parameters.The model incorporates a bulk normalization (BN)layer preceding each input layer to normalize the inputs to a mean of 0 and a variance of 1.This boosts the network’s complexity and non-linear characteristics while also speeding up the training procedure.

    Figure 2: Inception module

    2.3 Res2 Net

    Res2 Net was proposed by Mingming Cheng’s group at Nankai University in 2019 [30],and the main contribution is to improve the block module in the ResNet model with no increase in computational load and more powerful feature extraction.As shown in Fig.3,in simple terms,Res2 Net divides the input of 3×3 convolutional layers into four parts,and the network is connected internally in a residual style.Before finally going through the 1×1 convolution,after the newly added Squeeze-and-Excitation(SE)module,the expressive ability and performance of the neural network are enhanced by adaptively learning the weights of each channel so that the network can learn and utilize the relationship between the feature channels more efficiently,enabling Res2 Net to better extract multiscale features.

    Figure 3: Res2 Net module

    2.4 RRST Module

    Fig.4 a illustrates the network structure of the Res2 Net Residual Swin Transformer Block.It consists of Res2 Net and a sequence of Swin Transformer layers(STLs)with residual connections after the previous shallow extraction,followed by combining the extracted information into the Residual Swin Transformer Block network.Three Res2 Net modules and three Swin Transformer layers are employed in this study to extract intermediate global features.Res2 Net and Swin Transformer utilize many layers to capture global data effectively,with residual connections aggregating features across various levels.

    Figure 4: Res2 Net residual swin transformer block&swin transformer block

    The Swin Transformer layer,depicted in Fig.4b,as described in the literature[25],first employs an N×N sliding window to divide the input into non-overlappingHW/N2local windows and calculate their local attention.The matrices Q,K and V for the local windowΦz feature are computed using Eq.(1).

    The parametersWQ,WKandWV∈RN2are shared over three linear projection layers for distinct windows,whereas d represents the dimension of(Q,K).The sequence matrices of the self-attention process are described by Eq.(2).

    p represents the learnable parameter utilized for position decoding.The Swin Transformer layer then calculates the usual Multiple Self Attention(MSA)for the shift window.The system consists of a Weighted-Moving Standard Average(W-MSA)and a multilayer perceptron(MLP).A LayerNorm layer is placed before W-MSA and MLP,with a residual connection utilized for the module.

    2.5 Edge Enhancement Module

    In order to retain more edge features in the fusion stage,an edge feature extraction module is designed,as shown in Fig.5a.First,the Canny edge extraction operator is used to obtain preliminary edge features,and then two groups of convolution layers,both of which have convolution kerns of 3×3,and the Relu activation function is used in CONV-Block 1.In CONV-Block 2,the Tanh activation function is used and finally passes through the convolutional block attention module(CBAM)module,as shown in Fig.5b.The edge feature extraction of the source image is realized.

    Figure 5: Edge enhancement module

    EFis the feature of the edge extraction module,C is the edge extraction operator,infrared and visible images are through the edge extraction operator to get the featuref1,and finally through the CBAM module to get the final edge feature.

    3 Loss Functions and Fusion Strategies

    3.1 Loss Function for Image Decomposition(Coding)

    The background feature map is to get the same feature points of IR and visible light,and the detail feature map is to get the respective features of IR and visible light,so it should make the difference between the obtained same features a little less and make the difference between different features a little more,respectively,the loss function for image decomposition is defined as follows:

    BVandDVrepresent the background and detail feature maps of the visible image V,whereasBIandDIrepresent the background and detail feature maps of the infrared image I.The functionΦ(·)utilizes a tangent function with a restricted value range of(-1,1).

    3.2 Loss Function for Image Reconstruction

    The reconstruction loss function is defined by the following equation for image reconstruction,where the objective is to accurately maintain the input image’s pixel intensity and fine-grained texture information:

    Iandrepresent the input and reconstructed infrared pictures,whereasvand ?vrepresent the visible images.?symbolizes the gradient operator.

    The input and reconstructed pictures are designated by the lettersxand ?x,respectively,andλrepresents the hyperparameter.The original and reconstructed images’pixel intensity consistency is measured byL2-norm,while the difference in the images’brightness,contrast,and structural composition is determined byLSSIM.In this instance,LSSIMis calculated as

    Combining Eqs.(6)as well as(7),the total lossLtotalcan be expressed as

    The hyperparameters in the loss function are set as follows:α1=0.5,α2=3.5,α3=3.5,α4=10,λ=5(Subsequent experiments will justify the values taken).

    3.3 Integration Strategy

    The preceding section outlines the network structure and loss function.Following the training process,an encoder and a decoder will be generated.The objective during the fusion step is to combine the infrared and visible pictures.A fusion layer is added during the fusion stage to combine the background and detail feature maps together with the edge features.The fusion of the background features is calculated as Eq.(11)since the difference between the two is not much for the background information,but the difference between the detail information features is large.

    However,the detailed feature map is to show the difference between the two.The infrared detail feature map and the visible detail feature map are subtracted to calculate the difference between the two,and then the difference feature map and the maximum difference degree of the difference feature map are divided to obtain the relative difference weights and,as shown in Eqs.(12)and(13).

    The max in Eqs.(12) and (13) performs a global depth maximum pooling operation on the difference feature maps to obtain the maximum degree of difference between the two images.

    In order to assign weights to the pixels in the infrared and visible feature maps according to their importance,the relative difference weights obtained above are added to their absolute values and averaged as the final adaptive weighted maps λ1andλ2to guide the fusion process,shown by Eqs.(14)and(15),where λ1andλ2have the same dimensions as the initial feature maps.

    So he said to his love: Alas1! I must go off and leave you, but take this ring and wear it as a remembrance of me, and when I am King I will return and fetch you home

    This adaptive weighting map and the initial feature map are then multiplied to generate the prefused feature maps F1and F2,as shown by Eqs.(16)and(17).

    Finally,the two pre-fusion feature maps are averaged to obtain the final detailed fusion feature mapDF,as shown in Eq.(18).

    Finally,BFandDF,as well as the featureEFobtained through the edge module,carry out the final feature fusion.

    3.4 Complexity Analysis

    Suppose in the training phase,the number of images in the dataset in the training phase is M,the number of cycles of training is E,the number of images contained in one iteration,i.e.,the batch size,is N,the number of iterations (which denotes the number of iterations contained in one cycle) is I,and I=M/N.So,the time complexity obtained in the data extraction phase is O(E)?O(N)?O(I)=O(E ?N ?I).

    The article only includes the essential phases of the training cycle,as the primary time-consuming components are the Transformer and Res2 Net modules.The time complexity of the Transformer model is calculated as O(dim ?H ?W+d ?nW+H ?W ?dim ?dim ?mlp_ratio),where dim represents the input features’dimension,H is the input image height,W is the input image width,d is the input feature dimension within each window,and nW is the total number of windows.The time complexity of the 1x1 convolutional layer in Res2 Net is O(c_in ?c_out ?h ?w),whereas the time complexity of the BN layer and activation function may be ignored to get the total time complexity of Res2 Net.Res2 Net’s time complexity is O(c_in ?c_out ?h ?w ?scales),c_in and c_out denote the input and output channels of the convolutional layer,h represents the height of the input feature map,w represents the width of the input feature map,and scales indicate the number of feature groups in the residual block.

    In the training phase,since the rest of the steps are executed sequentially,the time complexity is all O(1),so the total time complexity is O(ENI)?[O(dim ?H ?W+d ?nW+H ?W ?dim ?dim ?mlp_ratio)+O(c_in ?c_out ?h ?w ?scales)].

    In the testing phase,this paper uses only one kind of dataset to test the time complexity and uses the already trained model.The time complexity depends on the number of times of forward propagation,so the time complexity of the testing phase is O(M ?I),where M is the number of samples tested and I is the prediction time of each sample.

    4 Experiment

    4.1 Experimental Setup

    This study uses six objective parameters to thoroughly assess the quality of the merged picture.Information entropy (EN) quantifies the information included in the combined features of the grayscale distribution inside the image.Mutual information (MI) quantifies the level of mutual interdependence between two variables.MI may assess the level of information overlap between two input pictures in image fusion.The average gradient(AG)is a measure of picture sharpness in fused images.A higher AG value implies greater image sharpness and overall fusion quality.Standard deviation(SD)is a quantitative measure that assesses the amount of information included in a picture.A higher standard deviation number signifies a wider spread of grayscale values and a greater quantity of information conveyed by the image,leading to improved fusion image quality.Visual Information Fidelity(VIF)is a measure used to assess how well visual information is maintained.Peak signal-tonoise ratio(PSNR)assesses picture distortion by analyzing the discrepancies between the original and reconstructed images.Higher PSNR values correspond to superior image quality.Six generally used parameters are utilized to objectively assess the fusion results of each approach.

    4.2 Experimental Results and Analysis

    4.2.1 Training Effect

    In the training phase,Adam optimizes the network for 120 epochs with a batch size of 24 during the training phase.In this study,the learning rate is set at 10-3and is lowered by a factor of 10 every 40 epochs.

    In order to verify the optimal values of each parameter value,the values of each parameter were adjusted to four sets of values(a),(b),(c),and(d),respectively,in the training phase and the optimal values were obtained according to the training results.The values of each parameter were set among them,as shown in Table 1.

    Table 1: Individual values of the parameters

    The obtained training effects are shown in Figs.6a–6d,define the horizontal and vertical axes as the x-axis and y-axis,respectively.The results show that all the loss curves are almost very flat after 120 rounds of training.Four training images were obtained for the four sets of data in the setup,and in Figs.6a and 6c,although the overall loss function and the recombined loss function curves reached convergence after about 20 rounds,the background loss function did not converge until about 40 rounds,and the convergence performance was slow and ineffective.Fig.6b shows that while the overall loss function and background loss function converge rapidly,the detail loss function decreases compared to group (d).A higher value in the detail loss function indicates better model extraction performance,making the image (d) more effective.Therefore,the data from group (d) is selected as the parameter value.

    4.2.2 Performance Evaluation

    This paper utilizes a dichotomous model support vector machine(SVM)to assess the benefits of the training model,as inspired by the literature [31].Fused images from each algorithm are used as inputs,and the variance of the area under the curve is calculated along with the 95%confidence interval using a normal distribution function.The 95%confidence interval,a statistical concept,represents the range of uncertainty in estimating parameter results in statistical inference,indicating the reliability of the inferred outcomes.The receiver operating characteristic (ROC) curve is obtained,which in turn gives the area under the curve (AUC) value [32].An AUC value of 0.9065 can be obtained for the model of this paper,0.8641 for U2Fusion,0.7301 for SwinFusion,and 0.7254 for NestFuse.Since there are more methods to compare,only three algorithms were selected for comparison,as shown in Fig.7.From the AUC results,it can be seen that the model proposed in this paper performs better.

    Figure 6: (Continued)

    Figure 7: ROC curve

    The performance comparison of various machine learning algorithms is shown in Table 2.

    Table 2: Comparison of different algorithms in SVM

    In order to verify the effectiveness of the method,Nestfuse [33],Seafusion [34],U2Fusion [35],IGNet [36],SwinFusion [37],DATFuse [38],PAIFusion [39],MetaFusion [40],and YDTR [41]were selected nine typical fusion algorithms are compared.The method’s performance is assessed subjectively and objectively.

    Fig.8 displays the fusion outcomes of combining a collection of infrared and visible photos from the TNO dataset using the nine approaches mentioned above.The infrared target perceives the brightness of (c),(f),(h),and (k) in Fig.8 as dim and indistinct,particularly inside the area shown in red.Only the basic shape of the window is seen,but individual details are unclear.In Figs.8d and 8g,although the brightness of the target is high,the colour of the figure is slightly dim in terms of the richness of texture details,and the key features are not highlighted.As can be seen from the green border marked in the figure,in images(e)and(i),although the brightness of the target is high,there are some distortions,especially in the image(i),where there are large black shadows.The original features cannot be accurately displayed.Compared with the target of the fusion image mentioned above,the fusion image target obtained by the model in this paper is bright and clear,which proves that the model can extract more feature information from the source image.

    Figure 8: Experimental results on the TNO dataset

    Fig.9 displays the fusion outcomes of an infrared and visible picture pair using several techniques in the NIR dataset.The target brightness of images (c),(f),(h),and (k) in Fig.9 is notably low,particularly in picture(h).The picture is too fuzzy,almost distorted,and cannot accurately reflect the texture and detail characteristics of the source image.In images(d),(g)and(j),the brightness effect looks good,but the texture of visible images cannot be well displayed.For example,in the marked red border,the shadow part of the tree is lighter,and the visual effect is unsatisfactory.In images(e)and(i),although the brightness is higher,the image is also clearer,but the tree and the shadow part(green border and red border)are a little too black,and it looks even darker than the colour of the source image,which is also relatively distorted.It can clearly display high brightness and does not excessively make a certain part of the image darker than the image(i).This method of fusion preserves the target features and background characteristics to the fullest extent in the resulting image.Hence,the fusion impact of this approach surpasses that of other comparison methods.

    Fig.10 displays the fusion outcomes of an infrared and visible picture pair from the FLIR dataset utilizing several techniques.In images(c),(e),(f),(h)and(k),some details of the target are lost,and the target point is almost invisible,especially in the image(e),although the overall image looks bright,the manhole cover in the figure is completely invisible,and the distortion is too serious,which can be seen by the marked red border.Although the target point can be seen in image(i),the overall image is still too black,except for the points marked.Almost all other places are black.In images (d),(g)and(j),although the overall image looks bright,the license plate in the figure is partially lost,and the texture details are not completely presented.In summary,it can be seen that the image fused by the method of this paper has a bright target,rich background and better quality.

    Figure 9: Experimental results of NIR dataset

    Figure 10: Experimental results of FLIR dataset

    The data metrics acquired by each algorithm in different datasets are presented in Tables 3–5 to confirm the benefits of the strategy suggested in this work.

    The approach presented in this study demonstrates the most effective results based on the objective assessment metrics of fusion pictures in Table 3 and Fig.8.The approach achieves the four highest fusion indicators of MI,AG,VIF,and PSNR among the six data indicators.In the AG index,the value is 7.2548,and the AG value of YDTR is 2.7249,which is 4.5299 greater than the AG value.This indicates that the picture has more edge features and demonstrates the success of the edge enhancement module.Furthermore,the MI and VIF values are greater compared to other methods,aligning with subjective perception.The fusion picture target seems brilliant,with rich local information and a pleasant visual impression.The study demonstrates that the proposed technique may extract additional information and provide a superior fusion effect in the TNO dataset.

    Table 3: Comparison of the performance of the different methods in Fig.8

    Table 4 documents the objective assessment metrics of the fusion pictures produced by the various approaches depicted in Fig.9.When using evaluation measures EN,the value may not be much greater compared to other techniques,but it can still indicate the level of similarity or information overlap between the original picture and the fused image.The MI value is 0.9708,more than the PAIFusion value of 3.897,indicating a higher level of feature information from the original picture.It indicates that the NIR data set yields more information and a higher correlation between the fused image and the source image.This approach effectively integrates picture information,texture,contrast,and visual effects,aligning with subjective judgment better than previous algorithms.

    Table 4: Comparison of the performance of the different methods in Fig.9

    The benefits of this paper’s approach are less noticeable in Table 5 when compared to the first two datasets.However,it does showcase some advantages in certain aspects.For instance,the EN value of this paper is 7.6973,whereas DATFuse is 6.6901,resulting in a difference of 1.0072.The fused image contains more information compared to the DATFuse algorithm.Additionally,the PSNR achieved the highest value,indicating that the fused image has more correlation or similarity with the source image in the FLIR dataset,resulting in better quality.This demonstrates the advantages of the method presented in the paper.

    Table 5: Performance comparison of different methods in Fig.10

    The data above shows that the fusion method suggested in this study completely integrates picture information compared to the other approaches.The fusion image comprises the most complimentary information from the source image and effectively preserves the edge texture details,demonstrating the benefits of the suggested approach.

    5 Conclusion

    This article proposes a novel infrared and visible light image fusion autoencoder-decoder based on the Res2 Net-Transformer structure.Firstly,most existing methods in image fusion either solely utilize CNN’s convolution layers or apply emerging transformer models for feature extraction.In contrast,the proposed algorithm combines the Res2 Net model for deep feature extraction with the Transformer model to incorporate global feature extraction,thereby improving the fusion quality.An edge extraction module is created to enhance the capture of edge characteristics in pictures,facilitating the extraction of more authentic edge features.A novel fusion approach has been developed based on optimizing the disparity between the infrared and visible light feature maps.The procedure assigns weights to pixels in feature maps based on their importance to guarantee the fused image retains crucial properties from the original photos adaptively.The autoencoder fusion architecture excels at merging infrared and visible light images,as demonstrated by subjective judgments and objective evaluation criteria.Pixel-level fusion aims to achieve high-quality fused pictures and support tasks like item identification and recognition.Hence,the subsequent action may involve creating a multitasking neural network capable of efficiently executing tasks like object identification while producing superior fused pictures.

    Acknowledgement:The authors would like to express their gratitude to the members of the research group for their support.

    Funding Statement:The authors received no specific funding for this study.

    Author Contributions:Study conception and design: Chunming Wu;data collection,analysis,and interpretation of results:Wukai Liu;draft manuscript preparation:Xin Ma.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The experimental data supporting the study’s conclusions may be obtained from the corresponding author upon request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    免费在线观看完整版高清| 亚洲国产中文字幕在线视频| 人妻丰满熟妇av一区二区三区| www日本在线高清视频| 国产一区二区三区视频了| 日韩免费av在线播放| 国产欧美日韩一区二区三区在线| 不卡一级毛片| 精品一区二区三区视频在线观看免费| 亚洲国产欧美日韩在线播放| 国产精品,欧美在线| 51午夜福利影视在线观看| 国内精品久久久久久久电影| av电影中文网址| 国产成人欧美| 91麻豆精品激情在线观看国产| 99国产精品一区二区蜜桃av| 午夜久久久在线观看| 欧美乱色亚洲激情| 国产精品久久久av美女十八| av在线天堂中文字幕| 三级毛片av免费| 中文字幕久久专区| 好男人电影高清在线观看| 91精品三级在线观看| 桃红色精品国产亚洲av| 我的亚洲天堂| 国产精品综合久久久久久久免费 | 女人被狂操c到高潮| 日日摸夜夜添夜夜添小说| 夜夜看夜夜爽夜夜摸| av视频免费观看在线观看| 日韩欧美免费精品| 少妇的丰满在线观看| av电影中文网址| 亚洲精品一区av在线观看| 久久精品人人爽人人爽视色| 国产97色在线日韩免费| 久久精品亚洲熟妇少妇任你| 男女做爰动态图高潮gif福利片 | 男男h啪啪无遮挡| 女人被躁到高潮嗷嗷叫费观| 色综合站精品国产| 国产亚洲欧美98| 亚洲中文日韩欧美视频| 最好的美女福利视频网| 日韩欧美免费精品| 多毛熟女@视频| 在线十欧美十亚洲十日本专区| 神马国产精品三级电影在线观看 | 国产精品一区二区三区四区久久 | 操美女的视频在线观看| 国产人伦9x9x在线观看| 天天一区二区日本电影三级 | 深夜精品福利| 91av网站免费观看| 亚洲人成电影免费在线| 人妻久久中文字幕网| 日韩精品中文字幕看吧| 一a级毛片在线观看| 欧美成人性av电影在线观看| 国产精品久久电影中文字幕| 超碰成人久久| 亚洲成人精品中文字幕电影| 亚洲七黄色美女视频| 大码成人一级视频| 欧美不卡视频在线免费观看 | 日本五十路高清| 国产一区二区激情短视频| 99精品久久久久人妻精品| 久久青草综合色| 老汉色∧v一级毛片| 亚洲精品美女久久久久99蜜臀| 禁无遮挡网站| 麻豆一二三区av精品| 亚洲专区中文字幕在线| 欧美激情高清一区二区三区| 日韩欧美在线二视频| 久热爱精品视频在线9| 国产精品久久久人人做人人爽| 最近最新免费中文字幕在线| 久久影院123| 一边摸一边做爽爽视频免费| 亚洲精品粉嫩美女一区| 女人被狂操c到高潮| 在线十欧美十亚洲十日本专区| 精品国产一区二区三区四区第35| 高清毛片免费观看视频网站| 国产1区2区3区精品| 日本免费一区二区三区高清不卡 | 又黄又爽又免费观看的视频| 欧美黄色片欧美黄色片| 国产成人一区二区三区免费视频网站| 美国免费a级毛片| 亚洲国产毛片av蜜桃av| 最近最新中文字幕大全免费视频| 久久影院123| 老汉色∧v一级毛片| 国产精品日韩av在线免费观看 | 日本 av在线| 一级片免费观看大全| www.熟女人妻精品国产| 午夜精品国产一区二区电影| 俄罗斯特黄特色一大片| 国内久久婷婷六月综合欲色啪| 激情在线观看视频在线高清| 亚洲激情在线av| 亚洲五月婷婷丁香| 国产亚洲精品综合一区在线观看 | 少妇粗大呻吟视频| 久久中文字幕一级| 亚洲国产精品久久男人天堂| 亚洲成人久久性| 国产精品日韩av在线免费观看 | 国产精品久久久人人做人人爽| 国产成人影院久久av| 精品久久久久久久久久免费视频| 欧美+亚洲+日韩+国产| 久久久精品欧美日韩精品| 欧美丝袜亚洲另类 | 久久久久九九精品影院| 看黄色毛片网站| av中文乱码字幕在线| 久久人妻熟女aⅴ| www.精华液| 久久香蕉精品热| 日本精品一区二区三区蜜桃| 亚洲欧美激情综合另类| 久久精品亚洲熟妇少妇任你| 两个人免费观看高清视频| 日韩高清综合在线| 国产精品野战在线观看| av欧美777| 午夜免费激情av| 精品国内亚洲2022精品成人| 精品国产一区二区三区四区第35| 午夜精品久久久久久毛片777| 黑人操中国人逼视频| av天堂在线播放| 国产91精品成人一区二区三区| 淫秽高清视频在线观看| 69av精品久久久久久| 男人的好看免费观看在线视频 | 欧美日韩瑟瑟在线播放| 国产精品一区二区精品视频观看| 国产一区二区三区视频了| 天天躁狠狠躁夜夜躁狠狠躁| 美女高潮喷水抽搐中文字幕| 狠狠狠狠99中文字幕| 国产色视频综合| 国产亚洲欧美精品永久| 脱女人内裤的视频| 午夜激情av网站| 在线视频色国产色| videosex国产| 后天国语完整版免费观看| svipshipincom国产片| 一本大道久久a久久精品| 国产精品影院久久| 黄色女人牲交| 亚洲,欧美精品.| 亚洲性夜色夜夜综合| 日本 欧美在线| 国产极品粉嫩免费观看在线| 久久影院123| 超碰成人久久| 男人舔女人的私密视频| 一进一出好大好爽视频| 日本a在线网址| e午夜精品久久久久久久| 人人澡人人妻人| 亚洲五月色婷婷综合| 亚洲欧美日韩无卡精品| 欧美黄色片欧美黄色片| 国产高清有码在线观看视频 | 久久人妻熟女aⅴ| 女生性感内裤真人,穿戴方法视频| aaaaa片日本免费| 国产亚洲精品一区二区www| 纯流量卡能插随身wifi吗| 中文字幕人成人乱码亚洲影| 男女床上黄色一级片免费看| 亚洲无线在线观看| 午夜福利成人在线免费观看| 不卡一级毛片| 亚洲欧美一区二区三区黑人| 亚洲精品av麻豆狂野| 成人三级黄色视频| 嫁个100分男人电影在线观看| 国产成人精品久久二区二区免费| 午夜免费观看网址| 成年女人毛片免费观看观看9| 在线观看免费视频日本深夜| 18禁美女被吸乳视频| 最近最新中文字幕大全电影3 | 国产又爽黄色视频| 中文字幕人妻熟女乱码| 9热在线视频观看99| 天堂√8在线中文| 久久草成人影院| 激情视频va一区二区三区| 天天添夜夜摸| 精品久久久久久久久久免费视频| 美女免费视频网站| 欧美人与性动交α欧美精品济南到| 不卡av一区二区三区| 亚洲男人天堂网一区| av片东京热男人的天堂| 久久国产精品影院| 国产伦人伦偷精品视频| 窝窝影院91人妻| 欧美在线黄色| 美女 人体艺术 gogo| 日韩精品中文字幕看吧| 亚洲性夜色夜夜综合| 日韩欧美国产在线观看| 欧美在线黄色| av福利片在线| 亚洲熟女毛片儿| 精品国产乱码久久久久久男人| 九色国产91popny在线| 亚洲美女黄片视频| 亚洲人成电影观看| 成人国产综合亚洲| av在线天堂中文字幕| 久久人妻av系列| 无遮挡黄片免费观看| 禁无遮挡网站| 国产精品一区二区精品视频观看| 国产伦人伦偷精品视频| 亚洲av片天天在线观看| www.999成人在线观看| 国产成人欧美在线观看| 亚洲国产欧美日韩在线播放| 日韩av在线大香蕉| 在线国产一区二区在线| 亚洲精品美女久久av网站| 老司机午夜福利在线观看视频| 欧美日韩黄片免| 国产高清有码在线观看视频 | 99精品久久久久人妻精品| 日韩精品中文字幕看吧| 国内毛片毛片毛片毛片毛片| 可以在线观看毛片的网站| 一区福利在线观看| 非洲黑人性xxxx精品又粗又长| 亚洲欧美日韩另类电影网站| 色播亚洲综合网| 亚洲av片天天在线观看| 在线免费观看的www视频| 久久久国产成人精品二区| 每晚都被弄得嗷嗷叫到高潮| 一夜夜www| 精品久久蜜臀av无| 少妇粗大呻吟视频| 国产色视频综合| 日韩精品中文字幕看吧| 日本免费a在线| 国产99久久九九免费精品| 精品人妻在线不人妻| 18禁黄网站禁片午夜丰满| 国产视频一区二区在线看| 性色av乱码一区二区三区2| 日日夜夜操网爽| 少妇被粗大的猛进出69影院| 精品久久蜜臀av无| 欧美日本亚洲视频在线播放| 人人妻人人澡人人看| 中文亚洲av片在线观看爽| 精品免费久久久久久久清纯| 亚洲一区二区三区不卡视频| 久久久精品欧美日韩精品| 我的亚洲天堂| av视频免费观看在线观看| 欧美黄色淫秽网站| 女人高潮潮喷娇喘18禁视频| 久久这里只有精品19| 久久精品91蜜桃| 欧美成人性av电影在线观看| 亚洲自偷自拍图片 自拍| 久久久久久久午夜电影| 亚洲 欧美一区二区三区| 色综合亚洲欧美另类图片| 精品久久蜜臀av无| 日韩成人在线观看一区二区三区| e午夜精品久久久久久久| 少妇粗大呻吟视频| 好男人在线观看高清免费视频 | 欧美午夜高清在线| 女人被狂操c到高潮| 侵犯人妻中文字幕一二三四区| 国产一区在线观看成人免费| 日韩av在线大香蕉| 精品久久蜜臀av无| 精品高清国产在线一区| 激情视频va一区二区三区| 一进一出好大好爽视频| 国产精品一区二区精品视频观看| www.自偷自拍.com| 日本a在线网址| 91大片在线观看| 妹子高潮喷水视频| 亚洲人成77777在线视频| 亚洲男人的天堂狠狠| 妹子高潮喷水视频| 男女之事视频高清在线观看| 欧洲精品卡2卡3卡4卡5卡区| 999精品在线视频| 国产精品av久久久久免费| 国产片内射在线| 亚洲精品在线观看二区| 欧美成人一区二区免费高清观看 | 国产精品美女特级片免费视频播放器 | 久久性视频一级片| 午夜免费鲁丝| 亚洲三区欧美一区| 女性生殖器流出的白浆| 精品午夜福利视频在线观看一区| 午夜福利成人在线免费观看| 国产成人一区二区三区免费视频网站| 正在播放国产对白刺激| 国产视频一区二区在线看| 欧美激情 高清一区二区三区| 成人国产一区最新在线观看| 亚洲视频免费观看视频| 黄色视频,在线免费观看| 国产极品粉嫩免费观看在线| 国产区一区二久久| 久久精品亚洲精品国产色婷小说| 亚洲第一av免费看| 免费看十八禁软件| 日韩大尺度精品在线看网址 | 亚洲性夜色夜夜综合| 欧美中文综合在线视频| 亚洲欧美日韩另类电影网站| 亚洲中文av在线| 精品欧美一区二区三区在线| 天天一区二区日本电影三级 | 天堂影院成人在线观看| 精品人妻1区二区| 欧美日韩福利视频一区二区| 欧美一区二区精品小视频在线| 大香蕉久久成人网| 国产99白浆流出| 桃红色精品国产亚洲av| 妹子高潮喷水视频| 久9热在线精品视频| 一边摸一边做爽爽视频免费| 午夜福利在线观看吧| 成年版毛片免费区| 1024视频免费在线观看| 国产视频一区二区在线看| 1024视频免费在线观看| 亚洲熟女毛片儿| 一边摸一边抽搐一进一小说| 成人特级黄色片久久久久久久| 国产精品 国内视频| 99国产精品99久久久久| 日韩视频一区二区在线观看| 韩国av一区二区三区四区| 91成年电影在线观看| 真人做人爱边吃奶动态| 久久久久久大精品| 搡老熟女国产l中国老女人| 嫩草影院精品99| 精品高清国产在线一区| av在线天堂中文字幕| 咕卡用的链子| 丝袜美腿诱惑在线| 少妇熟女aⅴ在线视频| 狠狠狠狠99中文字幕| 熟妇人妻久久中文字幕3abv| 久久精品国产亚洲av高清一级| 免费在线观看视频国产中文字幕亚洲| 俄罗斯特黄特色一大片| 欧美性长视频在线观看| 国产成人精品久久二区二区91| 欧美日韩瑟瑟在线播放| 日本 欧美在线| 一级a爱视频在线免费观看| 久久天躁狠狠躁夜夜2o2o| 亚洲精品一卡2卡三卡4卡5卡| 一个人观看的视频www高清免费观看 | 97碰自拍视频| 性少妇av在线| 色在线成人网| 精品久久蜜臀av无| 涩涩av久久男人的天堂| 桃色一区二区三区在线观看| 午夜影院日韩av| 亚洲成国产人片在线观看| 欧美av亚洲av综合av国产av| 99久久精品国产亚洲精品| 日本五十路高清| 久久久久久亚洲精品国产蜜桃av| 搡老岳熟女国产| 久久久久国产一级毛片高清牌| 亚洲美女黄片视频| 成人国语在线视频| 亚洲精品久久成人aⅴ小说| 久久伊人香网站| 一进一出好大好爽视频| 精品国产乱子伦一区二区三区| 99国产精品免费福利视频| 91老司机精品| 国产三级黄色录像| 在线观看舔阴道视频| 香蕉丝袜av| 久久精品亚洲熟妇少妇任你| 免费搜索国产男女视频| 黑人巨大精品欧美一区二区mp4| 久久精品成人免费网站| 亚洲欧美激情在线| 午夜免费激情av| 欧美亚洲日本最大视频资源| 女警被强在线播放| 亚洲欧美日韩无卡精品| svipshipincom国产片| 1024香蕉在线观看| 可以在线观看毛片的网站| 国产亚洲欧美在线一区二区| 91在线观看av| 69av精品久久久久久| 手机成人av网站| 午夜福利在线观看吧| 国产亚洲精品久久久久久毛片| 欧美成狂野欧美在线观看| 亚洲熟女毛片儿| cao死你这个sao货| 亚洲中文字幕日韩| 国产极品粉嫩免费观看在线| 亚洲成人精品中文字幕电影| 熟女少妇亚洲综合色aaa.| 久热这里只有精品99| 亚洲国产中文字幕在线视频| 亚洲男人天堂网一区| 成人欧美大片| 国产精品九九99| 国产亚洲精品av在线| 少妇裸体淫交视频免费看高清 | 91麻豆av在线| 一进一出好大好爽视频| 可以在线观看毛片的网站| 一个人免费在线观看的高清视频| 不卡av一区二区三区| a级毛片在线看网站| 亚洲 欧美 日韩 在线 免费| 日韩有码中文字幕| 国产成人系列免费观看| 日韩欧美国产在线观看| 亚洲中文av在线| 久久精品亚洲精品国产色婷小说| av超薄肉色丝袜交足视频| 国产午夜福利久久久久久| 黄色片一级片一级黄色片| 久久人人精品亚洲av| 色综合欧美亚洲国产小说| 精品国产美女av久久久久小说| 夜夜躁狠狠躁天天躁| 日韩欧美三级三区| 夜夜躁狠狠躁天天躁| 视频区欧美日本亚洲| 欧美乱色亚洲激情| 久久精品成人免费网站| 日韩免费av在线播放| 欧美丝袜亚洲另类 | 亚洲精品国产色婷婷电影| 国产欧美日韩精品亚洲av| 色播在线永久视频| av网站免费在线观看视频| 动漫黄色视频在线观看| 非洲黑人性xxxx精品又粗又长| 国产精品久久久av美女十八| 国产aⅴ精品一区二区三区波| 亚洲熟妇中文字幕五十中出| 99久久99久久久精品蜜桃| 亚洲一码二码三码区别大吗| 精品国产超薄肉色丝袜足j| 美女 人体艺术 gogo| 啦啦啦免费观看视频1| 级片在线观看| 久久久久久国产a免费观看| 国产成+人综合+亚洲专区| 电影成人av| 99久久久亚洲精品蜜臀av| 日本撒尿小便嘘嘘汇集6| 一区福利在线观看| 国产99久久九九免费精品| 欧美日韩亚洲综合一区二区三区_| 日韩精品中文字幕看吧| 后天国语完整版免费观看| 男女午夜视频在线观看| 亚洲精品国产色婷婷电影| 国产三级黄色录像| 亚洲一区中文字幕在线| 欧美乱妇无乱码| 免费看a级黄色片| 国产精品 欧美亚洲| 亚洲熟妇熟女久久| 久久精品国产99精品国产亚洲性色 | 丝袜美腿诱惑在线| 精品少妇一区二区三区视频日本电影| 黄色片一级片一级黄色片| 午夜福利欧美成人| 午夜福利成人在线免费观看| 美女大奶头视频| 中出人妻视频一区二区| 波多野结衣av一区二区av| 成人国语在线视频| 亚洲国产精品999在线| 精品久久蜜臀av无| 午夜精品国产一区二区电影| 这个男人来自地球电影免费观看| 少妇 在线观看| 日本 欧美在线| 黄色成人免费大全| 国产精品 国内视频| 亚洲精品一卡2卡三卡4卡5卡| 啪啪无遮挡十八禁网站| 日韩国内少妇激情av| 制服丝袜大香蕉在线| 国产亚洲精品av在线| 女性生殖器流出的白浆| 久久 成人 亚洲| 亚洲av电影在线进入| 国产精品98久久久久久宅男小说| 操出白浆在线播放| 在线观看免费视频网站a站| 亚洲精品av麻豆狂野| 99国产精品一区二区蜜桃av| 99久久久亚洲精品蜜臀av| 欧美一级毛片孕妇| 欧美日本中文国产一区发布| 99国产精品一区二区三区| 国产av又大| 巨乳人妻的诱惑在线观看| 91老司机精品| 亚洲精品中文字幕一二三四区| 欧美日韩黄片免| 精品久久蜜臀av无| 欧美日本中文国产一区发布| 亚洲午夜精品一区,二区,三区| 久9热在线精品视频| 欧美在线黄色| 亚洲精品av麻豆狂野| 一个人观看的视频www高清免费观看 | 国产91精品成人一区二区三区| 亚洲午夜理论影院| 嫁个100分男人电影在线观看| 涩涩av久久男人的天堂| 国产在线精品亚洲第一网站| 精品免费久久久久久久清纯| 亚洲成人久久性| 亚洲avbb在线观看| 亚洲无线在线观看| 在线观看免费视频日本深夜| 国产亚洲欧美98| 男女床上黄色一级片免费看| 色婷婷久久久亚洲欧美| 免费在线观看黄色视频的| 久热这里只有精品99| 亚洲av成人一区二区三| 午夜免费激情av| 欧美av亚洲av综合av国产av| 看免费av毛片| 一个人观看的视频www高清免费观看 | 国产三级黄色录像| 一级a爱视频在线免费观看| 老司机午夜十八禁免费视频| 国产精品久久久久久精品电影 | 日韩成人在线观看一区二区三区| 亚洲欧美日韩高清在线视频| 亚洲一码二码三码区别大吗| 男人舔女人的私密视频| 正在播放国产对白刺激| 少妇裸体淫交视频免费看高清 | 丁香六月欧美| 麻豆av在线久日| 午夜影院日韩av| 黑人欧美特级aaaaaa片| 久久久国产精品麻豆| 不卡av一区二区三区| 色精品久久人妻99蜜桃| 亚洲国产欧美日韩在线播放| 夜夜夜夜夜久久久久| 国产不卡一卡二| 久久国产精品人妻蜜桃| 久久久国产精品麻豆| 欧美国产精品va在线观看不卡| 亚洲精华国产精华精| 免费在线观看影片大全网站| 日本撒尿小便嘘嘘汇集6| 男人舔女人下体高潮全视频| 国产日韩一区二区三区精品不卡| 国产精品日韩av在线免费观看 | av视频在线观看入口| 国产视频一区二区在线看| 国产欧美日韩综合在线一区二区| 波多野结衣av一区二区av| 正在播放国产对白刺激| 国产精品九九99| 亚洲最大成人中文| 91国产中文字幕| 99国产极品粉嫩在线观看| 一卡2卡三卡四卡精品乱码亚洲| 岛国视频午夜一区免费看| 色播在线永久视频| 免费搜索国产男女视频| 亚洲国产欧美日韩在线播放| 国产成人欧美在线观看| 欧美老熟妇乱子伦牲交| 精品国产国语对白av| 久久天躁狠狠躁夜夜2o2o| 18禁观看日本| 精品久久久久久成人av|