• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    MIDNet:Deblurring Network for Material Microstructure Images

    2024-05-25 14:42:00JiaxiangWangZhengyiLiPengShiHongyingYuandDongbaiSun
    Computers Materials&Continua 2024年4期

    Jiaxiang Wang ,Zhengyi Li ,Peng Shi ,Hongying Yu and Dongbai Sun,3,?

    1National Center for Materials Service Safety,University of Science and Technology Beijing,Beijing,100083,China

    2School of Materials,Sun Yat-Sen University,Shenzhen,518107,China

    3School of Materials Science and Engineering,Southern Marine Science and Engineering Guangdong Laboratory(Zhuhai),Sun Yat-Sen University,Guangzhou,510006,China

    ABSTRACT Scanning electron microscopy(SEM)is a crucial tool in the field of materials science,providing valuable insights into the microstructural characteristics of materials.Unfortunately,SEM images often suffer from blurriness caused by improper hardware calibration or imaging automation errors,which present challenges in analyzing and interpreting material characteristics.Consequently,rectifying the blurring of these images assumes paramount significance to enable subsequent analysis.To address this issue,we introduce a Material Images Deblurring Network (MIDNet) built upon the foundation of the Nonlinear Activation Free Network (NAFNet).MIDNet is meticulously tailored to address the blurring in images capturing the microstructure of materials.The key contributions include enhancing the NAFNet architecture for better feature extraction and representation,integrating a novel soft attention mechanism to uncover important correlations between encoder and decoder,and introducing new multi-loss functions to improve training effectiveness and overall model performance.We conduct a comprehensive set of experiments utilizing the material blurry dataset and compare them to several state-of-theart deblurring methods.The experimental results demonstrate the applicability and effectiveness of MIDNet in the domain of deblurring material microstructure images,with a PSNR(Peak Signal-to-Noise Ratio)reaching 35.26 dB and an SSIM(Structural Similarity)of 0.946.Our dataset is available at:https://github.com/woshigui/MIDNet.

    KEYWORDS Image deblurring;material microstructure;attention mechanism;deep learning

    1 Introduction

    In the era of advanced imaging technology,modern material scientists delve into the microscopic realm,exploring and analyzing intricate phenomena.Among the array of methodologies available,scanning electron microscopy (SEM) emerges as a powerful tool for characterizing materials,and uncovering their morphologies,crystal structures,and chemical compositions [1].However,SEM images are susceptible to distortion,arising from instrument settings or operator inexperience,leading to blurred or defocused depictions that hinder research progress.When SEM images are blurry,the microstructural information of the material,such as crystal morphology,particle size,and pore structure,becomes less distinct,posing challenges to the accurate analysis of the material’s structural features.Additionally,the quantitative analysis of surface morphology,such as studying material texture and roughness,is also limited by the quality of SEM images.In the case of composite or multiphase materials,SEM images can reveal interface features between different phases.When the images are blurry,the interface structure may not be visible,thereby affecting the analysis and understanding of interface characteristics.SEM images are also employed for detecting defects in materials,such as cracks,voids,and particle non-uniformity.If the images are blurry,these defects may not be displayed,making defect detection and analysis difficult.The quest for effective deblurring techniques becomes paramount in ensuring the integrity of subsequent image analyses,particularly when grappling with suboptimal image quality.Our research is motivated by the imperative to investigate the paramount importance of this field,acknowledging the substantial adverse impact that blurry images can have on the further precise analysis of materials.Therefore,our study holds significant relevance in addressing this issue.We emphasize the pressing need for innovative deblurring solutions to address this issue effectively.

    Traditional image restoration techniques often lean on deconvolution methods that presuppose specific blur kernels,thereby crafting filters like local linear,nonlinear,non-local self-similarity,and Bayesian image restoration filters[2,3].However,their application in practical contexts remains challenging due to the prerequisite knowledge of blur kernels.The advent of deep learning revolutionizes image restoration,harnessing the prowess of deep neural networks to learn nonlinear mappings between degraded and sharp images,obviating the reliance on manually designed filters or blur kernels[4,5].Deep learning methods excel in preserving finer details,such as texture,edges,and structures,during the image reconstruction process [6].Furthermore,these methods demonstrate versatility in handling different levels of degradation and types of noise,allowing for image recovery across various scales [7].Its application has further extended to microscopic systems for enhancing image quality,encompassing optical microscopy[8],electromagnetic imaging,and scanning electron microscopy[9].

    While previous research has improved the quality of microscopic images,further investigation is warranted to explore the integration of deep learning for deblurring low-quality material microstructures.This inquiry begets key questions: (1) Can existing deblurring methods,which are applicable in real-world scenarios,be directly extended to address material data with blurred attributes using pre-trained weights?(2)Can retraining networks with material-specific blurry datasets lead to improvements in deblurring efficacy? (3)How can novel algorithms be developed to maximize their potential in enhancing the clarity of material microstructure images?

    In pursuit of these goals,we present a deep learning-based approach that combines soft attention mechanisms with multifaceted loss functions,aiming to enhance image quality while preserving intricate details.Our methodology tackles the challenge of image blurring in SEM images,arising from inaccurate hardware calibration or automation glitches.With our approach,researchers can efficiently rectify subpar images,saving significant time and resources that would otherwise be required for rescanning.This is particularly relevant for research projects that have limited budgets and require the rapid processing of numerous material samples within tight timeframes.In such circumstances,where image blurring continues to pose a recurring obstacle,our approach becomes especially crucial.In the dynamic field of high-throughput materials research,our innovation has the potential to enhance image quality and data fidelity,thereby accelerating the discovery and optimization of novel materials.In light of this,our approach emerges as a pivotal contribution,poised to catalyze diverse applications in the expanse of materials science research.The main contributions of this paper are as follows:

    (1) We propose a Material Images Deblurring Network (MIDNet) that specifically sharpens blurred images of material microstructures and outperforms current SOTA deblurring networks.

    (2) We introduce an attention mechanism that effectively mitigates the problem of inconsistent feature distributions by attending to the most informative features in both the encoder and decoder.This attention mechanism not only addresses the issue but also strengthens the interplay between components,enhancing overall performance.

    (3)We propose a novel multi-loss function that enhances the supervisory signal,thereby preserving intricate details and texture features more effectively.

    (4)Our MIDNet model’s superiority is thoroughly validated through rigorous experiments,both quantitatively and qualitatively.Through ablation experiments,we reveal the impact of different loss functions proposed in this paper on the model and demonstrate the effectiveness of constructing multiloss functions.

    2 Related Work

    2.1 Image Deblurring

    Several studies have combined computer science and materials science,with a particular emphasis on utilizing image processing methods for analyzing the microstructure images of materials.Varde[10]proposed a computational estimation method called AutoDomainMine,based on graph data mining.By integrating clustering and classification techniques,this method discovered knowledge from existing experimental data and utilized it for estimation.The main objective of this framework was to estimate the graphical results of experiments based on input conditions.Similar graph data mining methods can be employed for image deblurring tasks to analyze and extract patterns and features from image data to achieve image deblurring goals.Pan et al.[11] reviewed the evolution and impact of material microstructures during cutting processes,presenting a thermal-force-microstructure coupled modeling framework.They analyzed microstructural changes such as white layer formation,phase transformation,and dynamic recrystallization under different materials and cutting conditions,as well as the effects of these changes on cutting forces and surface integrity.Vibration of cutting tools or materials can cause motion in image acquisition devices (such as cameras) during the capturing process,resulting in image blurring.Therefore,studying the deblurring of material microstructures holds significant importance.

    Many traditional image enhancement methods employ regularization and manually crafted prior images for blur kernel estimation[12].Subsequent iterative optimization is used to gradually recover a clear image.However,this conventional approach involves intricate blur kernel estimation,leading to laborious sharpening,subpar real-time performance,and algorithmic limitations.To enhance the quality of image deblurring,many methods based on convolutional neural networks (CNN) have been proposed [13–15].Chakrabarti [13] designed a neural network to generate a global blur kernel for non-blind deconvolution.Song et al.[14]proposed a method using a neural network for reliable detection of motion blur kernels to detect image forgeries.Wang et al.[15]proposed a network-based framework that learned to remove raindrops by learning motion blur kernels.Sun et al.[16]predicted the probability distribution of non-uniform motion blur using CNNs.However,most neural-networkbased methods still rely on blur models to solve the blur kernel,limiting their performance.

    In recent years,with the development of deep learning,a series of methods based on deep learning have been used for image deblurring [17,18].Zhang et al.[19] proposed DMPHN,which is the first multi-scale network based on the multi-patch method for single-image deblurring.Chen et al.[20]proposed HINet,a deep image restoration network based on the HIN block.Fanous et al.[5]presented GANscan,a method for restoring sharp images from motion-blurred videos.The method was applied to reconstruct tissue sections under the microscope.Liang et al.[21]directly deblurred raw images using deep learning-based busy image-to-image blind-deblurring.DID-ANet[4]was designed specifically for single-image blur removal caused by camera misfocus.MedDeblur[18]was developed to remove blur in medical images due to patient movement or breathing.Xu et al.[22] proposed a deep-learning-based knowledge-enhanced image deblurring method for quality inspection in yarn production.Restormer[23]is an efficient transformer model that can be utilized for image restoration tasks at high resolutions.This model is effective for restoring high-resolution images.Chen et al.[7]found that nonlinear activation functions are not necessary and can be replaced or omitted,and developed NAFNet for both image denoising and deblurring.Due to the impressive performance of NAFNet in deblurring tasks,we are currently implementing modifications to its architecture.

    2.2 Attention-Based Deblurring Model

    In recent years,attention mechanisms have proven to be highly effective in various computer vision tasks[24,25].As a result,attention-based methods have gradually been adopted for the task of image deblurring[26,27].MSAN[28]is a convolutional neural network architecture based on attention that efficiently and effectively generalizes motion deblurring.D3-Net[26]can be used for deblurring,dehazing,and object detection,with the addition of a classification attention feature loss to improve deblurring and dehazing performance.Cui et al.[27] proposed a dual-domain attention mechanism that enhances feature expression in both spatial and frequency domains.Ma et al.[29] proposed an attention-based dehazing algorithm for deblurring to improve defect detection in inspection image pipelines.Shen et al.[30] introduced a supervised human-perception attention mechanism model,which performs exceptionally well in motion deblurring in an end-to-end manner.MALNET[31]is a lightweight network based on attention mechanisms,which also performs well in image deblurring.Zhang et al.[32]proposed an attention-based inter-frame compensation scheme for video deblurring.In this work,we also incorporate attention mechanisms into our image deblurring network to improve its deblurring capability.

    3 Method

    3.1 Architecture

    The network structure of this paper is shown in Fig.1.It follows a classical U-shape structure,which is an improvement from NAFNet[7].The structure comprises an encoder and a decoder,both belonging to the MID-Block.An attention mechanism is introduced between the blocks to improve the image restoration quality of the network.

    3.2 MID-Block

    MID-Block is the basic building block of MIDNet.To avoid high complexity between blocks,MID-Block does not use any nonlinear activation functions such as ReLU,GELU,and Softmax.We construct a MID-Block using analogies with NAFNet blocks,as illustrated in Fig.2.

    To stabilize the training process,the input is first passed through Layer Normalization.Next,the input undergoes convolution operations and is then processed by SimpleGate(SG)[7],which is a variant of Gated Linear Units(GLU)[33].The GLU formula is as follows:

    Figure 1: The MIDNet overview.The overall architecture of the network resembles a U-shape design,which is composed of MID-block and attention block

    Figure 2: Architecture of MID-block

    In Eq.(1),Xrepresents the feature map,fandgfunction as linear transformers,σrepresents a nonlinear activation function,such as Sigmoid,and ⊙represents element-wise multiplication.

    The GLU increases the intra-block complexity,which is not desirable.To remedy this issue,we reconsider the activation function in the block,specifically GELU[34],which is expressed as:

    whereφrepresents the cumulative distribution function of the standard normal distribution.According to reference[34],it is suggested that the Gaussian Error Linear Unit(GELU)activation function can be effectively approximated and implemented by employing the following methodology:

    In Eqs.(1)and(2),GELU is a specific case of GLU,where the activation functionsfandgare identity functions and the parameterσis substituted withφ.The GLU incorporates nonlinearity and is not reliant on the parameterσ.Even in the absence of the parameterσ,the expressionGate(X)=f(X) ⊙g(X)retains its nonlinearity.According to reference[7],we suggest a simple adjustment to GLU:Split the feature map into two parts along the channel dimension and multiply them.This could be done using a basic element-wise multiplication,which is represented by Eq.(4).

    In Eq.(4),XandYrepresent feature maps of equal proportions.

    The gating unit SG is a neural network component illustrated in Fig.3,which is used in the processing of feature maps.It operates by splitting the feature map into two parts along the channel dimension,which is then multiplied to generate the final output.By splitting the feature map in this manner,SG can selectively emphasize or de-emphasize specific channels in the feature map,which can be useful for enhancing certain features or suppressing noise in the signal.This process is often referred to as channel-wise gating.

    Figure 3: Simple gate as represented by Eq.(2).⊙:Element-wise

    Our novel approach introduces Simplified Channel Attention(SCA)[7],a new component that utilizes channel-wise attention to enhance relevant features in data.Compared to other approaches,SCA has a simpler structure which offers ease of implementation.Additionally,it adds minimal computational overhead to models,hence enhancing the efficiency of our approach.Please refer to Fig.4 for an illustration of SCA.

    Figure 4: Simplified channel attention(SCA).?:Channel-wise multiplicatio n

    SCA determines channel attention by computing the average of the feature map along the spatial dimensions and applying a fully connected layer to generate a channel-wise attention vector.This attention vector is then multiplied with the original feature map to selectively amplify important channels in the data while suppressing irrelevant or noisy channels.

    Our experiments demonstrate that incorporating SCA into a standard convolutional neural network yields improved performance,highlighting the efficacy of enhancing feature representation using channel attention.SCA can be easily integrated into existing neural network architectures and represents a useful tool for improving the performance of deep learning models in a variety of applications.

    SCA is derived from Channel Attention(CA)[35],which can be expressed by the Eq.(5).

    In Eq.(5),Xdenotes the feature map,pooldenotes the global average pooling operation,σdenotes an activation function such as Sigmoid,W1,andW2denote fully connected layers,and cross multiplication is the channel multiplication operation.By simplifying the Eq.(5),we can finally obtain SCA,as shown in Eq.(6).

    3.3 Attention Mechanism

    With the advancement of deep learning techniques,significant progress has been made in image restoration.The NAFNet model,in particular,has shown significant performance in various applications.However,a limitation of NAFNet is that the skip connections used for feature aggregation between the encoder and decoder have the potential to disrupt the feature distribution,resulting in inconsistencies between these components.Another shortcoming of NAFNet is that it only employs an intra-block attention mechanism and ignores attention-based skip connections.

    To address these challenges,we introduce a soft attention mechanism to capture the latent relationship between the encoder and decoder more adaptively.We refer to the proposed soft attention mechanism as ATT.The architecture of the attention gate ATT is shown in Fig.5.Specifically,the proposed attention gate ATT aggregates features from different blocks using a weighting scheme based on their relevance to the current image restoration task,instead of simple element-wise addition used in conventional skip connections.This allows the model to selectively focus on the most informative features while suppressing the irrelevant ones.

    Figure 5: The architecture of ATT

    Moreover,our attention mechanism enables us to incorporate attention-based skip connections,which further enhance the feature aggregation process.By attending to the most informative features in the encoder and decoder,the model can effectively alleviate issues related to feature distribution inconsistency and strengthen the correlation between these components.The formula of the soft attention mechanism can be expressed as follows:

    In Eqs.(7)and(8),σ1andσ2denote activation functions.The attention gate is represented by a set of parameters through mathematical formulas,including linear transformationsWa,Wb,ψ,and biasesbf,bψ.The linear transformations are obtained by performing convolution operations on the input tensors.The output of the attention gate is the product of the input feature map and the attention coefficient.

    3.4 Multi-Loss Function

    The paper utilizes multi-loss functions,as shown in Eq.(9),which comprise the deblurring loss,edge loss,and FFT loss.The hyperparametersλ1 andλ2 are assigned the values of 0.05 and 0.01,respectively.

    3.4.1 Deblurring Function

    The deblurred image is compared with its ground truth in the spatial domain,using the standardl1loss as shown in Eq.(10).We do not usel2loss because it sometimes over-penalizes errors and leads to poor deblurring performance.

    3.4.2 Edge Function

    To restore the high-frequency details of the image,we introduce an edge loss function.It aims to focus on the gradient information of the image and enhance the edge texture features.The edge loss function of this paper is as follows:

    In Eq.(11),Irrepresents the reconstructed image,Igtrepresents the clear ground truth image andΔdenotes the Laplacian operator.

    3.4.3 FFT Loss

    The FFT loss is a type of loss function based on the Fourier transform that is used for image restoration tasks.It aims to penalize the discrepancy between the reconstructed image and the ground truth image in the frequency domain.The FFT loss is represented as follows:

    In Eq.(12),the variablesWandHrefer to the width and height of the image being analyzed.The functionFrepresents the Fourier transform of the image,which is a mathematical technique used to analyze its frequency components.Wherewi,jrepresents the weight corresponding to the Fourier coefficient,andHrepresents the frequency response of the degradation function in the Fourier domain.

    Specifically,the FFT loss can be calculated as the weighted sum of the squared Euclidean distance between the discrete Fourier transform coefficients of the reconstructed image and the ground truth image.The weight factors,which correspond to different Fourier coefficients,are used to emphasize the importance of different frequencies in the loss function,allowing it to focus more on the crucial parts of the reconstructed image spectrum.In the Fourier domain,high-frequency information such as edges and textures has a more significant impact on the visual quality of the reconstructed image.Therefore,incorporating the FFT loss can help the network better preserve these details,ultimately leading to an improvement in the image quality.

    4 Experiments

    4.1 Dataset

    We utilize a dataset containing 120 paired images with both low and high quality to investigate material microstructure fuzziness.Specifically,low-quality images in this dataset are directly obtained from observations captured through the SEM rather than artificially blurred using blur kernels or algorithms.This approach replicates real-world scenarios more accurately while simultaneously presenting greater challenges for the process of deblurring.When low-quality images are captured in practice,operators take repeated images until high-quality ones are achieved.Consequently,we meticulously selected 120 matching low and high-quality images that met stringent criteria.All images are subsequently adjusted to 256 ?256 pixels.Several cropped images are displayed in Fig.6.The dataset is randomly divided into a training set comprising 108 image pairs and a test set containing 12 image pairs.

    4.2 Experiment Parameters

    We optimize the model using Adam (β1=0.9,β2=0.999) for 200 K iterations with a cosine annealing schedule that decreases the learning rate from 10–3 to 10–7.We crop the images to a size of 256 ?256 pixels and apply rotation and flipping as data augmentation techniques.We employed the skip-init method to ensure stable training and implemented our code in the PyTorch framework.We evaluate our model using peak signal-to-noise ratio(PSNR)and structural similarity(SSIM)metrics.All experiments are conducted on an NVIDIA Tesla V100 GPU.

    4.3 Experiments on SOTA Algorithms

    PSNR and SSIM are employed as quantitative evaluation metrics,with larger values indicating superior image quality.They are calculated according to Eqs.(13)and(14).

    Figure 6: A few sample images from our dataset.Column 1 shows the low-quality images,whereas Column 2 shows the high-quality images

    In Eq.(13),MAX represents the maximum pixel value of the image,typically 255 when each pixel is represented by an 8-bit binary.MSE(Mean Squared Error)is the mean squared error value between the blurred image and the clear image.In Eq.(14),xandydenote the original image and the deblurred image,respectively.μxandμyrepresent the mean pixel values of imagesxandy,σxandσyrepresent the standard deviations of pixel values in imagesxandy,andσxyis the covariance between the pixel values of the two images.C1andC2are constants introduced to prevent division by zero in the denominator.

    To assess the generalizability of models trained on natural images to material microstructure fuzziness data,we conduct a series of relevant studies.Specifically,we employ pre-trained weights from the original papers of DMPHN,HINet,Restormer,and NAFNet methods to conduct inference on material blurry images.The deblurred images are displayed in Fig.7,while the corresponding PSNR and SSIM values are summarized in Table 1.

    Table 1: Results of image deblurring by using pre-trained weights

    As observed in Fig.7,these methods exhibit certain levels of processing applied to the blurry images.However,their ability to achieve satisfactory deblurring outcomes remains limited,with minimal improvement over the initial blurry images.By referring to Table 1,the PSNR and SSIM values of both the original blurry and clear images are provided in the input row.Notably,these methods yield relatively low PSNR and SSIM scores,with instances where deblurred images demonstrate worse performance compared to their initial states.

    Figure 7: Image deblurring performance on the material blurry dataset is evaluated using several SOTA algorithms with pre-trained weights

    Interestingly,these methods have demonstrated proficiency on the GoPro dataset and have exhibited effective deblurring outcomes on real-world blurry images.Consequently,we postulate that their subpar performance on material images may be attributed to external factors rather than the inherent limitations of the methods themselves.

    Upon meticulous scrutiny of the GoPro dataset,a notable distinction emerges in the PSNR values of its blurry images,which average approximately 23.In contrast,the blurry images originating from our material microstructure exhibit a lower PSNR value of approximately 21.Building upon these observations,a hypothesis arises: The relatively lower quality of material images,resulting in reduced information content,poses a heightened challenge for the deblurring process.Consequently,this challenge could potentially contribute to network degradation and the suboptimal performance observed.

    Furthermore,an additional factor potentially influencing the subpar deblurring results is the unique visual characteristics inherent to material microstructures,setting them apart from real-world blurry images.This disparity in appearance might contribute to reduced reliability in the neural network’s performance when confronted with material microstructure fuzziness data.To address this challenge,we advocate for a proactive solution:Retraining and fine-tuning these methods using material blurry images.Our approach involves freezing the majority of the model layers and selectively unfreezing a small subset for training purposes.We apply data augmentation techniques,such as flip and rotate,to the dataset during the training process.Hyperparameters,including learning rate,batch size,and number of iterations,are adjusted based on the specific model to achieve optimal performance.Additionally,appropriate regularization strategies are employed to mitigate overfitting problems.Such an approach holds the promise of enhancing the network’s capability to effectively restore blurry images of materials.In line with this recommendation,we embarked on the process of retraining and fine-tuning these methods.To gauge the efficacy of this intervention,we present the deblurring outcomes in Fig.8.

    Figure 8: The outcomes of deblurring upon the retraining and fine-tuning of these methods with our blurry dataset

    This study utilizes a dataset of material blurry images to conduct a detailed analysis of the deblurring capability of the original method compared to the retraining and fine-tuning methods.The outcomes of this comparison reveal a significant enhancement in deblurring quality for material images through retraining and fine-tuning,surpassing the performance of the no-training scenario and yielding satisfactory results.Notably,the process of retraining and fine-tuning contributes to the restoration of intricate features within material images,underscoring the pivotal role of materialspecific data in optimizing deblurring effectiveness.These findings offer fresh insights into the efficacy of retraining and fine-tuning strategies in effectively addressing the intricate deblurring challenges posed by material images.Furthermore,they provide valuable guidance for the future development of more potent deblurring methodologies within the domain of material science and engineering.Importantly,this study also serves as a demonstration of the potential of deep learning techniques in enhancing the quality of visual data across a wide spectrum of scientific and industrial applications.

    4.4 Comparative Experiment

    4.4.1 Qualitative Results

    We undertake a comparative evaluation of MIDNet alongside several SOTA deblurring methods that have undergone retraining and fine-tuning,as discussed in the previous section.The deblurring outcomes produced by each of these methods are depicted in Fig.9.Within this array of tested approaches,Restormer’s results exhibit a residual blurriness accompanied by unclear edges,which implies a limited restorative impact.The HINet method,employing a patch-based testing strategy,manifests noticeable stripe artifacts,possibly attributed to boundary discontinuities.The DMPHN approach,although improved,still retains a degree of blurriness that hampers its ability to achieve significant image enhancement.The NAFNet method,while competent,sacrifices certain fine image details.In stark contrast,our proposed MIDNet method achieves a further elevation in image quality,facilitating the restoration of additional structural details without introducing any artifacts or related issues.By observing the image,we note that our method exhibits significantly clearer microstructural contours compared to other approaches,as indicated by the red arrow in Fig.9.This enhanced clarity allows for a more accurate analysis of the material’s surface morphology and structural features based on these finer details.

    Figure 9: Qualitative comparison of image deblurring methods on the dataset

    The comparison between the original image and the deblurred image obtained through the model proposed in this study is illustrated in Fig.10.In Fig.10a,we present the original image,while Fig.10b depicts the image after being processed by the model.Through visual observation,it is evident that the proposed model exhibits excellent deblurring performance.The outcomes of our study highlight the exceptional capabilities of MIDNet in effectively recovering intricate structures and details within material images.This showcases its potential as a promising solution for tackling intricate deblurring issues within the realm of materials science and engineering.

    4.4.2 Quantitative Results

    Table 2 outlines the quantitative findings of several deblurring techniques applied to material microstructure images.Our evaluation of image quality relies on two objective metrics: PSNR and SSIM,where higher values denote enhanced performance.Significant enhancements in PSNR are observed across HINet,Restormer,DMPHN,and NAFNet after the process of retraining and fine-tuning.The respective gains in PSNR are 7.89,9.43,10.13,and 13.53 dB.These compelling outcomes underscore the considerable potential of deep learning in addressing the intricate challenges associated with deblurring material microstructure images.This progress lays the foundation for practical applications within this domain.

    Table 2: Quantitative comparison of our proposed network with previous methods

    The insights provided by Table 2 highlight the substantial advancement brought forth by MIDNet,when compared with NAFNet,evaluated through both PSNR and SSIM metrics.Compared to NAFNet,MIDNet achieved an improvement of 1.45 dB in PSNR and 0.01 in SSIM.This indicates that our proposed method has an advantage in image deblurring.The efficacy of MIDNet in the deblurring task can be attributed to its integrative employment of an attention mechanism and a combination of diverse loss functions.

    Figure 10: Comparison between original images and deblurred images

    The attention mechanism significantly enhances the network’s ability to focus on pivotal features,leading to elevated deblurring performance.Our experiment results affirm that the simultaneous utilization of multiple loss functions empowers the network with enhanced image reconstruction supervision,consequently elevating image quality and augmenting fine detail preservation.

    4.4.3 Ablation Experiment

    To validate the efficacy of the newly introduced edge loss and FFT loss within the training process,we conduct ablation experiments.The outcomes of these experiments are meticulously presented in Table 3,showcasing the computed PSNR and SSIM values corresponding to each experimental configuration.The objective behind these ablation studies is to discern the impact and contribution of individual loss functions toward the process of image restoration.To achieve this,we train our model under different scenarios,each characterized by a distinct combination of loss functions.This systematic approach enables us to gain insights into the relative importance and effectiveness of each loss function in driving the enhancement of image quality.

    Table 3: Ablation experiments:We train our model using different combinations of loss functions to understand the importance of individual losses for image restoration

    In this study,we undertake a series of ablation experiments with the intent of examining the impact of integrating various loss functions during the training phase.To maintain consistency,the Ldloss function,which plays a pivotal role in image restoration,is kept constant across all experiments.The outcomes of these ablation studies are summarized in Table 3.We observe that the inclusion of the Leloss function results in noticeable improvements in both PSNR and SSIM metrics.This suggests that the network effectively retains more intricate edge details through the utilization of this loss function.Furthermore,the inclusion of Lfloss further improves the image quality by providing more structural guidance to the network solution,as observed in row 3.It is worth noting that by combining all the loss functions during training,the network achieved its best performance.These findings highlight the importance of the proposed multi-loss functions in enhancing image restoration capabilities and offer valuable insights for the advancement of effective image restoration methods.

    5 Conclusions and Future Work

    In this study,we propose a method named MIDNet to address the issue of blurry images in material microstructures.MIDNet is an end-to-end deblurring network that enhances the clarity of blurry images in material microstructures by incorporating an attention mechanism and introducing multiple loss functions.Thorough qualitative and quantitative analysis indicates that MIDNet surpasses other approaches in terms of the quality of reconstructed images,marked by enhanced clarity and texture richness.Ablation experiments have also showcased the effectiveness of different loss functions within the network.Our work has the potential to encourage the extended use of deep learning within materials science and promote advancements in the mutually beneficial partnership between computer science and materials science.

    The dataset utilized in this study comprises actual experimental material microstructural images.However,we acknowledge that the dataset size is relatively limited,which may potentially impact the accuracy of image deblurring when extrapolating our method to diverse materials.To address this limitation,our future research will emphasize the collection of SEM images encompassing a broader range of alloy materials,thereby expanding the dataset size.Through these endeavors,we aim to enhance the performance and adaptability of our model in the context of deblurring microstructural images across various materials.Our future work will be primarily focused on developing a video deblurring method that is specifically tailored to the demands of material science applications.Given the unique challenges posed by the complex and dynamic nature of material structures,a robust and effective video deblurring method would be of great value in enabling researchers to visualize and analyze material properties more accurately.

    Acknowledgement:The authors especially acknowledge Prof.Liwu Jiang of National Center for Materials Service Safety.

    Funding Statement:The current work was supported by the National Key R&D Program of China(Grant No.2021YFA1601104),National Key R&D Program of China(Grant No.2022YFA16038004),National Key R&D Program of China (Grant No.2022YFA16038002) and National Science and Technology Major Project of China(No.J2019-VI-0004-0117).

    Author Contributions:Study conception and design:J.X.Wang,H.Y.Yu and D.B.Sun;data collection:J.X.Wang,Z.Y.Li and P Shi;analysis and interpretation of result: J.X.Wang,P Shi.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:Some dataset for the experiments uploaded to the author’s github repository:https://github.com/woshigui/MIDNet.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    妹子高潮喷水视频| 日日爽夜夜爽网站| 国产 精品1| 纯流量卡能插随身wifi吗| 午夜影院在线不卡| 日韩熟女老妇一区二区性免费视频| 日产精品乱码卡一卡2卡三| 一级毛片 在线播放| av国产精品久久久久影院| 少妇熟女欧美另类| 尾随美女入室| 如日韩欧美国产精品一区二区三区| 一边亲一边摸免费视频| 国产成人aa在线观看| 亚洲精品av麻豆狂野| 下体分泌物呈黄色| 两个人免费观看高清视频| 全区人妻精品视频| 欧美精品国产亚洲| 精品一区二区三区视频在线| 极品少妇高潮喷水抽搐| 全区人妻精品视频| 免费久久久久久久精品成人欧美视频 | 亚洲天堂av无毛| 欧美另类一区| 亚洲精品日本国产第一区| 少妇精品久久久久久久| 免费观看在线日韩| av又黄又爽大尺度在线免费看| 美女内射精品一级片tv| 国产日韩欧美亚洲二区| 国产有黄有色有爽视频| 七月丁香在线播放| 热re99久久国产66热| 晚上一个人看的免费电影| 熟女av电影| 男的添女的下面高潮视频| 我的女老师完整版在线观看| 好男人视频免费观看在线| 国产精品久久久久久久久免| 欧美日本中文国产一区发布| 免费在线观看黄色视频的| av一本久久久久| 欧美精品av麻豆av| 乱码一卡2卡4卡精品| 精品99又大又爽又粗少妇毛片| 精品亚洲乱码少妇综合久久| 欧美老熟妇乱子伦牲交| 亚洲 欧美一区二区三区| 亚洲av福利一区| 国产亚洲欧美精品永久| 日韩三级伦理在线观看| 亚洲精品色激情综合| 免费大片18禁| 国产精品嫩草影院av在线观看| www日本在线高清视频| 久久人人爽av亚洲精品天堂| 亚洲av福利一区| 欧美bdsm另类| 国产 一区精品| 亚洲精品久久久久久婷婷小说| 亚洲激情五月婷婷啪啪| 亚洲精品,欧美精品| 国产一区二区三区av在线| 菩萨蛮人人尽说江南好唐韦庄| 日韩中文字幕视频在线看片| 五月玫瑰六月丁香| 国产精品三级大全| 男女国产视频网站| 啦啦啦视频在线资源免费观看| 久久午夜综合久久蜜桃| 人人澡人人妻人| 大片免费播放器 马上看| 熟女人妻精品中文字幕| 香蕉精品网在线| 亚洲精品美女久久久久99蜜臀 | 少妇猛男粗大的猛烈进出视频| 亚洲国产精品成人久久小说| 咕卡用的链子| 热99国产精品久久久久久7| 99久国产av精品国产电影| 最后的刺客免费高清国语| 黑人欧美特级aaaaaa片| 女人精品久久久久毛片| 久久99精品国语久久久| 国产免费福利视频在线观看| 国产探花极品一区二区| 午夜影院在线不卡| 欧美成人午夜免费资源| 99热这里只有是精品在线观看| 夜夜骑夜夜射夜夜干| 久久影院123| 日本爱情动作片www.在线观看| 日韩大片免费观看网站| 欧美日韩成人在线一区二区| 老司机影院毛片| 亚洲精品美女久久久久99蜜臀 | 性色av一级| 国产 精品1| 性色avwww在线观看| 在线观看国产h片| xxx大片免费视频| 少妇人妻 视频| 在线观看免费日韩欧美大片| 久久女婷五月综合色啪小说| 人妻一区二区av| 久久精品aⅴ一区二区三区四区 | 国产成人免费无遮挡视频| 成人影院久久| 久久人人爽人人爽人人片va| 亚洲美女视频黄频| 国产亚洲精品第一综合不卡 | 日本黄大片高清| 精品国产一区二区三区久久久樱花| 一本一本久久a久久精品综合妖精 国产伦在线观看视频一区 | 99热这里只有是精品在线观看| 国产一区有黄有色的免费视频| 国产精品不卡视频一区二区| 夫妻性生交免费视频一级片| 国产精品无大码| 丰满乱子伦码专区| 亚洲四区av| 亚洲图色成人| 久久久久久久久久成人| 婷婷色麻豆天堂久久| 激情五月婷婷亚洲| 91在线精品国自产拍蜜月| 婷婷色综合www| 高清欧美精品videossex| 99热这里只有是精品在线观看| 在线观看免费视频网站a站| 在线天堂中文资源库| 日本与韩国留学比较| 国产片内射在线| 99久久精品国产国产毛片| 欧美成人午夜免费资源| 久久久久国产网址| 大香蕉久久成人网| √禁漫天堂资源中文www| 高清欧美精品videossex| 中文字幕免费在线视频6| 成人综合一区亚洲| 亚洲精品美女久久久久99蜜臀 | 黑人猛操日本美女一级片| 久久久久精品性色| 国产一区有黄有色的免费视频| 国产在线免费精品| 美女主播在线视频| 亚洲欧洲精品一区二区精品久久久 | 亚洲,一卡二卡三卡| 久久精品国产鲁丝片午夜精品| 欧美日本中文国产一区发布| 日本-黄色视频高清免费观看| 在线看a的网站| 国产精品久久久久久久电影| 婷婷色麻豆天堂久久| 精品少妇久久久久久888优播| 免费黄网站久久成人精品| 亚洲成人av在线免费| 啦啦啦在线观看免费高清www| 亚洲五月色婷婷综合| 高清视频免费观看一区二区| 七月丁香在线播放| 国产亚洲一区二区精品| 久久鲁丝午夜福利片| 免费不卡的大黄色大毛片视频在线观看| 男人爽女人下面视频在线观看| 午夜免费男女啪啪视频观看| 日本午夜av视频| 亚洲成人手机| 日韩视频在线欧美| 欧美日韩成人在线一区二区| av不卡在线播放| 另类亚洲欧美激情| 亚洲综合色惰| 韩国高清视频一区二区三区| 国产免费福利视频在线观看| 午夜免费鲁丝| 亚洲精品一区蜜桃| 国产黄色免费在线视频| 日本爱情动作片www.在线观看| 97在线人人人人妻| 久久99蜜桃精品久久| 黄色毛片三级朝国网站| 午夜视频国产福利| 五月伊人婷婷丁香| 18禁在线无遮挡免费观看视频| 精品国产一区二区三区四区第35| 最黄视频免费看| 99久久精品国产国产毛片| 欧美老熟妇乱子伦牲交| 中文精品一卡2卡3卡4更新| 侵犯人妻中文字幕一二三四区| 在线精品无人区一区二区三| av又黄又爽大尺度在线免费看| 人妻系列 视频| 久久久久精品人妻al黑| 亚洲欧洲日产国产| av在线app专区| 天天躁夜夜躁狠狠久久av| 午夜福利视频精品| 另类亚洲欧美激情| 一级毛片 在线播放| 97人妻天天添夜夜摸| 伦理电影大哥的女人| 色吧在线观看| 欧美日韩亚洲高清精品| 1024视频免费在线观看| 春色校园在线视频观看| 伦理电影大哥的女人| 有码 亚洲区| 如何舔出高潮| 亚洲精品自拍成人| 欧美 亚洲 国产 日韩一| 久久久久久久久久成人| 精品一区在线观看国产| 黑人巨大精品欧美一区二区蜜桃 | 青春草国产在线视频| 精品国产一区二区三区久久久樱花| 制服人妻中文乱码| 91成人精品电影| 日本av免费视频播放| 免费日韩欧美在线观看| 色网站视频免费| 搡老乐熟女国产| av在线老鸭窝| 菩萨蛮人人尽说江南好唐韦庄| 少妇熟女欧美另类| 全区人妻精品视频| 精品少妇内射三级| 欧美激情国产日韩精品一区| 国产男女内射视频| 乱码一卡2卡4卡精品| 中文欧美无线码| 久久国内精品自在自线图片| 性高湖久久久久久久久免费观看| 午夜免费鲁丝| 18禁国产床啪视频网站| 国产精品国产三级国产av玫瑰| 黄网站色视频无遮挡免费观看| 欧美日韩综合久久久久久| 在线观看国产h片| 一级爰片在线观看| 亚洲av.av天堂| 三级国产精品片| 高清视频免费观看一区二区| 精品一区在线观看国产| 亚洲国产看品久久| 中文字幕人妻熟女乱码| 欧美3d第一页| 亚洲精品乱码久久久久久按摩| 王馨瑶露胸无遮挡在线观看| 尾随美女入室| 伦理电影免费视频| 精品国产一区二区三区四区第35| 18禁裸乳无遮挡动漫免费视频| 亚洲av综合色区一区| 亚洲欧美一区二区三区黑人 | 成年av动漫网址| 1024视频免费在线观看| 欧美日韩av久久| 色94色欧美一区二区| 伦理电影免费视频| 成年av动漫网址| 免费女性裸体啪啪无遮挡网站| 免费在线观看黄色视频的| 激情五月婷婷亚洲| 国产精品久久久久久精品电影小说| 妹子高潮喷水视频| 80岁老熟妇乱子伦牲交| 永久网站在线| 久久这里只有精品19| 欧美日韩国产mv在线观看视频| 女的被弄到高潮叫床怎么办| 成年女人在线观看亚洲视频| 熟女电影av网| 久久这里只有精品19| 久久国内精品自在自线图片| av在线app专区| 欧美老熟妇乱子伦牲交| 日本黄大片高清| 极品人妻少妇av视频| 美女福利国产在线| 精品人妻一区二区三区麻豆| 满18在线观看网站| 国产片内射在线| 国产高清三级在线| 美女国产视频在线观看| 国产在线一区二区三区精| 久久99蜜桃精品久久| 午夜激情av网站| 精品国产一区二区三区久久久樱花| 看免费av毛片| 国产又色又爽无遮挡免| 激情五月婷婷亚洲| 少妇被粗大的猛进出69影院 | 婷婷成人精品国产| 亚洲精品,欧美精品| 熟女人妻精品中文字幕| 母亲3免费完整高清在线观看 | 99热6这里只有精品| 亚洲av日韩在线播放| av卡一久久| 捣出白浆h1v1| 观看美女的网站| 免费人成在线观看视频色| 亚洲综合色惰| 涩涩av久久男人的天堂| 国产成人免费观看mmmm| 乱人伦中国视频| 国产精品一国产av| 久久99热6这里只有精品| 欧美精品一区二区大全| 国产精品熟女久久久久浪| 欧美日韩综合久久久久久| 久久 成人 亚洲| 午夜福利,免费看| √禁漫天堂资源中文www| 天天影视国产精品| 久久久久视频综合| 男女午夜视频在线观看 | 久久国产精品男人的天堂亚洲 | 蜜桃在线观看..| 国产男女内射视频| 久久久精品免费免费高清| 亚洲精品乱久久久久久| 午夜影院在线不卡| 一级片'在线观看视频| 极品少妇高潮喷水抽搐| 成人国产av品久久久| 国产精品久久久久成人av| av有码第一页| 亚洲高清免费不卡视频| 久久久久久久久久久免费av| 亚洲婷婷狠狠爱综合网| 成人亚洲欧美一区二区av| 亚洲,一卡二卡三卡| 久久精品国产亚洲av涩爱| 黑人欧美特级aaaaaa片| 久久精品国产亚洲av涩爱| 两个人看的免费小视频| 午夜福利,免费看| 久久久久人妻精品一区果冻| 免费在线观看完整版高清| 有码 亚洲区| 国产成人精品一,二区| 日本wwww免费看| 蜜桃国产av成人99| 成人黄色视频免费在线看| 99热这里只有是精品在线观看| 亚洲激情五月婷婷啪啪| 建设人人有责人人尽责人人享有的| 母亲3免费完整高清在线观看 | 亚洲成人一二三区av| 18在线观看网站| 亚洲欧美一区二区三区国产| 日韩av不卡免费在线播放| 97人妻天天添夜夜摸| 男的添女的下面高潮视频| 久久久久精品性色| 视频在线观看一区二区三区| 久久99热6这里只有精品| av在线观看视频网站免费| 少妇 在线观看| 亚洲精品视频女| a 毛片基地| 永久免费av网站大全| 国产老妇伦熟女老妇高清| 久久人妻熟女aⅴ| 一区二区三区乱码不卡18| 九九在线视频观看精品| 欧美人与性动交α欧美软件 | 麻豆乱淫一区二区| 欧美亚洲 丝袜 人妻 在线| 欧美成人午夜精品| 波多野结衣一区麻豆| 日韩一区二区三区影片| av播播在线观看一区| 建设人人有责人人尽责人人享有的| 只有这里有精品99| 晚上一个人看的免费电影| 午夜av观看不卡| 国产精品99久久99久久久不卡 | 色网站视频免费| 王馨瑶露胸无遮挡在线观看| 在线天堂最新版资源| 高清不卡的av网站| 18+在线观看网站| 97在线视频观看| 国产xxxxx性猛交| 免费久久久久久久精品成人欧美视频 | 成人亚洲欧美一区二区av| 咕卡用的链子| 国产成人精品无人区| 女性被躁到高潮视频| 男的添女的下面高潮视频| 国产黄色视频一区二区在线观看| 国产黄频视频在线观看| 妹子高潮喷水视频| 少妇人妻 视频| 精品人妻一区二区三区麻豆| 自拍欧美九色日韩亚洲蝌蚪91| 午夜影院在线不卡| www.熟女人妻精品国产 | 精品国产国语对白av| 亚洲人成77777在线视频| 精品卡一卡二卡四卡免费| 九草在线视频观看| 人妻人人澡人人爽人人| 中文天堂在线官网| 久久人人爽人人片av| 2022亚洲国产成人精品| 国产亚洲一区二区精品| 99热国产这里只有精品6| 中文字幕免费在线视频6| 日韩制服骚丝袜av| 久久国产精品男人的天堂亚洲 | 成年人免费黄色播放视频| 青青草视频在线视频观看| 看免费成人av毛片| 国产成人a∨麻豆精品| 99国产综合亚洲精品| 大片免费播放器 马上看| 一级毛片黄色毛片免费观看视频| 黄色视频在线播放观看不卡| 婷婷色麻豆天堂久久| 亚洲精品中文字幕在线视频| 国产欧美日韩综合在线一区二区| 伊人久久国产一区二区| av免费在线看不卡| 国产日韩欧美在线精品| 制服人妻中文乱码| 女人久久www免费人成看片| 18禁在线无遮挡免费观看视频| 中文字幕另类日韩欧美亚洲嫩草| 中文乱码字字幕精品一区二区三区| 中文天堂在线官网| 婷婷色av中文字幕| 久久久久久伊人网av| 精品一区二区三卡| 免费大片18禁| www.av在线官网国产| 亚洲 欧美一区二区三区| 久久久国产精品麻豆| 夜夜骑夜夜射夜夜干| 美女大奶头黄色视频| 激情视频va一区二区三区| 免费在线观看黄色视频的| 国产亚洲av片在线观看秒播厂| 亚洲伊人久久精品综合| 国产xxxxx性猛交| 国产精品三级大全| 不卡视频在线观看欧美| 少妇猛男粗大的猛烈进出视频| 欧美日韩综合久久久久久| 亚洲国产色片| 1024视频免费在线观看| 在线亚洲精品国产二区图片欧美| 桃花免费在线播放| 国产色爽女视频免费观看| 免费高清在线观看视频在线观看| 秋霞在线观看毛片| 亚洲欧洲国产日韩| 国产精品久久久av美女十八| 在线观看国产h片| 亚洲中文av在线| 亚洲综合精品二区| 边亲边吃奶的免费视频| 蜜桃国产av成人99| 亚洲综合色惰| 男人添女人高潮全过程视频| 国产亚洲最大av| 不卡视频在线观看欧美| 春色校园在线视频观看| 母亲3免费完整高清在线观看 | 男女午夜视频在线观看 | 九草在线视频观看| 热99国产精品久久久久久7| 汤姆久久久久久久影院中文字幕| 熟妇人妻不卡中文字幕| 日韩av免费高清视频| 欧美人与善性xxx| 午夜91福利影院| 纵有疾风起免费观看全集完整版| 婷婷成人精品国产| 少妇的丰满在线观看| kizo精华| videossex国产| 国产永久视频网站| 亚洲精品国产av蜜桃| 成人国产av品久久久| 人人澡人人妻人| 成人亚洲欧美一区二区av| 黄色视频在线播放观看不卡| 看非洲黑人一级黄片| 亚洲美女搞黄在线观看| 五月玫瑰六月丁香| 亚洲成av片中文字幕在线观看 | 亚洲人与动物交配视频| 五月伊人婷婷丁香| 国产色婷婷99| 亚洲情色 制服丝袜| 亚洲成色77777| 人妻少妇偷人精品九色| 免费在线观看黄色视频的| 精品一区二区三卡| 久久人人爽av亚洲精品天堂| 大陆偷拍与自拍| 欧美精品亚洲一区二区| 视频区图区小说| 在线观看免费视频网站a站| 啦啦啦中文免费视频观看日本| 国产av码专区亚洲av| 老司机影院成人| 蜜臀久久99精品久久宅男| 天天躁夜夜躁狠狠躁躁| 亚洲综合色惰| 国产精品国产av在线观看| 热re99久久国产66热| 欧美精品人与动牲交sv欧美| 天天躁夜夜躁狠狠躁躁| 多毛熟女@视频| 女性生殖器流出的白浆| 美女国产高潮福利片在线看| 亚洲熟女精品中文字幕| 夜夜爽夜夜爽视频| 校园人妻丝袜中文字幕| 久久99一区二区三区| 最近2019中文字幕mv第一页| 18在线观看网站| 97在线人人人人妻| 欧美性感艳星| 亚洲第一区二区三区不卡| 免费看av在线观看网站| 国产男人的电影天堂91| 精品少妇黑人巨大在线播放| www日本在线高清视频| 亚洲天堂av无毛| 国产深夜福利视频在线观看| 狂野欧美激情性xxxx在线观看| 国产高清不卡午夜福利| 亚洲欧美日韩另类电影网站| 亚洲av男天堂| 亚洲,欧美,日韩| 婷婷色综合www| 制服诱惑二区| 精品一区二区三卡| 卡戴珊不雅视频在线播放| 亚洲精品aⅴ在线观看| 女人被躁到高潮嗷嗷叫费观| 国产精品三级大全| 久久国产精品大桥未久av| 日韩一区二区三区影片| 午夜av观看不卡| 亚洲伊人色综图| 爱豆传媒免费全集在线观看| 美女中出高潮动态图| 亚洲精品色激情综合| 国产一区有黄有色的免费视频| 一本色道久久久久久精品综合| 免费黄网站久久成人精品| 少妇人妻久久综合中文| 亚洲图色成人| 色吧在线观看| 国产色婷婷99| 少妇猛男粗大的猛烈进出视频| 一级,二级,三级黄色视频| 亚洲国产精品国产精品| 一级毛片电影观看| 美女脱内裤让男人舔精品视频| 啦啦啦啦在线视频资源| 大片电影免费在线观看免费| 岛国毛片在线播放| 亚洲精品第二区| 18禁裸乳无遮挡动漫免费视频| 久久久久精品人妻al黑| 熟女人妻精品中文字幕| 久久久久久伊人网av| 欧美激情国产日韩精品一区| 久久久久精品性色| 少妇人妻久久综合中文| 欧美激情国产日韩精品一区| 亚洲欧美中文字幕日韩二区| 菩萨蛮人人尽说江南好唐韦庄| 尾随美女入室| www日本在线高清视频| 视频区图区小说| 国产精品不卡视频一区二区| 热99国产精品久久久久久7| 国产极品粉嫩免费观看在线| 少妇被粗大的猛进出69影院 | 欧美日本中文国产一区发布| 国产成人aa在线观看| av免费在线看不卡| 日韩,欧美,国产一区二区三区| 男人爽女人下面视频在线观看| 精品一区二区三区四区五区乱码 | 午夜久久久在线观看| 成人亚洲精品一区在线观看| 国产精品一二三区在线看| 亚洲综合色网址| 女性生殖器流出的白浆| 国产精品一二三区在线看| 久久久国产一区二区| 夫妻性生交免费视频一级片| 亚洲av综合色区一区| √禁漫天堂资源中文www| 中文字幕另类日韩欧美亚洲嫩草| 欧美日韩亚洲高清精品| 国产高清国产精品国产三级| 大话2 男鬼变身卡| 99久久人妻综合| 久久久久久久精品精品| 亚洲激情五月婷婷啪啪| 亚洲欧美色中文字幕在线|