• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    MIDNet:Deblurring Network for Material Microstructure Images

    2024-05-25 14:42:00JiaxiangWangZhengyiLiPengShiHongyingYuandDongbaiSun
    Computers Materials&Continua 2024年4期

    Jiaxiang Wang ,Zhengyi Li ,Peng Shi ,Hongying Yu and Dongbai Sun,3,?

    1National Center for Materials Service Safety,University of Science and Technology Beijing,Beijing,100083,China

    2School of Materials,Sun Yat-Sen University,Shenzhen,518107,China

    3School of Materials Science and Engineering,Southern Marine Science and Engineering Guangdong Laboratory(Zhuhai),Sun Yat-Sen University,Guangzhou,510006,China

    ABSTRACT Scanning electron microscopy(SEM)is a crucial tool in the field of materials science,providing valuable insights into the microstructural characteristics of materials.Unfortunately,SEM images often suffer from blurriness caused by improper hardware calibration or imaging automation errors,which present challenges in analyzing and interpreting material characteristics.Consequently,rectifying the blurring of these images assumes paramount significance to enable subsequent analysis.To address this issue,we introduce a Material Images Deblurring Network (MIDNet) built upon the foundation of the Nonlinear Activation Free Network (NAFNet).MIDNet is meticulously tailored to address the blurring in images capturing the microstructure of materials.The key contributions include enhancing the NAFNet architecture for better feature extraction and representation,integrating a novel soft attention mechanism to uncover important correlations between encoder and decoder,and introducing new multi-loss functions to improve training effectiveness and overall model performance.We conduct a comprehensive set of experiments utilizing the material blurry dataset and compare them to several state-of-theart deblurring methods.The experimental results demonstrate the applicability and effectiveness of MIDNet in the domain of deblurring material microstructure images,with a PSNR(Peak Signal-to-Noise Ratio)reaching 35.26 dB and an SSIM(Structural Similarity)of 0.946.Our dataset is available at:https://github.com/woshigui/MIDNet.

    KEYWORDS Image deblurring;material microstructure;attention mechanism;deep learning

    1 Introduction

    In the era of advanced imaging technology,modern material scientists delve into the microscopic realm,exploring and analyzing intricate phenomena.Among the array of methodologies available,scanning electron microscopy (SEM) emerges as a powerful tool for characterizing materials,and uncovering their morphologies,crystal structures,and chemical compositions [1].However,SEM images are susceptible to distortion,arising from instrument settings or operator inexperience,leading to blurred or defocused depictions that hinder research progress.When SEM images are blurry,the microstructural information of the material,such as crystal morphology,particle size,and pore structure,becomes less distinct,posing challenges to the accurate analysis of the material’s structural features.Additionally,the quantitative analysis of surface morphology,such as studying material texture and roughness,is also limited by the quality of SEM images.In the case of composite or multiphase materials,SEM images can reveal interface features between different phases.When the images are blurry,the interface structure may not be visible,thereby affecting the analysis and understanding of interface characteristics.SEM images are also employed for detecting defects in materials,such as cracks,voids,and particle non-uniformity.If the images are blurry,these defects may not be displayed,making defect detection and analysis difficult.The quest for effective deblurring techniques becomes paramount in ensuring the integrity of subsequent image analyses,particularly when grappling with suboptimal image quality.Our research is motivated by the imperative to investigate the paramount importance of this field,acknowledging the substantial adverse impact that blurry images can have on the further precise analysis of materials.Therefore,our study holds significant relevance in addressing this issue.We emphasize the pressing need for innovative deblurring solutions to address this issue effectively.

    Traditional image restoration techniques often lean on deconvolution methods that presuppose specific blur kernels,thereby crafting filters like local linear,nonlinear,non-local self-similarity,and Bayesian image restoration filters[2,3].However,their application in practical contexts remains challenging due to the prerequisite knowledge of blur kernels.The advent of deep learning revolutionizes image restoration,harnessing the prowess of deep neural networks to learn nonlinear mappings between degraded and sharp images,obviating the reliance on manually designed filters or blur kernels[4,5].Deep learning methods excel in preserving finer details,such as texture,edges,and structures,during the image reconstruction process [6].Furthermore,these methods demonstrate versatility in handling different levels of degradation and types of noise,allowing for image recovery across various scales [7].Its application has further extended to microscopic systems for enhancing image quality,encompassing optical microscopy[8],electromagnetic imaging,and scanning electron microscopy[9].

    While previous research has improved the quality of microscopic images,further investigation is warranted to explore the integration of deep learning for deblurring low-quality material microstructures.This inquiry begets key questions: (1) Can existing deblurring methods,which are applicable in real-world scenarios,be directly extended to address material data with blurred attributes using pre-trained weights?(2)Can retraining networks with material-specific blurry datasets lead to improvements in deblurring efficacy? (3)How can novel algorithms be developed to maximize their potential in enhancing the clarity of material microstructure images?

    In pursuit of these goals,we present a deep learning-based approach that combines soft attention mechanisms with multifaceted loss functions,aiming to enhance image quality while preserving intricate details.Our methodology tackles the challenge of image blurring in SEM images,arising from inaccurate hardware calibration or automation glitches.With our approach,researchers can efficiently rectify subpar images,saving significant time and resources that would otherwise be required for rescanning.This is particularly relevant for research projects that have limited budgets and require the rapid processing of numerous material samples within tight timeframes.In such circumstances,where image blurring continues to pose a recurring obstacle,our approach becomes especially crucial.In the dynamic field of high-throughput materials research,our innovation has the potential to enhance image quality and data fidelity,thereby accelerating the discovery and optimization of novel materials.In light of this,our approach emerges as a pivotal contribution,poised to catalyze diverse applications in the expanse of materials science research.The main contributions of this paper are as follows:

    (1) We propose a Material Images Deblurring Network (MIDNet) that specifically sharpens blurred images of material microstructures and outperforms current SOTA deblurring networks.

    (2) We introduce an attention mechanism that effectively mitigates the problem of inconsistent feature distributions by attending to the most informative features in both the encoder and decoder.This attention mechanism not only addresses the issue but also strengthens the interplay between components,enhancing overall performance.

    (3)We propose a novel multi-loss function that enhances the supervisory signal,thereby preserving intricate details and texture features more effectively.

    (4)Our MIDNet model’s superiority is thoroughly validated through rigorous experiments,both quantitatively and qualitatively.Through ablation experiments,we reveal the impact of different loss functions proposed in this paper on the model and demonstrate the effectiveness of constructing multiloss functions.

    2 Related Work

    2.1 Image Deblurring

    Several studies have combined computer science and materials science,with a particular emphasis on utilizing image processing methods for analyzing the microstructure images of materials.Varde[10]proposed a computational estimation method called AutoDomainMine,based on graph data mining.By integrating clustering and classification techniques,this method discovered knowledge from existing experimental data and utilized it for estimation.The main objective of this framework was to estimate the graphical results of experiments based on input conditions.Similar graph data mining methods can be employed for image deblurring tasks to analyze and extract patterns and features from image data to achieve image deblurring goals.Pan et al.[11] reviewed the evolution and impact of material microstructures during cutting processes,presenting a thermal-force-microstructure coupled modeling framework.They analyzed microstructural changes such as white layer formation,phase transformation,and dynamic recrystallization under different materials and cutting conditions,as well as the effects of these changes on cutting forces and surface integrity.Vibration of cutting tools or materials can cause motion in image acquisition devices (such as cameras) during the capturing process,resulting in image blurring.Therefore,studying the deblurring of material microstructures holds significant importance.

    Many traditional image enhancement methods employ regularization and manually crafted prior images for blur kernel estimation[12].Subsequent iterative optimization is used to gradually recover a clear image.However,this conventional approach involves intricate blur kernel estimation,leading to laborious sharpening,subpar real-time performance,and algorithmic limitations.To enhance the quality of image deblurring,many methods based on convolutional neural networks (CNN) have been proposed [13–15].Chakrabarti [13] designed a neural network to generate a global blur kernel for non-blind deconvolution.Song et al.[14]proposed a method using a neural network for reliable detection of motion blur kernels to detect image forgeries.Wang et al.[15]proposed a network-based framework that learned to remove raindrops by learning motion blur kernels.Sun et al.[16]predicted the probability distribution of non-uniform motion blur using CNNs.However,most neural-networkbased methods still rely on blur models to solve the blur kernel,limiting their performance.

    In recent years,with the development of deep learning,a series of methods based on deep learning have been used for image deblurring [17,18].Zhang et al.[19] proposed DMPHN,which is the first multi-scale network based on the multi-patch method for single-image deblurring.Chen et al.[20]proposed HINet,a deep image restoration network based on the HIN block.Fanous et al.[5]presented GANscan,a method for restoring sharp images from motion-blurred videos.The method was applied to reconstruct tissue sections under the microscope.Liang et al.[21]directly deblurred raw images using deep learning-based busy image-to-image blind-deblurring.DID-ANet[4]was designed specifically for single-image blur removal caused by camera misfocus.MedDeblur[18]was developed to remove blur in medical images due to patient movement or breathing.Xu et al.[22] proposed a deep-learning-based knowledge-enhanced image deblurring method for quality inspection in yarn production.Restormer[23]is an efficient transformer model that can be utilized for image restoration tasks at high resolutions.This model is effective for restoring high-resolution images.Chen et al.[7]found that nonlinear activation functions are not necessary and can be replaced or omitted,and developed NAFNet for both image denoising and deblurring.Due to the impressive performance of NAFNet in deblurring tasks,we are currently implementing modifications to its architecture.

    2.2 Attention-Based Deblurring Model

    In recent years,attention mechanisms have proven to be highly effective in various computer vision tasks[24,25].As a result,attention-based methods have gradually been adopted for the task of image deblurring[26,27].MSAN[28]is a convolutional neural network architecture based on attention that efficiently and effectively generalizes motion deblurring.D3-Net[26]can be used for deblurring,dehazing,and object detection,with the addition of a classification attention feature loss to improve deblurring and dehazing performance.Cui et al.[27] proposed a dual-domain attention mechanism that enhances feature expression in both spatial and frequency domains.Ma et al.[29] proposed an attention-based dehazing algorithm for deblurring to improve defect detection in inspection image pipelines.Shen et al.[30] introduced a supervised human-perception attention mechanism model,which performs exceptionally well in motion deblurring in an end-to-end manner.MALNET[31]is a lightweight network based on attention mechanisms,which also performs well in image deblurring.Zhang et al.[32]proposed an attention-based inter-frame compensation scheme for video deblurring.In this work,we also incorporate attention mechanisms into our image deblurring network to improve its deblurring capability.

    3 Method

    3.1 Architecture

    The network structure of this paper is shown in Fig.1.It follows a classical U-shape structure,which is an improvement from NAFNet[7].The structure comprises an encoder and a decoder,both belonging to the MID-Block.An attention mechanism is introduced between the blocks to improve the image restoration quality of the network.

    3.2 MID-Block

    MID-Block is the basic building block of MIDNet.To avoid high complexity between blocks,MID-Block does not use any nonlinear activation functions such as ReLU,GELU,and Softmax.We construct a MID-Block using analogies with NAFNet blocks,as illustrated in Fig.2.

    To stabilize the training process,the input is first passed through Layer Normalization.Next,the input undergoes convolution operations and is then processed by SimpleGate(SG)[7],which is a variant of Gated Linear Units(GLU)[33].The GLU formula is as follows:

    Figure 1: The MIDNet overview.The overall architecture of the network resembles a U-shape design,which is composed of MID-block and attention block

    Figure 2: Architecture of MID-block

    In Eq.(1),Xrepresents the feature map,fandgfunction as linear transformers,σrepresents a nonlinear activation function,such as Sigmoid,and ⊙represents element-wise multiplication.

    The GLU increases the intra-block complexity,which is not desirable.To remedy this issue,we reconsider the activation function in the block,specifically GELU[34],which is expressed as:

    whereφrepresents the cumulative distribution function of the standard normal distribution.According to reference[34],it is suggested that the Gaussian Error Linear Unit(GELU)activation function can be effectively approximated and implemented by employing the following methodology:

    In Eqs.(1)and(2),GELU is a specific case of GLU,where the activation functionsfandgare identity functions and the parameterσis substituted withφ.The GLU incorporates nonlinearity and is not reliant on the parameterσ.Even in the absence of the parameterσ,the expressionGate(X)=f(X) ⊙g(X)retains its nonlinearity.According to reference[7],we suggest a simple adjustment to GLU:Split the feature map into two parts along the channel dimension and multiply them.This could be done using a basic element-wise multiplication,which is represented by Eq.(4).

    In Eq.(4),XandYrepresent feature maps of equal proportions.

    The gating unit SG is a neural network component illustrated in Fig.3,which is used in the processing of feature maps.It operates by splitting the feature map into two parts along the channel dimension,which is then multiplied to generate the final output.By splitting the feature map in this manner,SG can selectively emphasize or de-emphasize specific channels in the feature map,which can be useful for enhancing certain features or suppressing noise in the signal.This process is often referred to as channel-wise gating.

    Figure 3: Simple gate as represented by Eq.(2).⊙:Element-wise

    Our novel approach introduces Simplified Channel Attention(SCA)[7],a new component that utilizes channel-wise attention to enhance relevant features in data.Compared to other approaches,SCA has a simpler structure which offers ease of implementation.Additionally,it adds minimal computational overhead to models,hence enhancing the efficiency of our approach.Please refer to Fig.4 for an illustration of SCA.

    Figure 4: Simplified channel attention(SCA).?:Channel-wise multiplicatio n

    SCA determines channel attention by computing the average of the feature map along the spatial dimensions and applying a fully connected layer to generate a channel-wise attention vector.This attention vector is then multiplied with the original feature map to selectively amplify important channels in the data while suppressing irrelevant or noisy channels.

    Our experiments demonstrate that incorporating SCA into a standard convolutional neural network yields improved performance,highlighting the efficacy of enhancing feature representation using channel attention.SCA can be easily integrated into existing neural network architectures and represents a useful tool for improving the performance of deep learning models in a variety of applications.

    SCA is derived from Channel Attention(CA)[35],which can be expressed by the Eq.(5).

    In Eq.(5),Xdenotes the feature map,pooldenotes the global average pooling operation,σdenotes an activation function such as Sigmoid,W1,andW2denote fully connected layers,and cross multiplication is the channel multiplication operation.By simplifying the Eq.(5),we can finally obtain SCA,as shown in Eq.(6).

    3.3 Attention Mechanism

    With the advancement of deep learning techniques,significant progress has been made in image restoration.The NAFNet model,in particular,has shown significant performance in various applications.However,a limitation of NAFNet is that the skip connections used for feature aggregation between the encoder and decoder have the potential to disrupt the feature distribution,resulting in inconsistencies between these components.Another shortcoming of NAFNet is that it only employs an intra-block attention mechanism and ignores attention-based skip connections.

    To address these challenges,we introduce a soft attention mechanism to capture the latent relationship between the encoder and decoder more adaptively.We refer to the proposed soft attention mechanism as ATT.The architecture of the attention gate ATT is shown in Fig.5.Specifically,the proposed attention gate ATT aggregates features from different blocks using a weighting scheme based on their relevance to the current image restoration task,instead of simple element-wise addition used in conventional skip connections.This allows the model to selectively focus on the most informative features while suppressing the irrelevant ones.

    Figure 5: The architecture of ATT

    Moreover,our attention mechanism enables us to incorporate attention-based skip connections,which further enhance the feature aggregation process.By attending to the most informative features in the encoder and decoder,the model can effectively alleviate issues related to feature distribution inconsistency and strengthen the correlation between these components.The formula of the soft attention mechanism can be expressed as follows:

    In Eqs.(7)and(8),σ1andσ2denote activation functions.The attention gate is represented by a set of parameters through mathematical formulas,including linear transformationsWa,Wb,ψ,and biasesbf,bψ.The linear transformations are obtained by performing convolution operations on the input tensors.The output of the attention gate is the product of the input feature map and the attention coefficient.

    3.4 Multi-Loss Function

    The paper utilizes multi-loss functions,as shown in Eq.(9),which comprise the deblurring loss,edge loss,and FFT loss.The hyperparametersλ1 andλ2 are assigned the values of 0.05 and 0.01,respectively.

    3.4.1 Deblurring Function

    The deblurred image is compared with its ground truth in the spatial domain,using the standardl1loss as shown in Eq.(10).We do not usel2loss because it sometimes over-penalizes errors and leads to poor deblurring performance.

    3.4.2 Edge Function

    To restore the high-frequency details of the image,we introduce an edge loss function.It aims to focus on the gradient information of the image and enhance the edge texture features.The edge loss function of this paper is as follows:

    In Eq.(11),Irrepresents the reconstructed image,Igtrepresents the clear ground truth image andΔdenotes the Laplacian operator.

    3.4.3 FFT Loss

    The FFT loss is a type of loss function based on the Fourier transform that is used for image restoration tasks.It aims to penalize the discrepancy between the reconstructed image and the ground truth image in the frequency domain.The FFT loss is represented as follows:

    In Eq.(12),the variablesWandHrefer to the width and height of the image being analyzed.The functionFrepresents the Fourier transform of the image,which is a mathematical technique used to analyze its frequency components.Wherewi,jrepresents the weight corresponding to the Fourier coefficient,andHrepresents the frequency response of the degradation function in the Fourier domain.

    Specifically,the FFT loss can be calculated as the weighted sum of the squared Euclidean distance between the discrete Fourier transform coefficients of the reconstructed image and the ground truth image.The weight factors,which correspond to different Fourier coefficients,are used to emphasize the importance of different frequencies in the loss function,allowing it to focus more on the crucial parts of the reconstructed image spectrum.In the Fourier domain,high-frequency information such as edges and textures has a more significant impact on the visual quality of the reconstructed image.Therefore,incorporating the FFT loss can help the network better preserve these details,ultimately leading to an improvement in the image quality.

    4 Experiments

    4.1 Dataset

    We utilize a dataset containing 120 paired images with both low and high quality to investigate material microstructure fuzziness.Specifically,low-quality images in this dataset are directly obtained from observations captured through the SEM rather than artificially blurred using blur kernels or algorithms.This approach replicates real-world scenarios more accurately while simultaneously presenting greater challenges for the process of deblurring.When low-quality images are captured in practice,operators take repeated images until high-quality ones are achieved.Consequently,we meticulously selected 120 matching low and high-quality images that met stringent criteria.All images are subsequently adjusted to 256 ?256 pixels.Several cropped images are displayed in Fig.6.The dataset is randomly divided into a training set comprising 108 image pairs and a test set containing 12 image pairs.

    4.2 Experiment Parameters

    We optimize the model using Adam (β1=0.9,β2=0.999) for 200 K iterations with a cosine annealing schedule that decreases the learning rate from 10–3 to 10–7.We crop the images to a size of 256 ?256 pixels and apply rotation and flipping as data augmentation techniques.We employed the skip-init method to ensure stable training and implemented our code in the PyTorch framework.We evaluate our model using peak signal-to-noise ratio(PSNR)and structural similarity(SSIM)metrics.All experiments are conducted on an NVIDIA Tesla V100 GPU.

    4.3 Experiments on SOTA Algorithms

    PSNR and SSIM are employed as quantitative evaluation metrics,with larger values indicating superior image quality.They are calculated according to Eqs.(13)and(14).

    Figure 6: A few sample images from our dataset.Column 1 shows the low-quality images,whereas Column 2 shows the high-quality images

    In Eq.(13),MAX represents the maximum pixel value of the image,typically 255 when each pixel is represented by an 8-bit binary.MSE(Mean Squared Error)is the mean squared error value between the blurred image and the clear image.In Eq.(14),xandydenote the original image and the deblurred image,respectively.μxandμyrepresent the mean pixel values of imagesxandy,σxandσyrepresent the standard deviations of pixel values in imagesxandy,andσxyis the covariance between the pixel values of the two images.C1andC2are constants introduced to prevent division by zero in the denominator.

    To assess the generalizability of models trained on natural images to material microstructure fuzziness data,we conduct a series of relevant studies.Specifically,we employ pre-trained weights from the original papers of DMPHN,HINet,Restormer,and NAFNet methods to conduct inference on material blurry images.The deblurred images are displayed in Fig.7,while the corresponding PSNR and SSIM values are summarized in Table 1.

    Table 1: Results of image deblurring by using pre-trained weights

    As observed in Fig.7,these methods exhibit certain levels of processing applied to the blurry images.However,their ability to achieve satisfactory deblurring outcomes remains limited,with minimal improvement over the initial blurry images.By referring to Table 1,the PSNR and SSIM values of both the original blurry and clear images are provided in the input row.Notably,these methods yield relatively low PSNR and SSIM scores,with instances where deblurred images demonstrate worse performance compared to their initial states.

    Figure 7: Image deblurring performance on the material blurry dataset is evaluated using several SOTA algorithms with pre-trained weights

    Interestingly,these methods have demonstrated proficiency on the GoPro dataset and have exhibited effective deblurring outcomes on real-world blurry images.Consequently,we postulate that their subpar performance on material images may be attributed to external factors rather than the inherent limitations of the methods themselves.

    Upon meticulous scrutiny of the GoPro dataset,a notable distinction emerges in the PSNR values of its blurry images,which average approximately 23.In contrast,the blurry images originating from our material microstructure exhibit a lower PSNR value of approximately 21.Building upon these observations,a hypothesis arises: The relatively lower quality of material images,resulting in reduced information content,poses a heightened challenge for the deblurring process.Consequently,this challenge could potentially contribute to network degradation and the suboptimal performance observed.

    Furthermore,an additional factor potentially influencing the subpar deblurring results is the unique visual characteristics inherent to material microstructures,setting them apart from real-world blurry images.This disparity in appearance might contribute to reduced reliability in the neural network’s performance when confronted with material microstructure fuzziness data.To address this challenge,we advocate for a proactive solution:Retraining and fine-tuning these methods using material blurry images.Our approach involves freezing the majority of the model layers and selectively unfreezing a small subset for training purposes.We apply data augmentation techniques,such as flip and rotate,to the dataset during the training process.Hyperparameters,including learning rate,batch size,and number of iterations,are adjusted based on the specific model to achieve optimal performance.Additionally,appropriate regularization strategies are employed to mitigate overfitting problems.Such an approach holds the promise of enhancing the network’s capability to effectively restore blurry images of materials.In line with this recommendation,we embarked on the process of retraining and fine-tuning these methods.To gauge the efficacy of this intervention,we present the deblurring outcomes in Fig.8.

    Figure 8: The outcomes of deblurring upon the retraining and fine-tuning of these methods with our blurry dataset

    This study utilizes a dataset of material blurry images to conduct a detailed analysis of the deblurring capability of the original method compared to the retraining and fine-tuning methods.The outcomes of this comparison reveal a significant enhancement in deblurring quality for material images through retraining and fine-tuning,surpassing the performance of the no-training scenario and yielding satisfactory results.Notably,the process of retraining and fine-tuning contributes to the restoration of intricate features within material images,underscoring the pivotal role of materialspecific data in optimizing deblurring effectiveness.These findings offer fresh insights into the efficacy of retraining and fine-tuning strategies in effectively addressing the intricate deblurring challenges posed by material images.Furthermore,they provide valuable guidance for the future development of more potent deblurring methodologies within the domain of material science and engineering.Importantly,this study also serves as a demonstration of the potential of deep learning techniques in enhancing the quality of visual data across a wide spectrum of scientific and industrial applications.

    4.4 Comparative Experiment

    4.4.1 Qualitative Results

    We undertake a comparative evaluation of MIDNet alongside several SOTA deblurring methods that have undergone retraining and fine-tuning,as discussed in the previous section.The deblurring outcomes produced by each of these methods are depicted in Fig.9.Within this array of tested approaches,Restormer’s results exhibit a residual blurriness accompanied by unclear edges,which implies a limited restorative impact.The HINet method,employing a patch-based testing strategy,manifests noticeable stripe artifacts,possibly attributed to boundary discontinuities.The DMPHN approach,although improved,still retains a degree of blurriness that hampers its ability to achieve significant image enhancement.The NAFNet method,while competent,sacrifices certain fine image details.In stark contrast,our proposed MIDNet method achieves a further elevation in image quality,facilitating the restoration of additional structural details without introducing any artifacts or related issues.By observing the image,we note that our method exhibits significantly clearer microstructural contours compared to other approaches,as indicated by the red arrow in Fig.9.This enhanced clarity allows for a more accurate analysis of the material’s surface morphology and structural features based on these finer details.

    Figure 9: Qualitative comparison of image deblurring methods on the dataset

    The comparison between the original image and the deblurred image obtained through the model proposed in this study is illustrated in Fig.10.In Fig.10a,we present the original image,while Fig.10b depicts the image after being processed by the model.Through visual observation,it is evident that the proposed model exhibits excellent deblurring performance.The outcomes of our study highlight the exceptional capabilities of MIDNet in effectively recovering intricate structures and details within material images.This showcases its potential as a promising solution for tackling intricate deblurring issues within the realm of materials science and engineering.

    4.4.2 Quantitative Results

    Table 2 outlines the quantitative findings of several deblurring techniques applied to material microstructure images.Our evaluation of image quality relies on two objective metrics: PSNR and SSIM,where higher values denote enhanced performance.Significant enhancements in PSNR are observed across HINet,Restormer,DMPHN,and NAFNet after the process of retraining and fine-tuning.The respective gains in PSNR are 7.89,9.43,10.13,and 13.53 dB.These compelling outcomes underscore the considerable potential of deep learning in addressing the intricate challenges associated with deblurring material microstructure images.This progress lays the foundation for practical applications within this domain.

    Table 2: Quantitative comparison of our proposed network with previous methods

    The insights provided by Table 2 highlight the substantial advancement brought forth by MIDNet,when compared with NAFNet,evaluated through both PSNR and SSIM metrics.Compared to NAFNet,MIDNet achieved an improvement of 1.45 dB in PSNR and 0.01 in SSIM.This indicates that our proposed method has an advantage in image deblurring.The efficacy of MIDNet in the deblurring task can be attributed to its integrative employment of an attention mechanism and a combination of diverse loss functions.

    Figure 10: Comparison between original images and deblurred images

    The attention mechanism significantly enhances the network’s ability to focus on pivotal features,leading to elevated deblurring performance.Our experiment results affirm that the simultaneous utilization of multiple loss functions empowers the network with enhanced image reconstruction supervision,consequently elevating image quality and augmenting fine detail preservation.

    4.4.3 Ablation Experiment

    To validate the efficacy of the newly introduced edge loss and FFT loss within the training process,we conduct ablation experiments.The outcomes of these experiments are meticulously presented in Table 3,showcasing the computed PSNR and SSIM values corresponding to each experimental configuration.The objective behind these ablation studies is to discern the impact and contribution of individual loss functions toward the process of image restoration.To achieve this,we train our model under different scenarios,each characterized by a distinct combination of loss functions.This systematic approach enables us to gain insights into the relative importance and effectiveness of each loss function in driving the enhancement of image quality.

    Table 3: Ablation experiments:We train our model using different combinations of loss functions to understand the importance of individual losses for image restoration

    In this study,we undertake a series of ablation experiments with the intent of examining the impact of integrating various loss functions during the training phase.To maintain consistency,the Ldloss function,which plays a pivotal role in image restoration,is kept constant across all experiments.The outcomes of these ablation studies are summarized in Table 3.We observe that the inclusion of the Leloss function results in noticeable improvements in both PSNR and SSIM metrics.This suggests that the network effectively retains more intricate edge details through the utilization of this loss function.Furthermore,the inclusion of Lfloss further improves the image quality by providing more structural guidance to the network solution,as observed in row 3.It is worth noting that by combining all the loss functions during training,the network achieved its best performance.These findings highlight the importance of the proposed multi-loss functions in enhancing image restoration capabilities and offer valuable insights for the advancement of effective image restoration methods.

    5 Conclusions and Future Work

    In this study,we propose a method named MIDNet to address the issue of blurry images in material microstructures.MIDNet is an end-to-end deblurring network that enhances the clarity of blurry images in material microstructures by incorporating an attention mechanism and introducing multiple loss functions.Thorough qualitative and quantitative analysis indicates that MIDNet surpasses other approaches in terms of the quality of reconstructed images,marked by enhanced clarity and texture richness.Ablation experiments have also showcased the effectiveness of different loss functions within the network.Our work has the potential to encourage the extended use of deep learning within materials science and promote advancements in the mutually beneficial partnership between computer science and materials science.

    The dataset utilized in this study comprises actual experimental material microstructural images.However,we acknowledge that the dataset size is relatively limited,which may potentially impact the accuracy of image deblurring when extrapolating our method to diverse materials.To address this limitation,our future research will emphasize the collection of SEM images encompassing a broader range of alloy materials,thereby expanding the dataset size.Through these endeavors,we aim to enhance the performance and adaptability of our model in the context of deblurring microstructural images across various materials.Our future work will be primarily focused on developing a video deblurring method that is specifically tailored to the demands of material science applications.Given the unique challenges posed by the complex and dynamic nature of material structures,a robust and effective video deblurring method would be of great value in enabling researchers to visualize and analyze material properties more accurately.

    Acknowledgement:The authors especially acknowledge Prof.Liwu Jiang of National Center for Materials Service Safety.

    Funding Statement:The current work was supported by the National Key R&D Program of China(Grant No.2021YFA1601104),National Key R&D Program of China(Grant No.2022YFA16038004),National Key R&D Program of China (Grant No.2022YFA16038002) and National Science and Technology Major Project of China(No.J2019-VI-0004-0117).

    Author Contributions:Study conception and design:J.X.Wang,H.Y.Yu and D.B.Sun;data collection:J.X.Wang,Z.Y.Li and P Shi;analysis and interpretation of result: J.X.Wang,P Shi.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:Some dataset for the experiments uploaded to the author’s github repository:https://github.com/woshigui/MIDNet.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    有码 亚洲区| 久久人人爽人人爽人人片va| 久久99热这里只有精品18| xxx大片免费视频| 丰满乱子伦码专区| 偷拍熟女少妇极品色| 精品少妇久久久久久888优播| 男男h啪啪无遮挡| 欧美成人一区二区免费高清观看| 亚洲欧美清纯卡通| 亚洲精品第二区| 如何舔出高潮| 国产免费又黄又爽又色| 99re6热这里在线精品视频| 三级男女做爰猛烈吃奶摸视频| 22中文网久久字幕| 少妇丰满av| 国产亚洲91精品色在线| 免费观看性生交大片5| 性色avwww在线观看| 男人爽女人下面视频在线观看| 又爽又黄无遮挡网站| 涩涩av久久男人的天堂| 国产在线男女| 亚洲经典国产精华液单| 国产午夜福利久久久久久| 街头女战士在线观看网站| 国产成人免费观看mmmm| 舔av片在线| 王馨瑶露胸无遮挡在线观看| 久久精品夜色国产| 狠狠精品人妻久久久久久综合| 男人添女人高潮全过程视频| 伊人久久精品亚洲午夜| 永久网站在线| 哪个播放器可以免费观看大片| 日韩欧美一区视频在线观看 | 国产亚洲最大av| 久久韩国三级中文字幕| 简卡轻食公司| 在线亚洲精品国产二区图片欧美 | 日韩av在线免费看完整版不卡| 最新中文字幕久久久久| 在线观看三级黄色| 精品少妇久久久久久888优播| 热99国产精品久久久久久7| 成人二区视频| 夜夜爽夜夜爽视频| 熟女av电影| 看十八女毛片水多多多| 中文天堂在线官网| 精品久久久久久电影网| 欧美少妇被猛烈插入视频| 一个人观看的视频www高清免费观看| 毛片女人毛片| 亚洲av不卡在线观看| 蜜桃久久精品国产亚洲av| 免费av观看视频| 国产精品久久久久久久电影| 街头女战士在线观看网站| 欧美精品国产亚洲| 日日摸夜夜添夜夜爱| 男人爽女人下面视频在线观看| 婷婷色av中文字幕| 亚洲不卡免费看| 精品一区二区三卡| 白带黄色成豆腐渣| 九九在线视频观看精品| 国产女主播在线喷水免费视频网站| 成人黄色视频免费在线看| 中国三级夫妇交换| 啦啦啦中文免费视频观看日本| 成年人午夜在线观看视频| 人妻系列 视频| 在线亚洲精品国产二区图片欧美 | 99久久人妻综合| 久久人人爽人人片av| 日本三级黄在线观看| 国产淫语在线视频| 亚洲国产日韩一区二区| 国产高潮美女av| 视频区图区小说| 搡女人真爽免费视频火全软件| 国产成人免费观看mmmm| 日韩av免费高清视频| 精品一区在线观看国产| 亚洲婷婷狠狠爱综合网| 国产视频内射| 男人添女人高潮全过程视频| 特级一级黄色大片| 亚洲国产日韩一区二区| 国产毛片a区久久久久| 日韩,欧美,国产一区二区三区| 欧美区成人在线视频| 日产精品乱码卡一卡2卡三| 亚洲国产精品专区欧美| 一级二级三级毛片免费看| 亚洲av免费高清在线观看| 美女视频免费永久观看网站| 夜夜爽夜夜爽视频| 亚洲色图av天堂| 神马国产精品三级电影在线观看| 色吧在线观看| 久久6这里有精品| 国产精品一区www在线观看| 直男gayav资源| 日韩国内少妇激情av| 亚洲精品亚洲一区二区| 亚洲av男天堂| 亚洲人与动物交配视频| 五月开心婷婷网| 国产成人91sexporn| 99久久精品国产国产毛片| av专区在线播放| 男人舔奶头视频| 午夜免费鲁丝| 国产成人a区在线观看| 边亲边吃奶的免费视频| 国产精品爽爽va在线观看网站| 天堂俺去俺来也www色官网| 日韩,欧美,国产一区二区三区| 国产精品秋霞免费鲁丝片| 亚洲综合精品二区| 夜夜爽夜夜爽视频| 精品久久久久久久久亚洲| 大片免费播放器 马上看| 国产日韩欧美亚洲二区| 亚洲av日韩在线播放| 亚洲国产欧美人成| 亚洲人成网站在线播| 国产av码专区亚洲av| 大片电影免费在线观看免费| 晚上一个人看的免费电影| 国产高清不卡午夜福利| 搡老乐熟女国产| 嘟嘟电影网在线观看| 国产高清不卡午夜福利| 日本猛色少妇xxxxx猛交久久| 免费在线观看成人毛片| 99精国产麻豆久久婷婷| 男女边吃奶边做爰视频| 青春草国产在线视频| 久久鲁丝午夜福利片| 亚洲伊人久久精品综合| 午夜激情福利司机影院| 久久久久久久久久人人人人人人| 国产精品爽爽va在线观看网站| 国产v大片淫在线免费观看| 亚洲av二区三区四区| 丰满人妻一区二区三区视频av| 国产人妻一区二区三区在| 亚洲无线观看免费| 国产欧美另类精品又又久久亚洲欧美| 蜜桃亚洲精品一区二区三区| 欧美国产精品一级二级三级 | 国精品久久久久久国模美| 新久久久久国产一级毛片| 麻豆久久精品国产亚洲av| 搡老乐熟女国产| 精品99又大又爽又粗少妇毛片| 在线天堂最新版资源| 久久精品综合一区二区三区| 国产乱人偷精品视频| 精品人妻偷拍中文字幕| 亚洲av一区综合| 亚洲av在线观看美女高潮| 人妻夜夜爽99麻豆av| 欧美性感艳星| 久久久久久久大尺度免费视频| 国产成人freesex在线| 精品人妻视频免费看| 制服丝袜香蕉在线| 成人亚洲精品一区在线观看 | 国产成人aa在线观看| 成人美女网站在线观看视频| 美女脱内裤让男人舔精品视频| 少妇被粗大猛烈的视频| av国产精品久久久久影院| 日韩 亚洲 欧美在线| 午夜精品一区二区三区免费看| 国产av码专区亚洲av| 久久久成人免费电影| 毛片一级片免费看久久久久| 亚洲精品中文字幕在线视频 | 一级毛片电影观看| 国产中年淑女户外野战色| 99re6热这里在线精品视频| eeuss影院久久| 欧美激情在线99| 日本一本二区三区精品| 黄片wwwwww| 国内精品宾馆在线| 亚洲av二区三区四区| 久久久久久国产a免费观看| 欧美成人精品欧美一级黄| 欧美三级亚洲精品| av在线播放精品| 一级爰片在线观看| 亚洲成人精品中文字幕电影| 国产 精品1| 看免费成人av毛片| 高清av免费在线| 蜜桃亚洲精品一区二区三区| 久久久亚洲精品成人影院| 日日摸夜夜添夜夜爱| 联通29元200g的流量卡| 哪个播放器可以免费观看大片| 亚洲四区av| 伦精品一区二区三区| 高清日韩中文字幕在线| 特大巨黑吊av在线直播| 边亲边吃奶的免费视频| 亚洲婷婷狠狠爱综合网| 特级一级黄色大片| 91久久精品国产一区二区三区| 直男gayav资源| 神马国产精品三级电影在线观看| 五月天丁香电影| 十八禁网站网址无遮挡 | 亚洲欧美成人综合另类久久久| 老司机影院成人| 亚洲欧美日韩东京热| 日韩av免费高清视频| 亚洲人成网站在线播| 在线观看av片永久免费下载| 日韩免费高清中文字幕av| 高清视频免费观看一区二区| 国产成人福利小说| 久久久a久久爽久久v久久| 欧美日韩在线观看h| 国产精品嫩草影院av在线观看| 久久精品国产亚洲av涩爱| 久久人人爽人人爽人人片va| 久久久久久久国产电影| av播播在线观看一区| 国产亚洲av嫩草精品影院| 亚洲最大成人中文| 青春草亚洲视频在线观看| 国产综合懂色| 久久久久精品久久久久真实原创| 97精品久久久久久久久久精品| 人人妻人人澡人人爽人人夜夜| 男女国产视频网站| 成人综合一区亚洲| av免费在线看不卡| 99久久人妻综合| 2021天堂中文幕一二区在线观| 亚洲精品国产av成人精品| 一个人看视频在线观看www免费| 一边亲一边摸免费视频| 日本一本二区三区精品| 亚洲国产精品专区欧美| 狠狠精品人妻久久久久久综合| 亚洲激情五月婷婷啪啪| 最近手机中文字幕大全| 久久久久久久久久久免费av| 久久综合国产亚洲精品| 国产精品国产av在线观看| 久久久久久久久大av| 免费看av在线观看网站| 日产精品乱码卡一卡2卡三| 久久国内精品自在自线图片| 99精国产麻豆久久婷婷| av卡一久久| 中文资源天堂在线| av一本久久久久| 国产大屁股一区二区在线视频| 国产真实伦视频高清在线观看| 久久久久精品性色| 99视频精品全部免费 在线| 日韩成人av中文字幕在线观看| 一级毛片aaaaaa免费看小| 99久久精品热视频| 99热这里只有精品一区| 亚洲成人中文字幕在线播放| 热re99久久精品国产66热6| 国产女主播在线喷水免费视频网站| 18禁在线无遮挡免费观看视频| 亚洲无线观看免费| 中国国产av一级| 青春草视频在线免费观看| 九九久久精品国产亚洲av麻豆| 啦啦啦啦在线视频资源| 日产精品乱码卡一卡2卡三| 国产老妇伦熟女老妇高清| 欧美最新免费一区二区三区| 欧美一区二区亚洲| 亚洲无线观看免费| av在线天堂中文字幕| 欧美97在线视频| 免费av毛片视频| 蜜桃亚洲精品一区二区三区| 亚洲一区二区三区欧美精品 | 欧美日韩在线观看h| 亚洲欧美一区二区三区国产| 成人毛片60女人毛片免费| 精品久久久久久久末码| 99久久九九国产精品国产免费| 成人漫画全彩无遮挡| 午夜免费观看性视频| 欧美高清性xxxxhd video| 蜜臀久久99精品久久宅男| 亚洲欧美中文字幕日韩二区| av在线蜜桃| 插阴视频在线观看视频| 久久久久国产网址| 18禁在线播放成人免费| av在线亚洲专区| 男女边吃奶边做爰视频| 秋霞伦理黄片| 少妇的逼好多水| 国产精品麻豆人妻色哟哟久久| 搡女人真爽免费视频火全软件| 婷婷色av中文字幕| 成人黄色视频免费在线看| 99热全是精品| 欧美亚洲 丝袜 人妻 在线| 麻豆成人午夜福利视频| 欧美三级亚洲精品| 亚洲av免费高清在线观看| 精品国产一区二区三区久久久樱花 | 亚洲av国产av综合av卡| 亚洲人与动物交配视频| 亚洲精品乱码久久久久久按摩| 久久久久网色| 永久网站在线| 一区二区三区免费毛片| 小蜜桃在线观看免费完整版高清| 日韩精品有码人妻一区| 男人和女人高潮做爰伦理| 观看免费一级毛片| h日本视频在线播放| 大片电影免费在线观看免费| 在线免费观看不下载黄p国产| 人人妻人人爽人人添夜夜欢视频 | 日韩一区二区三区影片| 三级经典国产精品| 一边亲一边摸免费视频| 国产一级毛片在线| 永久网站在线| 午夜亚洲福利在线播放| 尤物成人国产欧美一区二区三区| 亚洲国产高清在线一区二区三| 亚洲最大成人av| 国产综合懂色| 国产午夜精品久久久久久一区二区三区| 好男人视频免费观看在线| 免费看不卡的av| 麻豆国产97在线/欧美| 欧美一级a爱片免费观看看| 97超视频在线观看视频| 黄色欧美视频在线观看| 一本—道久久a久久精品蜜桃钙片 精品乱码久久久久久99久播 | 2022亚洲国产成人精品| 国产亚洲一区二区精品| 亚洲天堂av无毛| 欧美丝袜亚洲另类| 国产精品成人在线| 嘟嘟电影网在线观看| 国产真实伦视频高清在线观看| 日本黄色片子视频| 亚洲精品日韩av片在线观看| 国产欧美日韩一区二区三区在线 | 男女那种视频在线观看| 中文精品一卡2卡3卡4更新| 国产成人精品一,二区| 最近最新中文字幕免费大全7| 极品少妇高潮喷水抽搐| 国产欧美另类精品又又久久亚洲欧美| 亚洲三级黄色毛片| 欧美少妇被猛烈插入视频| a级一级毛片免费在线观看| 男女边吃奶边做爰视频| 亚洲欧美成人精品一区二区| 午夜亚洲福利在线播放| 五月开心婷婷网| 少妇猛男粗大的猛烈进出视频 | 五月天丁香电影| 我的老师免费观看完整版| 国产精品熟女久久久久浪| 永久网站在线| 在线 av 中文字幕| 男的添女的下面高潮视频| 乱系列少妇在线播放| 国产永久视频网站| 建设人人有责人人尽责人人享有的 | 国产精品久久久久久久电影| 在线 av 中文字幕| 亚洲精品视频女| 99热全是精品| 97人妻精品一区二区三区麻豆| 香蕉精品网在线| eeuss影院久久| 亚洲久久久久久中文字幕| 亚洲av成人精品一二三区| 精品亚洲乱码少妇综合久久| 一本色道久久久久久精品综合| 99热这里只有是精品50| 欧美国产精品一级二级三级 | 精品久久久噜噜| 中文精品一卡2卡3卡4更新| 男女无遮挡免费网站观看| 国产黄a三级三级三级人| 久热这里只有精品99| 亚洲最大成人av| 人人妻人人爽人人添夜夜欢视频 | 久久精品国产自在天天线| 亚洲美女搞黄在线观看| 日韩大片免费观看网站| 成人欧美大片| 欧美一区二区亚洲| 久久久久久久久久久免费av| 午夜福利视频1000在线观看| 男的添女的下面高潮视频| 亚洲欧美日韩另类电影网站 | 久久99热这里只有精品18| 国产综合懂色| 男女边吃奶边做爰视频| 久久久久久久久久成人| 国产成人免费无遮挡视频| 夫妻性生交免费视频一级片| 国产亚洲精品久久久com| 边亲边吃奶的免费视频| a级毛色黄片| 日日啪夜夜撸| 国产片特级美女逼逼视频| 成人亚洲欧美一区二区av| 色视频在线一区二区三区| 日日摸夜夜添夜夜添av毛片| av在线播放精品| 国产精品一区二区在线观看99| 一本久久精品| 你懂的网址亚洲精品在线观看| 嫩草影院入口| 国产毛片a区久久久久| 舔av片在线| 欧美一级a爱片免费观看看| av免费观看日本| 亚洲欧美日韩另类电影网站 | 在线播放无遮挡| av又黄又爽大尺度在线免费看| 亚洲av不卡在线观看| 国产精品福利在线免费观看| 亚洲国产色片| 国产精品久久久久久久电影| 国产成人a∨麻豆精品| 国产成人一区二区在线| 大片电影免费在线观看免费| videos熟女内射| 亚洲美女视频黄频| 极品教师在线视频| 国产精品国产三级专区第一集| 中文字幕久久专区| 亚洲图色成人| 亚洲精品日韩在线中文字幕| 日本欧美国产在线视频| 精品国产一区二区三区久久久樱花 | 99热这里只有精品一区| 男女边摸边吃奶| 国产美女午夜福利| 精品国产一区二区三区久久久樱花 | 人妻少妇偷人精品九色| 精品熟女少妇av免费看| 国产视频首页在线观看| 亚洲精品456在线播放app| 中文天堂在线官网| 久久午夜福利片| 成人美女网站在线观看视频| av免费观看日本| 久久精品夜色国产| 国产爱豆传媒在线观看| 老司机影院成人| av天堂中文字幕网| 国产成人免费无遮挡视频| 十八禁网站网址无遮挡 | 成人亚洲欧美一区二区av| 国产免费视频播放在线视频| 亚洲在久久综合| 18禁在线播放成人免费| 久久久欧美国产精品| 夫妻午夜视频| 成人综合一区亚洲| 日韩三级伦理在线观看| 国产亚洲av片在线观看秒播厂| 菩萨蛮人人尽说江南好唐韦庄| 精品99又大又爽又粗少妇毛片| 看非洲黑人一级黄片| 91精品国产九色| 亚洲自拍偷在线| 建设人人有责人人尽责人人享有的 | 老司机影院毛片| av在线天堂中文字幕| 欧美变态另类bdsm刘玥| 国产男人的电影天堂91| 丝袜喷水一区| av免费在线看不卡| 97在线人人人人妻| www.色视频.com| 高清毛片免费看| 亚洲av成人精品一二三区| 熟妇人妻不卡中文字幕| eeuss影院久久| 中文字幕制服av| 国产一区二区在线观看日韩| 女的被弄到高潮叫床怎么办| 亚洲熟女精品中文字幕| 亚洲精品456在线播放app| 久久国产乱子免费精品| 亚洲精品456在线播放app| 少妇人妻一区二区三区视频| 国产亚洲一区二区精品| 久久精品久久久久久噜噜老黄| 又黄又爽又刺激的免费视频.| 激情 狠狠 欧美| 夫妻性生交免费视频一级片| 久久久久国产精品人妻一区二区| 搡女人真爽免费视频火全软件| 成人一区二区视频在线观看| 视频中文字幕在线观看| 国产成人精品久久久久久| 成人鲁丝片一二三区免费| 日韩大片免费观看网站| 亚洲性久久影院| 亚洲综合精品二区| 国产精品成人在线| 又黄又爽又刺激的免费视频.| 成人漫画全彩无遮挡| 欧美区成人在线视频| 久久久久性生活片| 我的女老师完整版在线观看| 日韩成人伦理影院| 性色avwww在线观看| 国产亚洲5aaaaa淫片| 精品人妻一区二区三区麻豆| 国产精品精品国产色婷婷| 美女脱内裤让男人舔精品视频| 日韩一区二区三区影片| 99久久中文字幕三级久久日本| kizo精华| 国产在视频线精品| 亚洲自拍偷在线| 国产精品麻豆人妻色哟哟久久| 免费大片18禁| 国产成人精品久久久久久| 免费看日本二区| 啦啦啦在线观看免费高清www| 国产一区二区三区综合在线观看 | tube8黄色片| 少妇裸体淫交视频免费看高清| 精品熟女少妇av免费看| 蜜桃亚洲精品一区二区三区| 一边亲一边摸免费视频| 伦理电影大哥的女人| 熟女人妻精品中文字幕| 精品久久久久久电影网| 国产精品不卡视频一区二区| 日韩免费高清中文字幕av| 亚洲自拍偷在线| 黄色欧美视频在线观看| 亚洲aⅴ乱码一区二区在线播放| 国产高潮美女av| 国产爽快片一区二区三区| 欧美少妇被猛烈插入视频| 老司机影院毛片| 天堂俺去俺来也www色官网| 黑人高潮一二区| 成人高潮视频无遮挡免费网站| 人妻少妇偷人精品九色| 亚洲伊人久久精品综合| 国产毛片在线视频| 国产精品嫩草影院av在线观看| 国产一区二区在线观看日韩| 午夜视频国产福利| 亚洲综合精品二区| av免费观看日本| 久热这里只有精品99| 久久午夜福利片| 国产精品成人在线| 男女无遮挡免费网站观看| 少妇猛男粗大的猛烈进出视频 | 高清毛片免费看| 久久影院123| 寂寞人妻少妇视频99o| 精品99又大又爽又粗少妇毛片| 亚洲精品国产成人久久av| 高清午夜精品一区二区三区| 国产精品久久久久久久久免| 欧美国产精品一级二级三级 | www.av在线官网国产| 狂野欧美白嫩少妇大欣赏| 免费电影在线观看免费观看| 日韩国内少妇激情av| 国产精品99久久99久久久不卡 | 小蜜桃在线观看免费完整版高清| 性色avwww在线观看| 免费观看a级毛片全部| 午夜视频国产福利| 美女cb高潮喷水在线观看| 3wmmmm亚洲av在线观看| 欧美高清成人免费视频www| 免费观看的影片在线观看| 99热国产这里只有精品6| 免费少妇av软件| 亚洲最大成人av| 欧美激情久久久久久爽电影| 一区二区三区免费毛片| 日日啪夜夜撸| 夜夜爽夜夜爽视频| 亚洲精品影视一区二区三区av| 国产精品成人在线| 亚洲丝袜综合中文字幕| 好男人视频免费观看在线| 深夜a级毛片| 波多野结衣巨乳人妻| 亚洲最大成人手机在线|