• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Multi-Layer Deep Sparse Representation for Biological Slice Image Inpainting

    2023-10-26 13:15:30HaitaoHuHongmeiMaandShuliMei
    Computers Materials&Continua 2023年9期

    Haitao Hu ,Hongmei Ma and Shuli Mei,?

    1College of Information and Electrical Engineering,China Agricultural University,Beijing,100000,China

    2Yantai Research Institute,China Agricultural University,Yantai,264670,China

    ABSTRACT Biological slices are an effective tool for studying the physiological structure and evolution mechanism of biological systems.However,due to the complexity of preparation technology and the presence of many uncontrollable factors during the preparation processing,leads to problems such as difficulty in preparing slice images and breakage of slice images.Therefore,we proposed a biological slice image small-scale corruption inpainting algorithm with interpretability based on multi-layer deep sparse representation,achieving the high-fidelity reconstruction of slice images.We further discussed the relationship between deep convolutional neural networks and sparse representation,ensuring the high-fidelity characteristic of the algorithm first.A novel deep wavelet dictionary is proposed that can better obtain image prior and possess learnable feature.And multi-layer deep sparse representation is used to implement dictionary learning,acquiring better signal expression.Compared with methods such as NLABH,Shearlet,Partial Differential Equation(PDE),K-Singular Value Decomposition(K-SVD),Convolutional Sparse Coding,and Deep Image Prior,the proposed algorithm has better subjective reconstruction and objective evaluation with small-scale image data,which realized high-fidelity inpainting,under the condition of small-scale image data.And the O(n2)-level time complexity makes the proposed algorithm practical.The proposed algorithm can be effectively extended to other cross-sectional image inpainting problems,such as magnetic resonance images,and computed tomography images.

    KEYWORDS Deep sparse representation;image inpainting;convolutional sparse modelling;deep neural network

    1 Introduction

    Biological slice is a technique that uses frozen or paraffin slicing to obtain thin slices of biological tissue,which is an important approach to studying the interaction mechanism of biological tissue and the system.For example,mouse brain slices are an important model for studying the development of mouse neural networks,synapses,and brain area function[1].In addition,lust slice images can provide strong support for the reconstruction of lust 3D models[2].Biological slice images,as cross-sectional images,contain rich texture and contour information,which are important features for identifying objects in the images[3].However,during the slicing process,image degradation such as corruption may occur,which affects further research on biological mechanisms.Therefore,we need to perform inpainting on small-scale corruption in images.Besides,The uniqueness and individual differences of biological organisms make high-fidelity inpainting of biological slice images necessary to ensure the authenticity of the inpainting.

    The example-based method[4]uses the structure and redundancy of an image to search for the best-matched block in the image and fill in the missing areas.However,this method is only suitable for images with self-similarity and repetitive textures.In recent years,the use of deep learning to achieve image inpainting has gained some popularity.Pathak et al.[5]proposed the use of Deep Convolutional Neural Network(DCNN)to obtain high-dimensional information to guide image inpainting,showing some promising results.Generative deep learning methods (represented by Generative Adversarial Networks and denoising diffusion inpainting models[6])have demonstrated powerful abilities in the inpainting of large-scale corrupted images.However,deep learning relies on statistical information and only focuses on visual plausibility,and lacks interpretability.The reliability of the image reconstruction results cannot be guaranteed,casting doubts on the application of deep learning in biological slice image inpainting[7].

    Sparse representation(SR)[8,9]is a research direction in solving ill-posed problems.It has been extensively used in solving inverse problems like image denoising,restoration,and deblurring.Studying the interpretability problem of deep learning through sparse representation has become a feasible method[10–12].Usually,sparse representation solves the following constrained optimization problem:

    where D represents the dictionary,z represents the sparse coding,and λ represents the regularization parameter.

    The sparse coding reconstruction of an image is achieved by the linear combination of atoms in the dictionary [13].The dictionary serves as a set of basis vectors for signal representation and can be divided into fixed and learned dictionaries.Fixed dictionaries are no longer considered due to their inflexibility and rigidity.Learned dictionaries are initialized appropriately and obtained through training.The form of the initial dictionary affects the optimal results of dictionary learning.The initialization of the dictionary can be obtained by calculating Hilbert space basis functions,such as using Discrete Cosine Transform (DCT),Wavelet Transform,etc.[14].Some scholars use the convolution kernel function as the initial dictionary or construct convolutional sparse dictionaries by using convolutional methods[15].The lack of prior knowledge of the image,and the presence of noise and random initialization of the dictionary,make the learning of sparse coding difficult.

    The K-SVD algorithm[16]can obtain a set of overcomplete basis vectors,which can effectively represent signals.However,the time complexity of the optimization process is daunting.Although denoising algorithms such as BM3D[17]and DnCNN[18]have surpassed the denoising performance of the K-SVD algorithm,the K-SVD algorithm still has a wide range of applications.Can the K-SVD algorithm be revitalized in the era of deep learning? Scetbon,Elad,and other scholars proposed the Deep K-SVD algorithm[19],which re-interpreted the dictionary update strategy in the K-SVD algorithm as a differentiable form and constructed an interpretable end-to-end deep neural network model with a lightweight architecture that achieves a denoising performance approaching that of the DCNN.

    In high-reliability requirements for the task of restoring biological slice images,deep learning uses generative means,but the authenticity of the reconstruction results cannot be guaranteed.Using deep neural networks such as Autoencoder and restricted Boltzmann machines as deep sparse representation models [20],the interpretability is not clear.Deep sparse representation is a deep learning model based on sparse representation,which possesses a certain level of interpretability and learning capability.It enables better high-fidelity restoration and meets the authenticity requirements of biological slice images.Deep sparse representation still requires a large amount of data,and selecting hyperparameters for training poses certain difficulties.

    We summarize the contributions of this paper as follows: (1) We propose an end-to-end deep neural network model based on deep sparse representation to address the task of small-scale damage restoration in biological slice images.(2)We investigated the learnability of wavelet dictionaries and proposed a deep wavelet dictionary along with its updating algorithm.(3)We conducted an in-depth analysis of the relationship between sparse representation and deep neural networks,providing a comprehensive discussion on the interpretability of deep sparse representation.

    The paper is organized as follows.Section 2 introduces related work,Section 3 discusses the proposed algorithm,Section 4 reports experimental results and analysis,and Section 5 concludes the paper.

    2 Related Work

    2.1 Wavelet Transform

    Wavelet transform can capture local features of an image from multiple perspectives and achieve energy concentration.But,wavelet basis functions are manually designed,and fixed basis functions may not adapt well to signal families.The wavelet function can be obtained through the discretization of the parameters a and b of the continuous wavelet function.

    The response of the wavelet function to signalS(t)∈L2(R)is shown in Eq.(3).

    whereψ(t)is the wavelet function.

    In the L2(R)space,the one-dimensional discrete wavelet transform convolves and downsamples the signalf(t)using a high-pass filterψ(x)and a low-pass filterφ(x)to obtain approximation componentsdjand detail componentsaj[21].For an imageY∈L2(R2),the two-dimensional discrete wavelet filter is obtained by taking the tensor product of the one-dimensional wavelet filter,defined as follows[22]:

    In the equation,Φ(x,y)represents the scaling function,Ψ(x,y)represents the wavelet function.Fast Mallat algorithm is commonly used to efficiently compute the discrete wavelet transform,which uses the alternating convolution and down-sampling operations of the low-pass filterΦ(x,y)and high-pass filterΨ(x,y)to obtain the approximation and detail components of the image.

    Deep learning has shown great charm with its powerful fitting ability[23,24].This leads people to think about whether it is feasible to re-examine traditional methods,including wavelet transform,from the perspective of deep learning.Fortunately,following the train of thought of traditional methods and reconstructing algorithms from the perspective of deep learning has become a trend nowadays[25].Efforts have been made to improve the interpretability of DCNN,and it has been found that its learned results tend to approach wavelet transform or sparse representations[12].We consider building a trainable wavelet dictionary using learnable wavelet functions to obtain a better dictionary.

    2.2 Convolutional Sparse Modelling

    Convolutional Sparse Coding(CSC)is one approach to sparse representation in image processing,which is supported by strong theoretical foundations and has good biological plausibility.However,in recent years,the performance of CSC has been surpassed by deep learning.Building a“deep”CSC model has potential application values in various fields such as image restoration[15,26],image classification[27],and image registration[28].

    Sparsity has been integrated into the development of deep neural networks[29].CSC can also be used to construct deep sparse models by applying deep architecture,which transforms the optimization problem in Eq.(1)into the following form:

    since solving the?1norm is an NP-hard problem,it is usually related to the?1norm,which can be transformed into a convex optimization problem.

    Autoencoders,Restricted Boltzmann Machines,and other deep neural networks lack clear interpretability when solving Eq.(1).Sparse coding usually uses greedy algorithms or iterative thresholding algorithms,with the latter being able to approximate basis pursuit and implement sparsification in the form of network unfolding.Daubechies proposed Iterative shrinkage-thresholding algorithm(ISTA)[30]to approximate layered basis pursuit,updating the sparse coefficients through j rounds of iteration,with small values set to zero in each round while the rest remains almost unchanged.It is defined as follows:

    in the ISTA algorithm,Lis the step size with a value of the maximum eigenvalue ofDTD.λis the regularization parameter,and the soft thresholding operator is denoted asSθand is defined as:

    when the dictionaryDis enforced to be shared,the thresholding scheme can be approximately viewed as a“recurrent neural network”.Deep sparse representation is a deep learning model based on sparse representation.It has advantages such as stronger representation power and lower time complexity.However,it poses challenges in terms of selecting model hyperparameters and constructing datasets.

    3 Sparse Representation and Deep Neural Network

    3.1 Connection Between Sparse Representation and Deep Neural Networks

    Deep neural networks are developed based on the study of biological neural systems.The representation of signals in deep neural networks is non-linear,and the feature extraction is complex with a multi-scale network hierarchy.Deep convolutional neural networks(DCNN),as the representative of deep neural networks,use operations such as convolutional layers,linear layers,and pooling layers.The convolutional layer contains operations like convolution,activation,and bias.The linear layer performs linear transformations on the convolutional coefficients.The pooling layer implements the multi-resolution analysis of the neural network.

    The convolutional layer simulates the functions and structures of biological neurons.For input signalsy∈RM?N,the convolution operation is a group of convolutions applied toy,i.e.,x=φy,whereφrepresents a convolution matrix.The result is strengthened by the biasband then passed through an unsaturated non-linear activation function,such as the Rectified Linear Unit(ReLU),to achieve sparse activation.Therefore,neurons in DCNN can be defined as follows:

    in the equation,ReLU(α)=max(0,α).The multi-layer forward propagation mechanism in DCNN allows for multi-scale analysis.The parameter updates using gradient descent and backpropagation allow DCNN to provide feedback to the model based on the loss function.The transition from linear activation functions to saturated non-linear activation functions and to unsaturated non-linear activation functions enhances the non-linear characteristics of DCNN while increasing the sparsity.This indicates that neural networks are evolving towards sparser responses that resemble those of biological neural systems[30].

    The starting point for the construction of sparse representation theory is the sparse response of neurons in the visual cortex of the brain to visual signals.The update of sparse coefficients is generally achieved by using an iterative thresholding function to realize sparsification,such as the ISTA algorithm shown in Eqs.(6)and(7).It can be observed that the sparse coding algorithm in deep sparse representation theory and the structure of neurons(convolution operation)in neural networks have equivalence.Both apply a convolution followed by a non-linear transformation with a threshold to output an activation or inhibition state.

    As shown in Fig.1,the ReLU function and the ISTA algorithm both achieve sparseness and nonlinearity as a variant of the threshold function.The convolution operation applied to the convolutional coefficients in DCNN is equivalent to the convolution between dictionaryφand sparse coefficientsαin sparse representation theory.The biasbfunctionally equivalent to the thresholdθ,and the contraction operation both result in the sparsity of coefficients.Therefore,the functionality of neurons in the forward propagation of DCNN is equivalent to the sparse coding of sparse representation theory.

    Figure 1:Sparse regularization strategy

    The above discussion confines the problem to a single-layer model.The multiscale nature of DCNN is evident,formally defined as follows:

    By extending the basic deep sparse representation to multiple layers,we can obtain the Multi-layer Deep Sparse Representation(ML-DSR)model,formally defined below:

    this indicates that deep sparse representation also has a multi-scale analysis mechanism.

    DCNN and sparse representation share similarities in optimization and feedback mechanisms.In DCNN,backpropagation is applied in each iteration to update the network parameters with respect to the loss function,using the gradient descent method.Such optimization and feedback mechanisms ensure that the final solution moves towards minimizing the error.In sparse representation,the dictionary is updated based on the feedback from the reconstruction error,and the update direction is towards minimizing the loss.This optimization mechanism is evident in convolutional sparse models,improving the learnability of dictionaries and network parameters.

    3.2 High-Fidelity Reconstruction

    Starting from the basic form of deep sparse representationX=Dα,assumingX={x1,x2,...,xn-1,xn}Tandα={α1,α2,...,αn-1,αn}T,to achieve the reconstruction of the true valueX,we need to pursueargmin,where i=1···n,to obtain the optimal sparse coefficientα.Reference[10] provided the conditions for stable recovery of the original signal in the worst-case scenario.As follows:

    μ(D)represents the coherence of the dictionaryD,which is defined by the maximum inner product of different atoms in the dictionary.The solution obtained by applying the pursuit algorithm to this is stable and reliable,satisfying the condition.

    the symbolεrepresents bounded noise disturbance to the signal,which indicates that perfect signal reconstruction can be achieved when,thereby achieving high-fidelity reconstruction.However,the precise pursuit of the sparse coefficientαis very difficult,so we relax the constraint on,and achieve a high-fidelity representation of the image under the condition thatδ=<σ.

    Based on the Deep K-SVD algorithm,we constructed an artificial image dataset to validate its ability to effectively represent images through deep sparse representation,as shown in Fig.2.The root mean square error(RMSE)was used to measure the errorδof the reconstructed artificial images.Fig.3 shows the results of several artificial images reconstructed using deep sparse representation,which met our theoretical expectations in terms of visual effects and metrics analysis.The RMSE errors of the four test images reconstructed were 0.0185,0.0667,0.0326,and 0.0108,respectively,and the errors were within an acceptable range.This indicates that deep sparse representation can achieve the precise reconstruction of biological slice images,which establishes the basis for high-fidelity image inpainting.

    Figure 2:Artificial image datasets

    Figure 3:The reconstructed results of the test images.a1,b1,c1,and d1 are the original artificial images,while a2,b2,c2,and d2 are the reconstructed results

    4 Deep Sparse Representation Inpainting

    Our end-to-end biological slice image inpainting model is shown in Fig.4.First,we construct a deep wavelet dictionary (DWDict) for the imagey.Next,we decomposeyinto a collection of overlapping image blocks and generate corresponding sparse code.We then iteratively update the dictionary and sparse code in k layers.Finally,we obtain the reconstruction result by simple averaging.The error between the reconstruction result and the ground truth is fed back to the model to update the network’s parameters.Next,we will explain the details of dictionary construction and sparse coding.

    Figure 4:The algorithm architecture.The upper part represents the process of reconstructing damaged images,while the lower part shows the details of the encoding stage and dictionary update stage

    4.1 Deep Wavelet Dictionary

    Multiscale geometric analysis tools can achieve sparse representation of target images[31].Using wavelet transform to construct a wavelet dictionary can further exploit the sparsity performance of the algorithm,and the prior knowledge obtained from the images can enhance the performance of deep learning algorithms[32].However,wavelet transform basis functions rely on laborious designs,obtaining the optimal sparse representation of signal sets difficult.Therefore,we designed learnable wavelet basis functions and used the image prior knowledge to construct a Deep Wavelet Dictionary(DWDict),as shown in Fig.5.

    To meet the multiscale characteristics of wavelet decomposition,we first construct 2D discrete wavelet filters based on Eq.(5)and then construct a cascaded deep wavelet network as the convolutional kernel function to achieve convolutional operation and parameter learnability.Meanwhile,the down-sampling operation with a step size of 2 is performed to meet the requirements of each layer decomposition in the Fast Mallat algorithm.The form of the deep wavelet decomposition network is shown in Fig.6.LetFi(n)be the non-linear function of the convolutional kernel function,and for signalS(t),the response of the deep wavelet decomposition network is shown in Eq.(13):

    where ↓2 represents the down-sampling operator in the equation.

    The imageYis decomposed into a coefficient setα={αi|i=1,...,n} by the deep wavelet decomposition network,and we decompose the coefficient setαinto a normalized coefficient block collectionaccording to the step sizepand specifications.The coefficient block retains local information such as the edges of the image.

    The dictionary should cover the entire space of x in spaceRNas much as possible.Does the K-SVD algorithm tend to learn what kind of atoms and does it meet the above ideas?We constructed some artificial images and used the K-SVD algorithm to generate their dictionaries,as shown in Fig.7.

    Figure 5:Algorithm flow for the construction of the deep wavelet dictionary

    Figure 6:Deep wavelet decomposition network

    The K-SVD algorithm tends to favor learning the parts of the image with obvious features and selects representative parts from many feature areas as atoms.Based on this characteristic,we decided to prioritize image blocks with clear edge texture features in the construction of the deep wavelet dictionary.Therefore,we use the mean gradient as a measuring factor,which can characterize the degree of grayscale change,to select potential image blocks.The formula is defined as follows:

    in the formula,Prepresents the decomposed small blocks.To avoid the disturbance caused by the damaged areas on the dictionary construction,the image blocks in these areas will be excluded in advance.

    Figure 7:The results of K-SVD dictionary learning:(a)ring,(b)dictionary;(c)square,(d)dictionary

    4.2 Sparse Coding

    Our sparsification strategy is to use a learnable ISTA algorithm to achieve sparse coding.The parameters of the ISTA algorithm rely entirely on manual design,which makes it difficult to achieve optimal sparse representation,especially for ill-posed inverse problems.To address this issue,Gregor et al.[33] proposed the Learned-ISTA algorithm to learn the model’s parameters,but its adaptability is not strong.

    Due to the powerful representation capability of DCNN,the ISTA-Net algorithm[34]showed that nonlinear transformation functions can achieve sparse representation of images.TheF(·)operator constructed using nonlinear operators and ReLU linear units can achieve a richer expression of images.We use the learnability and nonlinearity of theF(·)operator to improve the sparsification performance of the sparse coding algorithm.Sparse coefficientαgets byα=F(α),and then as input to Eq.(15).

    in the equation,ρ=λ/Lis a learnable parameter and its value is obtained by the MLP module to satisfy an appropriate stopping criterion for each small block.

    ML-DSR shows that D orαcan be a component of a layer,or can be used as the target signal.Therefore,we use the dictionary D as the target signal and deploy a shallow deep neural network model based on ML-DSR,which is updated in each iteration as defined below:

    the dictionary also uses a deep differentiable method to optimize it during model feedback.

    4.3 General Structure

    The issue of small-scale corruption can be described asy=A⊙x,where ⊙represents the elementwise product andArepresents the limited operator.When addressing the damage caused by the limited operator,a new optimization objective function is obtained as shown in Eq.(19).

    In the Deep K-SVD algorithm,the?2norm is used to measure the error between the true value and the reconstructed result,i.e.,loss=‖x-f(y)‖2.However,the?2norm alone does not necessarily produce good sparse solutions and may cause artifacts,ignoring the visual perception demands of the human eye.Therefore,according to the suggestion in reference[33],a mixed form ofms_ssimand?1loss is used as the loss function.

    As it is a global reconstruction for images with small-scale damage,information loss in the nondamaged areas of the image is inevitable.Therefore,we adopt the complement operation on the reconstructed result X,which allows us to focus on the reconstruction of the damaged areas and obtain better reconstruction results.The complement operation is defined as follows:

    the symbol“~”represents a negation operation on the mask M.Finally,we can get the following algorithm flow.

    5 Experiment

    5.1 Data Sets and Preparatory Work

    We used the frozen slice method to prepare the biological slices and stained them with uranyl acetate and lead citrate.Then we obtained 90 to 100 micrographs of lust microscopic slices by photographing them with a microscope.Processing the slice image by image enhancement methods such as shading correction and cropping.The process of slice image acquisition is shown in Fig.8,The final obtained lust slice images contain rich texture information,complex edge contour structures,and segmental smoothness.In some tissue slices,there are also self-similar fractal structures[35],which reflect the complexity of biological structures.Sliced image inpainting has the following challenges:first,sliced images have complex texture contour information and thus need to be restored realistically at the time of inpainting,second,the low amount of sliced image data,and third,the need to avoid application difficulties due to high time complexity.

    Figure 8:Biological sliced image preparation process

    Due to the complexity of the slicing process and the small size of the data,we extract a specified number of images from the training set to effectively address the problem of insufficient data.Specifically,am?nsized image is cropped into(m-k+1)?(n-k+1)sub-images of sizek?k.A manually crafted set of around 200 small-scale masks with different shapes are used to initialize the position of masks on the sliced images during data loading,to improve the generalization performance of the model.

    Our model is built using the PyTorch framework and trained using the ADMM optimizer with a learning rate of 1e-4.The patch size is 8.The computer used for model training is equipped with an Intel(R) Xeon(R) Platinum 8255C CPU @ 2.50 GHz and NVIDIA GeForce RTX 3080.For model testing,both MATLAB R2021a and Python 3.7.4 environments were used on a computer with an Intel(R)Core(TM)i7-9750H CPU@2.60 GHz and NVIDIA GeForce GTX 1650.

    5.2 Experiment Result

    We design experiments from several aspects such as restoration effect,model complexity,and practicality,and use both simulated breakage and real breakage to show the restoration effect in the evaluation of inpainting effect.Fig.9 shows the training loss,which was completed in just over 3 h.During testing and analysis,we used grayscale images of size 512 ?512 pixels for evaluation,including subjective and objective evaluations,time complexity comparison,model size comparison,and other assessment experiments that will be discussed in the following.

    Figure 9:Training loss

    As shown in Fig.10,we constructed a DCT dictionary with 256 atoms and compared it to a deep wavelet dictionary with the same number of atoms.We found that the deep wavelet dictionary captures local feature information such as curved stripes,which are edge-like features,whereas the DCT dictionary mainly consists of striped and grid-like information.This indicates that the variety of atoms in the deep wavelet dictionary is more diverse,and the prior knowledge obtained from images can lead to better recovery performance of the algorithm.

    Figure 10:Dictionary construction algorithm,(a)is the DCT dictionary,(b)is the DWDict

    To evaluate the effectiveness of our proposed algorithm,we compared it with classical image inpainting algorithms and deep learning-based image inpainting algorithms,including NLABH[36],Shearlet [37],Mumford-Shah [38],K-SVD [16],Deep Image Prior (DIP) [32],and Local Block Coordinate Descent Algorithm(LoBCoD)[39].To assess the reconstruction quality of the algorithms,we used both subjective visual evaluation and objective evaluation metrics such as peak signal to noise ratio (PSNR),structural similarity index (SSIM),and root mean square error (RMSE),which can reflect the quality of image inpainting.

    Fig.11 shows the reconstruction results of different models after inpainting,and their evaluation metrics can be obtained from Table 1,which used simulated breakage masks.It can be observed that PDE-based inpainting methods such as Mumford-Shah algorithm exhibit strong competitiveness in small-scale damage inpainting by reasoning from the damaged edges towards the inside.LoBCoD algorithm,as a convolutional sparse coding model,uses the local block coordinate descent method to find the optimal basis vector and achieves good reconstruction performance.Compared with K-SVD algorithm,NLABH algorithm,our proposed algorithm achieves excellent subjective visual effects and optimal PSNR and RMSE scores,with SSIM score also approaching the optimal,reflecting the superior performance of our algorithm in reconstruction quality.

    Figure 11:Slice image inpainting experiment:An original image,(b)masked image,(c)Mumford-Shah,(d)Shearlet,(e)NLABH,(f)K-SVDm,(g)DIP,(h)LoBCoD,and(i)Proposed algorithm

    Table 1:Evaluation of image contrast inpainting effect

    The time complexity of the algorithm isO(n2),which the bottleneck lies in the construction of the deep wavelet dictionary.Table 2 provides the running time of each model.We aim to evaluate the application value of the algorithms through a rough analysis of their running time.On the other hand,our proposed algorithm only takes less than 13 s to achieve comparable inpainting results to the KSVD algorithm.Although the K-SVD algorithm also has good performance,its resource consumption is huge,reaching an intolerable time scale,making it difficult to apply in practical work.Table 2 and the reconstruction results discussed above demonstrate that our proposed algorithm not only achieves competitive inpainting results but also has practical application value.This is the goal we are pursuing.

    Table 2:Run time comparison of algorithms

    We conducted experimental comparisons on different forms of biological slice images to demonstrate the effectiveness of our proposed algorithm,as shown in Fig.12 and Table 3.The highlighted red boxes demonstrate the rationality of our proposed algorithm in the inpainting of details.It can be seen that the PSNR indicator of the image after reconstruction using the multi-scale geometric analysis tool as a“sparse”expression tool is not ideal.This is because the geometric analysis tool adopts an approximation method,which brings information loss in the decomposition and reconstruction process.Compared to algorithms such as K-SVD and DIP,the algorithm proposed in this paper achieves a more natural and continuous transition at the boundaries of damaged regions.The SSIM value of the proposed algorithm reaches 0.993,which is an optimal result compared with other algorithms.

    To further highlight the advantages of our proposed model in terms of small scale and low complexity,we used two indicators:the total number of parameters and the number of floating-point operations per second(FLOPS).The total number of parameters reflects the scale of the model,while FLOPS reflects the complexity of the model.As shown in Table 4,compared with the lightweight model of DIP,it can be seen that our proposed algorithm has a total number of parameters of only 1.789M,which is lower than the 2.15M of DIP.The GFLOPS indicator of our proposed model is 19.288G,much lower than the 1.930T of DIP.The significant difference in GFLOPS is that our proposed algorithm does not drastically increase in complexity with the scale of the problem.

    Figure 12:Slice image inpainting experiment:An original image,(b)masked image,(c)Mumford-Shah,(d)Shearlet,(e)NLABH,(f)K-SVDm,(g)DIP,(h)LoBCoD,and(i)Proposed algorithm

    Another issue worth discussing is whether it is possible to train the model with fewer data to cope with the difficulty of acquiring images.As shown in Table 5,we trained the model with 16,32,and 44 images,respectively,and evaluated the results.It can be seen that even with a small amount of data,the reconstructed results are acceptable,and the error level is kept below 0.004.This indicates that our proposed model has value in situations where data is difficult to obtain,and can adapt well to scenarios where it is hard to acquire biological image data.

    Table 3:Evaluation of image contrast inpainting effect

    Lastly,to demonstrate the practical application capability of the proposed algorithm in this paper,this chapter focuses on the application of the high-fidelity restoration algorithm to the problem of small-scale damages in biological slice images.As shown in Fig.13,the location marked by the red box in the left image indicates the damaged area,which is labeled with a mask.It can be observed that the algorithm presented in this chapter achieves excellent restoration results for this real small-scale damage.

    Figure 13:Realistic corruptions inpainting in slice images

    6 Conclusion

    This article focuses on the issue of small-scale corruption in biological slice image preparation,particularly in the case of limited data.We analyzed the relationship between sparse representation and deep neural networks and established a deep network model based on deep sparse representation.Our proposed model can effectively inpainting small-scale corruption in biological slice images while preserving the edge texture and contour structure of the slice images.We conducted tests using simulated damages,compared with other methods such as PDE-based methods,Shearlet algorithm,DIP algorithm,K-SVD algorithm,and sparse coding algorithm,our proposed model achieved good results in terms of effectiveness,time,and model scale.And then we demonstrated the application capability of our proposed method in addressing true corruptions in biological slice images that highfidelity inpainting has been achieved.

    Unlike algorithms such as K-SVD and LoBCOD,the algorithm proposed in this article not only achieves high-fidelity inpainting of biological slice images but also benefits from well-time complexity,making it valuable for practical applications.The proposed algorithm can also be effectively applied to other cross-sectional image restoration tasks,such as MRI and CT scan images.In future work,we will focus on explainable deep learning research based on deep sparse representation.Furthermore,the efficiency of the proposed model still needs further improvement,we will focus on dictionary learning with an emphasis on the learnable wavelet dictionary.

    Acknowledgement:The authors extend their appreciation to the anonymous reviewers for their constructive comments and suggestions.

    Funding Statement:This work was supported by the National Natural Science Foundation of China (Grant No.61871380),the Shandong Provincial Natural Science Foundation (Grant No.ZR2020MF019),and Beijing Natural Science Foundation(Grant No.4172034).

    Author Contributions:Study conception and design: Haitao Hu,Shuli Mei;data collection: Haitao Hu,Hongmei Ma;analysis and interpretation of results:Haitao Hu,Shuli Mei,Hongmei Ma;draft manuscript preparation: Haitao Hu,Shuli Mei,Hongmei Ma.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data that support the findings of this study are available from the corresponding author upon reasonable request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    国产成人午夜福利电影在线观看| 老女人水多毛片| 69人妻影院| 国产黄a三级三级三级人| 日韩中字成人| 亚洲精品自拍成人| 麻豆久久精品国产亚洲av| 欧美日韩综合久久久久久| 欧美激情久久久久久爽电影| 国产成人午夜福利电影在线观看| 少妇的逼水好多| 久久久精品免费免费高清| 国产大屁股一区二区在线视频| 午夜精品国产一区二区电影 | 国产不卡一卡二| 国产黄频视频在线观看| 天堂影院成人在线观看| 全区人妻精品视频| 深夜a级毛片| 一级毛片 在线播放| 久久草成人影院| 麻豆av噜噜一区二区三区| 51国产日韩欧美| 亚洲精品第二区| 国产美女午夜福利| 国产精品美女特级片免费视频播放器| 国产精品久久久久久av不卡| 在线观看免费高清a一片| 日韩强制内射视频| 一级a做视频免费观看| 综合色av麻豆| 免费看美女性在线毛片视频| 听说在线观看完整版免费高清| 男女国产视频网站| 九色成人免费人妻av| 日韩强制内射视频| 欧美性猛交╳xxx乱大交人| 日韩成人av中文字幕在线观看| 亚洲av一区综合| 亚洲精品久久午夜乱码| 日韩成人av中文字幕在线观看| 卡戴珊不雅视频在线播放| 成人无遮挡网站| 国产免费视频播放在线视频 | 99re6热这里在线精品视频| 久久精品国产亚洲av涩爱| 国产欧美日韩精品一区二区| 欧美一级a爱片免费观看看| 日韩在线高清观看一区二区三区| 久久久久久久久久久免费av| 大陆偷拍与自拍| 寂寞人妻少妇视频99o| 免费高清在线观看视频在线观看| 人妻制服诱惑在线中文字幕| 亚洲在久久综合| 99re6热这里在线精品视频| 人人妻人人澡人人爽人人夜夜 | 3wmmmm亚洲av在线观看| 精品久久久久久久久av| 日本av手机在线免费观看| 18禁裸乳无遮挡免费网站照片| 久久久久网色| 亚洲av不卡在线观看| 国产黄色视频一区二区在线观看| 又爽又黄无遮挡网站| 99久久人妻综合| 久久久精品免费免费高清| 精品国内亚洲2022精品成人| 中文字幕人妻熟人妻熟丝袜美| 秋霞伦理黄片| 久久久a久久爽久久v久久| 国产av国产精品国产| 亚洲精品国产成人久久av| 亚洲精品乱久久久久久| 精品人妻偷拍中文字幕| 九草在线视频观看| 赤兔流量卡办理| 亚洲,欧美,日韩| 日韩,欧美,国产一区二区三区| av在线亚洲专区| 女人久久www免费人成看片| 黄色配什么色好看| 亚洲色图av天堂| 一区二区三区高清视频在线| 午夜免费男女啪啪视频观看| 国语对白做爰xxxⅹ性视频网站| 久久久久网色| 啦啦啦啦在线视频资源| 国产乱人偷精品视频| 老女人水多毛片| 欧美97在线视频| 波多野结衣巨乳人妻| 特级一级黄色大片| 国产白丝娇喘喷水9色精品| 国产一级毛片七仙女欲春2| 午夜日本视频在线| 街头女战士在线观看网站| 亚洲欧洲国产日韩| 777米奇影视久久| 久久99热这里只频精品6学生| 久久国内精品自在自线图片| a级毛色黄片| 在线观看人妻少妇| 亚洲性久久影院| av在线蜜桃| 国产三级在线视频| 亚洲国产av新网站| 日韩不卡一区二区三区视频在线| 国产亚洲最大av| 久久99热这里只频精品6学生| 国产在视频线精品| 秋霞在线观看毛片| 午夜福利成人在线免费观看| 亚洲在线观看片| 久久久久国产网址| 欧美日韩一区二区视频在线观看视频在线 | 美女被艹到高潮喷水动态| 国产伦一二天堂av在线观看| 亚洲经典国产精华液单| 91久久精品电影网| 99久久精品热视频| 综合色av麻豆| 亚洲婷婷狠狠爱综合网| 国产美女午夜福利| 中文字幕制服av| 国产成人精品一,二区| 亚洲无线观看免费| 国产高清不卡午夜福利| 婷婷色麻豆天堂久久| 高清日韩中文字幕在线| 丰满乱子伦码专区| 一级毛片我不卡| 精品人妻熟女av久视频| 日日啪夜夜爽| 国产伦一二天堂av在线观看| 亚洲欧美一区二区三区黑人 | 夫妻性生交免费视频一级片| 身体一侧抽搐| 亚洲欧美日韩无卡精品| 国产 亚洲一区二区三区 | 久久精品综合一区二区三区| 18禁在线无遮挡免费观看视频| 国产精品一区www在线观看| 少妇熟女欧美另类| 午夜福利成人在线免费观看| 国产成人午夜福利电影在线观看| 亚洲av二区三区四区| 亚洲精品日韩av片在线观看| 色综合亚洲欧美另类图片| a级毛片免费高清观看在线播放| 两个人视频免费观看高清| 国产中年淑女户外野战色| 国产综合懂色| 免费av观看视频| 免费观看在线日韩| 日韩制服骚丝袜av| 久久97久久精品| 韩国高清视频一区二区三区| 国产一区有黄有色的免费视频 | 欧美xxⅹ黑人| 精品久久久久久久久亚洲| 久久精品综合一区二区三区| 国内精品宾馆在线| 亚洲图色成人| 一级av片app| 能在线免费观看的黄片| 尤物成人国产欧美一区二区三区| 色综合色国产| 精华霜和精华液先用哪个| 久久99热这里只频精品6学生| 三级毛片av免费| 在线免费十八禁| 欧美xxxx性猛交bbbb| 亚洲精品乱码久久久v下载方式| 又黄又爽又刺激的免费视频.| 欧美最新免费一区二区三区| 国产午夜精品久久久久久一区二区三区| 亚洲精品成人久久久久久| 亚州av有码| 日韩强制内射视频| 国产亚洲av嫩草精品影院| 国产精品一区www在线观看| 国产色爽女视频免费观看| 不卡视频在线观看欧美| 久久久久久九九精品二区国产| 日韩大片免费观看网站| 插阴视频在线观看视频| 亚洲精品乱码久久久v下载方式| 久久久久久久久久成人| 神马国产精品三级电影在线观看| 丰满乱子伦码专区| 免费大片18禁| 亚洲精品国产av成人精品| 亚洲国产精品国产精品| 特级一级黄色大片| 亚洲精品自拍成人| 亚洲精品日韩av片在线观看| 青青草视频在线视频观看| 亚洲欧美一区二区三区国产| 天堂av国产一区二区熟女人妻| 不卡视频在线观看欧美| 国产精品国产三级专区第一集| 亚洲av成人av| 丝袜美腿在线中文| 国产69精品久久久久777片| 夜夜爽夜夜爽视频| 人人妻人人看人人澡| 99久久九九国产精品国产免费| 亚洲欧美精品专区久久| 美女xxoo啪啪120秒动态图| 熟女人妻精品中文字幕| 极品教师在线视频| 97超碰精品成人国产| av网站免费在线观看视频 | 久久97久久精品| 久久草成人影院| av在线亚洲专区| 男的添女的下面高潮视频| av播播在线观看一区| 综合色丁香网| 亚洲一区高清亚洲精品| 久久精品国产亚洲av涩爱| 97人妻精品一区二区三区麻豆| 日本av手机在线免费观看| 亚洲精品456在线播放app| 天堂av国产一区二区熟女人妻| 国产永久视频网站| 成人特级av手机在线观看| 亚洲国产色片| 一级毛片aaaaaa免费看小| 一区二区三区高清视频在线| 欧美精品国产亚洲| 三级毛片av免费| 久久亚洲国产成人精品v| 午夜日本视频在线| 噜噜噜噜噜久久久久久91| 午夜激情欧美在线| 国产淫语在线视频| 国产av码专区亚洲av| 亚洲av.av天堂| av国产久精品久网站免费入址| 亚洲不卡免费看| 国产毛片a区久久久久| 美女主播在线视频| 99九九线精品视频在线观看视频| 国产精品久久久久久精品电影小说 | 免费观看在线日韩| 久久久成人免费电影| 国产精品久久久久久av不卡| 亚洲久久久久久中文字幕| 伊人久久精品亚洲午夜| 亚洲av国产av综合av卡| 亚洲国产欧美人成| 久久人人爽人人片av| 尾随美女入室| 人人妻人人澡欧美一区二区| 乱人视频在线观看| 久久精品久久久久久久性| 日韩在线高清观看一区二区三区| 亚洲av中文av极速乱| 一级毛片久久久久久久久女| 三级国产精品片| 中文资源天堂在线| 2022亚洲国产成人精品| 麻豆国产97在线/欧美| 亚洲精品第二区| 午夜免费观看性视频| 国产免费一级a男人的天堂| 国产黄a三级三级三级人| 人妻少妇偷人精品九色| 午夜免费男女啪啪视频观看| 亚洲精品,欧美精品| 啦啦啦中文免费视频观看日本| 久久精品久久久久久久性| 熟妇人妻久久中文字幕3abv| 免费观看a级毛片全部| 两个人视频免费观看高清| 久久精品国产亚洲av天美| 精品国产一区二区三区久久久樱花 | 亚洲av电影在线观看一区二区三区 | videossex国产| 国产免费又黄又爽又色| 男女边摸边吃奶| 免费看美女性在线毛片视频| 国产精品一区二区三区四区久久| 国产精品精品国产色婷婷| 99久国产av精品| 色综合色国产| 国产 亚洲一区二区三区 | 高清欧美精品videossex| 亚洲av中文字字幕乱码综合| 免费高清在线观看视频在线观看| 九色成人免费人妻av| 国产一级毛片在线| 少妇高潮的动态图| 最近最新中文字幕免费大全7| 亚洲国产精品成人久久小说| 小蜜桃在线观看免费完整版高清| 国内精品美女久久久久久| 精品久久国产蜜桃| 亚洲av不卡在线观看| 国产高清不卡午夜福利| 亚洲自偷自拍三级| 水蜜桃什么品种好| 欧美+日韩+精品| 精品一区二区免费观看| 成人性生交大片免费视频hd| 人人妻人人看人人澡| 亚洲美女视频黄频| 国产老妇伦熟女老妇高清| 人人妻人人看人人澡| 亚洲欧美成人综合另类久久久| 久久国内精品自在自线图片| 久久精品国产亚洲av涩爱| 精品午夜福利在线看| 国产69精品久久久久777片| 成年女人看的毛片在线观看| 日韩强制内射视频| 老女人水多毛片| 欧美日韩在线观看h| 一本久久精品| 亚洲国产日韩欧美精品在线观看| 国产成人免费观看mmmm| 午夜激情福利司机影院| 欧美一级a爱片免费观看看| 少妇被粗大猛烈的视频| 能在线免费看毛片的网站| 亚洲,欧美,日韩| 日韩精品有码人妻一区| 亚洲av免费高清在线观看| 色5月婷婷丁香| 精品人妻视频免费看| 成人无遮挡网站| 精品国内亚洲2022精品成人| 中文字幕av成人在线电影| 少妇被粗大猛烈的视频| 插阴视频在线观看视频| 可以在线观看毛片的网站| 国产一区二区三区av在线| 欧美zozozo另类| 国产成人精品一,二区| 街头女战士在线观看网站| 男女视频在线观看网站免费| 国产大屁股一区二区在线视频| 边亲边吃奶的免费视频| 淫秽高清视频在线观看| 熟妇人妻不卡中文字幕| 国产av码专区亚洲av| 日韩av免费高清视频| 欧美3d第一页| 九九在线视频观看精品| 丰满乱子伦码专区| 91精品国产九色| 三级国产精品欧美在线观看| 久久综合国产亚洲精品| 在线观看人妻少妇| 在线观看av片永久免费下载| 亚洲国产精品成人综合色| 欧美潮喷喷水| 少妇人妻一区二区三区视频| 高清欧美精品videossex| 免费看不卡的av| 亚洲国产av新网站| 99热6这里只有精品| 欧美最新免费一区二区三区| a级毛片免费高清观看在线播放| 亚洲av电影不卡..在线观看| 夜夜看夜夜爽夜夜摸| 激情 狠狠 欧美| 丝瓜视频免费看黄片| 欧美xxⅹ黑人| 日韩成人av中文字幕在线观看| 国产大屁股一区二区在线视频| 日日啪夜夜撸| 美女内射精品一级片tv| 久久久久久久久中文| 精品国产三级普通话版| 中文乱码字字幕精品一区二区三区 | 精品国产露脸久久av麻豆 | 三级国产精品片| 波野结衣二区三区在线| 亚洲欧洲日产国产| 男人狂女人下面高潮的视频| 国产熟女欧美一区二区| 校园人妻丝袜中文字幕| 精品人妻一区二区三区麻豆| 日韩av在线免费看完整版不卡| 久久99热这里只频精品6学生| 国产探花在线观看一区二区| 美女大奶头视频| av女优亚洲男人天堂| 少妇熟女aⅴ在线视频| 97超视频在线观看视频| 亚洲精品456在线播放app| 欧美xxⅹ黑人| 欧美zozozo另类| 91精品一卡2卡3卡4卡| 精品一区二区免费观看| 99热6这里只有精品| 最近中文字幕2019免费版| 国产亚洲最大av| 最近最新中文字幕大全电影3| 亚洲欧美精品专区久久| 亚洲精品日本国产第一区| 免费看光身美女| 又粗又硬又长又爽又黄的视频| 国产v大片淫在线免费观看| 午夜福利高清视频| 丰满少妇做爰视频| 老师上课跳d突然被开到最大视频| 在线观看av片永久免费下载| 91精品伊人久久大香线蕉| 色尼玛亚洲综合影院| 自拍偷自拍亚洲精品老妇| 国语对白做爰xxxⅹ性视频网站| www.av在线官网国产| av又黄又爽大尺度在线免费看| 特大巨黑吊av在线直播| 夜夜看夜夜爽夜夜摸| 丝袜美腿在线中文| 18+在线观看网站| 美女黄网站色视频| 亚洲国产成人一精品久久久| 日韩精品青青久久久久久| 亚洲精品国产av成人精品| 一个人看视频在线观看www免费| 又粗又硬又长又爽又黄的视频| 一级黄片播放器| 免费观看av网站的网址| 亚洲av一区综合| 联通29元200g的流量卡| 热99在线观看视频| 麻豆av噜噜一区二区三区| 纵有疾风起免费观看全集完整版 | 99久国产av精品| 久久久久九九精品影院| 日本一本二区三区精品| 国产伦精品一区二区三区视频9| 特大巨黑吊av在线直播| 亚洲欧美日韩无卡精品| 秋霞在线观看毛片| 午夜福利网站1000一区二区三区| 精品久久久精品久久久| 国产毛片a区久久久久| 综合色丁香网| 伊人久久精品亚洲午夜| 又粗又硬又长又爽又黄的视频| 国产美女午夜福利| 国产精品一区二区在线观看99 | 久久6这里有精品| 美女被艹到高潮喷水动态| 成年人午夜在线观看视频 | 亚洲伊人久久精品综合| 亚洲精品第二区| 亚洲av免费在线观看| 免费黄网站久久成人精品| 亚洲熟女精品中文字幕| 亚洲一区高清亚洲精品| 91午夜精品亚洲一区二区三区| 国语对白做爰xxxⅹ性视频网站| 伦理电影大哥的女人| 国产高清国产精品国产三级 | 春色校园在线视频观看| 免费av不卡在线播放| 成人午夜高清在线视频| 特级一级黄色大片| 精品一区在线观看国产| 夜夜爽夜夜爽视频| 国产成人免费观看mmmm| 欧美高清性xxxxhd video| av免费观看日本| 成年免费大片在线观看| 精品人妻偷拍中文字幕| 精品欧美国产一区二区三| 久久久久久久久久人人人人人人| 欧美激情在线99| 岛国毛片在线播放| 一夜夜www| 91精品一卡2卡3卡4卡| 国国产精品蜜臀av免费| 国产免费又黄又爽又色| av免费在线看不卡| 亚洲精品456在线播放app| 久久99精品国语久久久| 亚洲成人久久爱视频| 日本-黄色视频高清免费观看| 性色avwww在线观看| 成年版毛片免费区| 天美传媒精品一区二区| 在线播放无遮挡| 欧美日韩视频高清一区二区三区二| kizo精华| 久久久精品欧美日韩精品| 看黄色毛片网站| 国产精品日韩av在线免费观看| 联通29元200g的流量卡| 亚洲国产欧美在线一区| videos熟女内射| 日日撸夜夜添| 高清毛片免费看| 亚洲精品日韩av片在线观看| 麻豆av噜噜一区二区三区| 国产av在哪里看| 日韩一区二区三区影片| freevideosex欧美| 能在线免费看毛片的网站| 亚洲av成人精品一区久久| 国产精品av视频在线免费观看| 午夜福利成人在线免费观看| 26uuu在线亚洲综合色| 蜜桃亚洲精品一区二区三区| 蜜臀久久99精品久久宅男| 又粗又硬又长又爽又黄的视频| 国产探花在线观看一区二区| 麻豆国产97在线/欧美| 久久久精品免费免费高清| 国产高清国产精品国产三级 | 国产精品人妻久久久影院| 日本黄大片高清| a级毛色黄片| 我的老师免费观看完整版| 日韩亚洲欧美综合| 大话2 男鬼变身卡| 2021少妇久久久久久久久久久| 国产一级毛片七仙女欲春2| 久久国内精品自在自线图片| 亚洲第一区二区三区不卡| 亚洲av免费高清在线观看| 午夜福利在线观看免费完整高清在| 午夜老司机福利剧场| av福利片在线观看| av国产久精品久网站免费入址| 国产有黄有色有爽视频| 日本免费a在线| 久久精品国产亚洲av涩爱| 91在线精品国自产拍蜜月| 亚洲无线观看免费| 日本av手机在线免费观看| www.av在线官网国产| 久久精品久久精品一区二区三区| 九草在线视频观看| videossex国产| 老女人水多毛片| 天堂√8在线中文| 成人欧美大片| 国产亚洲最大av| 91狼人影院| 两个人视频免费观看高清| 伦理电影大哥的女人| 97热精品久久久久久| 肉色欧美久久久久久久蜜桃 | 97热精品久久久久久| av.在线天堂| 99久国产av精品国产电影| 国产女主播在线喷水免费视频网站 | 晚上一个人看的免费电影| 久久这里只有精品中国| 又爽又黄a免费视频| 亚洲av福利一区| 美女被艹到高潮喷水动态| 成人特级av手机在线观看| 午夜免费男女啪啪视频观看| 偷拍熟女少妇极品色| 亚洲色图av天堂| 亚洲欧美清纯卡通| 国产免费福利视频在线观看| 男人爽女人下面视频在线观看| 久久久亚洲精品成人影院| 三级国产精品片| 哪个播放器可以免费观看大片| 午夜福利成人在线免费观看| 亚洲美女视频黄频| 丝袜喷水一区| 国产精品人妻久久久影院| 国产成人午夜福利电影在线观看| 最近视频中文字幕2019在线8| 小蜜桃在线观看免费完整版高清| 一级毛片久久久久久久久女| 欧美不卡视频在线免费观看| 日本一二三区视频观看| 国产免费一级a男人的天堂| 少妇人妻精品综合一区二区| 亚洲国产精品sss在线观看| 亚洲精品乱码久久久v下载方式| 免费不卡的大黄色大毛片视频在线观看 | 国产男女超爽视频在线观看| 在线 av 中文字幕| 亚洲精品自拍成人| 又爽又黄无遮挡网站| 精品久久久久久电影网| 精品一区二区三卡| 欧美性感艳星| 免费av不卡在线播放| 在线免费十八禁| 亚洲欧美日韩无卡精品| 狠狠精品人妻久久久久久综合| 美女国产视频在线观看| 久久精品熟女亚洲av麻豆精品 | 国产精品伦人一区二区| 免费播放大片免费观看视频在线观看| 真实男女啪啪啪动态图| 亚洲18禁久久av| 午夜视频国产福利| 久久精品久久精品一区二区三区| 精品国产一区二区三区久久久樱花 | 又爽又黄a免费视频| 九色成人免费人妻av| 亚洲三级黄色毛片| 国产精品女同一区二区软件| 人妻制服诱惑在线中文字幕| 色播亚洲综合网| 亚洲av日韩在线播放| 久久精品国产亚洲网站| 最新中文字幕久久久久| 大又大粗又爽又黄少妇毛片口|