• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    HCSP-Net:A Novel Model of Age-Related Macular Degeneration Classification Based on Color Fundus Photography

    2024-05-25 14:39:38ChengWanJianiZhaoXiangqianHongWeihuaYangandShaochongZhang
    Computers Materials&Continua 2024年4期

    Cheng Wan ,Jiani Zhao ,Xiangqian Hong ,Weihua Yang,? and Shaochong Zhang,?

    1College of Electronic and Information Engineering/College of Integrated Circuits,Nanjing University of Aeronautics and Astronautics,Nanjing,211106,China

    2Shenzhen Eye Institute,Shenzhen Eye Hospital,Jinan University,Shenzhen,518040,China

    ABSTRACT Age-related macular degeneration(AMD)ranks third among the most common causes of blindness.As the most conventional and direct method for identifying AMD,color fundus photography has become prominent owing to its consistency,ease of use,and good quality in extensive clinical practice.In this study,a convolutional neural network (CSPDarknet53) was combined with a transformer to construct a new hybrid model,HCSP-Net.This hybrid model was employed to tri-classify color fundus photography into the normal macula(NM),dry macular degeneration (DMD),and wet macular degeneration (WMD) based on clinical classification manifestations,thus identifying and resolving AMD as early as possible with color fundus photography.To further enhance the performance of this model,grouped convolution was introduced in this study without significantly increasing the number of parameters.HCSP-Net was validated using an independent test set.The average precision of HCSPNet in the diagnosis of AMD was 99.2%,the recall rate was 98.2%,the F1-Score was 98.7%,the PPV (positive predictive value)was 99.2%,and the NPV(negative predictive value)was 99.6%.Moreover,a knowledge distillation approach was also adopted to develop a lightweight student network(SCSP-Net).The experimental results revealed a noteworthy enhancement in the accuracy of SCSP-Net,rising from 94%to 97%,while remarkably reducing the parameter count to a quarter of HCSP-Net.This attribute positions SCSP-Net as a highly suitable candidate for the deployment of resource-constrained devices,which may provide ophthalmologists with an efficient tool for diagnosing AMD.

    KEYWORDS Computer-aided diagnosis;deep learning;age-related macular degeneration;transformer

    1 Introduction

    According to a study published inThe Lancet Global Health,the number of patients with severe visual impairment or blindness due to age-related macular degeneration(AMD)is anticipated to reach 288 million by 2040,bringing a significant burden to ophthalmologists[1].As reported in some studies,AMD is influenced by age,genetics,and complex environmental factors(such as smoking and diet)[2,3].However,it remains unclear about the specific pathogenesis of AMD.Aging of macular tissues is believed to be the primary cause of AMD,which is more prevalent in individuals over the age of 50 years[4–6].There are serious challenges to global eye health in the aging population[1,3,7].Although AMD can exert serious effects on patient’s health,it has not yet attracted a high degree of societal attention or investment in medical resources.Insufficient knowledge and medical resources pose dire threats to patients and hinder limited ophthalmologists from providing convenient and comprehensive diagnostic service for such a large patient population.

    AMD can be categorized into Normal Macula(NM),Dry Macular Degeneration(DMD),and Wet Macular Degeneration (WMD) based on its clinical manifestations and imaging features,as shown in Fig.1 [3,8,9].Over 80% of patients with AMD present with DMD,defined by the early development of drusen and later development of geographic atrophy,resulting in a gradual loss of vision and vision distortion[6,10].In contrast,WMD develops from DMD,in which drusen of varying sizes form rapidly in the maculae.As the affected area rapidly invades the surrounding tissues,this condition can cause severe vision loss or even blindness[11].In this study,the classification based on clinical manifestations could aid ophthalmologists in making quick diagnoses and selecting effective therapies.

    Figure 1: (A)Normal macula;(B)dry macular degeneration;(C)wet macular degeneration

    It is important to note that conventional diagnostic techniques have been improved by deeplearning (DL) technology.DL techniques can be employed to avoid misdiagnosis in subjective assessments and make efficient and accurate diagnoses by objectively analyzing numerous data.This can reduce the burden on physicians and facilitate better therapeutic outcomes.A convolutional neural network(CNN)is the main model used in DL.As the depth of the CNN increases,the backpropagation algorithm can be utilized to solve the contribution distribution problem of each layer of the network,so that the model can be applied to prediction.Medical images feature uniform specifications,ease of use,and high quality in long-term clinical practice.The use of DL to process medical images exhibits extremely broad application possibilities and outperforms human experts in diagnosing certain diseases [12–16].Priya et al.proposed a probabilistic graphical model and a series of image-preprocessing techniques to classify AMD [17].To improve the visibility of lesions,they first extracted the green channel.Then,the discrete wavelet transforms in combination with the Kirsch operator were used to locate vessels and identify potential lesions.This optimized preprocessing pipeline effectively extracted pathological discriminative features,achieving a classification accuracy of 96%.In contrast to single-model approaches,Grassmann et al.proposed an integrated framework of convolutional neural networks based on random forests for the automated detection of macular degeneration [18].Through the integration of multiple independently trained CNNs,the integrated model demonstrated superior classification performance compared with individual human experts.Motozawa et al.constructed another dual-model approach for diagnosing AMD[19].The first model was designed to differentiate between NM and AMD,and the second one was used to detect exudative changes in AMD,differentiating between DMD and WMD.However,the construction of two separate convolutional neural networks was required,and a diagnostic accuracy of 93.9%was only achieved for AMD.Vague et al.explored the application of multimodal image analysis methods in three cohorts:Young normal,old normal,and dry macular maculopathy [20].The results showed a diagnostic accuracy of 96%for this task when combined with multimodal training,thus validating the excellence of multimodality for the diagnosis of macular lesions.However,multimodal models are more complex than normal ones and require more computational resources and time for training and optimization.

    Nonetheless,there are still limitations in accurately diagnosing AMD through existing models.Firstly,manual feature engineering and image preprocessing are highly dependent,and more automation and intelligence are required.Secondly,sample imbalance still needs to be addressed.Thirdly,the architecture of existing models lacks lightweight and efficient deployment in real-world applications.Given that,a more automated and efficient diagnostic framework was introduced in this study to overcome these challenges.The proposed HCSP-Net,a DL model with a hybrid architecture,aims to tackle the issues resulting from uneven data distribution.Notably,HCSP-Net achieves an impressive diagnostic accuracy of 99%.Furthermore,knowledge distillation was employed to develop SCSPNet—a lightweight network with enhanced feasibility on resource-limited platforms just based on 1.05 MB parameters.

    2 Structure

    2.1 Data Acquisition

    This study has obtained ethical approval to collect retinal images from the Shenzhen Eye Hospital for research purposes,following the principles outlined in theDeclaration of Helsinki[21].To address privacy concerns associated with patient data usage,careful steps were taken before color fundus photographs were incorporated into the dataset.Utilizing the OpenCV toolkit,all patient-related information,including patient name,age,and date of diagnosis,was systematically removed from fundus photographs[22].During the data preprocessing phase,which involved transforming fundus photographs into binary images,a 10×10 operator was applied to eliminate any remaining patientrelated information.Subsequently,contour screening(extracting the contour with the most significant area) was performed to determine the retinal area and cropping the retinal area.Therefore,it is important to highlight that this study does not disclose patient-related statistical information.

    Recognizing the critical role of image quality in influencing the data analysis of models,745 data points were carefully selected,excluding those with inferior exposure.Besides,a meticulous approach using random number seeds was used for dataset division.This ensured a balanced proportion of each macular data in both Datasets A and B,which achieved equitable data distribution and mitigated the impact of subjective factors,thus ensuring fairness in the experimental process.Additional details about the data division are presented in Table 1.Owing to the limited number of dry maculae(only half as many as the other two maculae),the network’s generalization and convergence may be severely hampered by the uneven spread of data[23].Section 2.2 of this paper provides details of the approach we used.

    Table 1: Number of fundus images in Datasets A and B from clinical classification of AMD

    2.2 Composition of Model Structure

    Convolutional neural networks have been extensively applied to image processing and are constantly evolving,with many excellent networks emerging[24–26].However,when an image is processed locally by CNNs via convolution and pooling,the influence of the context surrounding the image is ignored.This study combines a convolutional neural network with a transformer to access the specific and general contextual information of an image.The core feature extraction component of HCSPNet involves an improved convolutional neural,CSPDarknet53.CSPDarknet53 reduces the number of model parameters while resolving redundant gradients[24].The structure of HCSP-Net is shown in Fig.2.The improved CSPDarknet53 is the backbone of YOLOv5(You Only Look Once-v5)and comprises CSPDarknet53 and SPPF.YOLOv5 comprises four models with various depths and widths:S,M,L,and X.In this study,the backbone of the S model,which has the smallest width and depth,was used to construct a lightweight model.

    Figure 2: The structure of HCSP-Net

    CSPDarknet53 reduces the number of model parameters while resolving redundant gradients.In addition,the data will be deformed when the model input part uniformly scales the image to a fixed resolution.In this study,the SPPF module was retained in the design of the model,making the HCSP-Net more resistant to object deformation[27].Yet,the original SPPF module relies solely on the maximum pooling layer to emphasize the most salient or active features within a given region,disregarding others.In this study,it was maintained that after maximum pooling,there was a need for further integration of features to ensure the effective capture of information from the entire feature map.Consequently,a more intricate integration strategy was adopted in the SPPF module.The convolutional operation was incorporated after each maximal pooling layer to enhance the network’s perceptual capabilities.Additionally,in consideration of the model’s parameter count,the conventional convolution operation was substituted with grouped convolution,aiming to reduce parameters without compromising performance.The enhanced SPPF module is illustrated in Fig.3.The introduction of grouped convolution provided HCSP-Net with more potent feature extraction and processing capabilities.This established a robust foundation for the model to adeptly handle diverse and dynamic data in practical applications.

    Figure 3: ESPPF:Enhanced SPPF module with GCONV(grouped convolution)

    The transformer,a model based on a self-attention mechanism originally proposed by Google for machine translation tasks,has demonstrated superiority over conventional convolutional neural networks in processing long sequence data and modeling global dependencies [28].It has become a pivotal technique in various domains,including image processing[28–31].The classification module employs a Transformer block to capture long-term dependencies in the data while preserving sequence information [28].The structure of the Transformer block is illustrated in Fig.4.Besides,in light of the uneven data distribution and the shared characteristics between DMD and WMD,this study was conducted to enhance the model’s ability to distinguish between the two by incorporating a transformer module.This addition seeks to mitigate the impact of data imbalance on the model’s performance.

    Of note,due to the small size of the feature map output from the convolutional neural network(7×7),the single-headed self-attention mechanism was directly used for calculations.The self-attention can be calculated as follows:

    After the calculation by the Transformer module,its output was transformed into a 3×1 vector using a linear layer,where each vector represents the probability of corresponding to NM,DMD,and WMD,respectively.

    2.3 SCSP-Net Boosted by Knowledge Distillation

    The Teacher-Student Training (TST) method is a widely recognized approach to knowledge distillation[32].TST involves the construction of a teacher network with significant depth and width,alongside generating a comparatively lightweight student network.Then,the trained teacher network is used to supervise the training of the student network,aiming to enhance its overall performance.In this study,the well-trained HCSP-Net served as the teacher network,while a lightweight student network,SCSP-Net,was constructed (refer to Table 2 for model structures).HCSP-Net and SCSPNet exhibited notable distinctions in channel configuration and module settings.Notably,SCSPNet featured half the number of channels across the entire network compared with HCSP-Net.In addition,differences in the number of modules between the two models were observed at specific layers.For example,in the third C3 layer,HCSP-Net incorporated three duplicate C3 modules,whereas SCSP-Net employed only two duplicate stacked modules.HCSP-Net’s model parameters amounted to 4.19 MB,while SCSP-Net,through a reasonable reduction in width and depth,achieved a parameter count of 1.05 MB—approximately one-fourth of that of HCSP-Net.The amalgamation of lightweight design and knowledge distillation rendered the more pragmatic and viable deployment of SCSP-Net on embedded devices.In the knowledge distillation process,the generalization performance of the student model was improved by transferring the soft labels of the teacher model (HCSP-Net).This enabled SCSP-Net to effectively incorporate knowledge from HCSP-Net.

    Table 2: Architectures for HCSP-Net and SCSP-Net.In the table,‘n’represents the number of modules,‘c’represents the number of channels,‘w’represents the width of the feature map,and‘h’represents the height of the feature map

    Figure 4: The structure of the Transformer block

    The loss function comprises two key componesnts,namely,KL divergence and cross-entropy loss.These components can be employed to measure the difference in the output distribution between the teacher-student networks and the matching degree between the student network and the real label,respectively.The specific expressions for KL divergence and cross-entropy loss are shown as follows:

    In the above equation,the symbol ∝represents the weighting coefficient in the loss function,the symbolγrepresents the soft labeling of the output of the teacher’s network,the symbol p represents the predicted value of the student’s network,and the symbol y represents the true labeling of the input image.

    2.4 Implementation

    The HCSP-Net model was built in PyTorch based on Python 3.7.11,and a GPU (NVIDIA GeForce RTX 1080)was used for the experiments[33].Due to the limited data collected,Dataset A was expanded by skimming the data horizontally and flipping them vertically,with the final amount of data expanded to four times the original size.To prevent the model from overfitting,data enhancement techniques,including color space variation,random luminance contrast variation,translation scaling,and random orientation rotation,were applied to the fundus map with probabilities of 0.2,0.2,0.5,and 1.0,respectively.In the training process,Dataset A was divided according to the ratio of 8:2,which was used for model training and validation,respectively,and the weight file with the lowest epoch number and the highest accuracy was saved as the optimal model.

    2.5 Statistical Method

    The Scikit-learn toolkit was used to conduct the statistical study [34].The precision,recall,F1-Score,PPV (positive predictive value),and NPV (negative predictive value) of the HCSP-Net for NM,DMD,and WMD were determined using binary indicators.The areas under the readings and receiver operating characteristic curves(ROC)were also computed.AUC values were classified as poor diagnostic values if they were between 0.5 and 0.70,average diagnostic values if they were between 0.75 and 0.85,and excellent diagnostic values if they were above 0.85.

    The multi-classification indicator Kappa value was used to assess the degree of agreement between the true diagnostic findings and the CSPDarknet53 and HCSP-Net results.The Kappa value ranged from 0 to 1,and a higher Kappa value indicated greater agreement between the model’s predicted and actual results.The Kappa values were calculated as follows:

    wherep0represents the overall classification accuracy.pecan be calculated as follows:

    whereairepresents the number of actual samples in class i,birepresents the number of samples predicted for class i,and n represents the total number of samples.The Jaccard similarity coefficient is also a simple and effective statistical indicator of similarity and diversity.It is defined as the ratio of the number of elements in the intersection of two sets to the number of elements in the concatenation,which can be expressed as follows:

    where A and B represent the real and predicted label sets,respectively.

    3 Results

    3.1 Model Performance Evaluation

    The confusion matrix provided a more intuitive view of the classification performance of the model.CSPDarknet53 can be utilized to identify 100%of the NM;however,owing to a lack of data and similarities between DMD and WMD in terms of lesion features,it is difficult for the model to distinguish between them,as shown in Fig.5A.In Dataset B,26.3% of DMD cases and 2.4% of WMD cases were incorrectly categorized.Notably,with the addition of the transformer module,the misdiagnosis of DMD was reduced by the HCSP-Net(SPPF),indicating the improved ability of this model to capture the difference between DMD and WMD.Following a careful integration of features extracted from the SPPF through the addition of grouped convolution,only one DMD lesion was misdiagnosed as a WMD lesion by HCSP-Net (ESPPF).The corresponding confusion matrices for the two improved networks based on CSPDarknet53 are shown in Figs.5B and 5C,respectively.Hence,the problem of uneven data distribution was resolved,resulting in improved model efficiency.

    Figure 5: Confusion matrix.(A)CSPDarknet53;(B)HCSP-Net(SPPF);(C)HCSP-Net(ESPPF).0 for normal macula,1 for dry macular degeneration,2 for wet macular degeneration

    To objectively evaluate the performance of HCSP-Net,the receiver operating curve (ROC) was used in this study.Fig.6 illustrates the AUC values for NM,DMD,and WMD,which were 1.0000,0.9974,and 0.9880,respectively.

    Figure 6: Roc curve

    3.2 Enhanced SPPF Performance

    In the comparison between SPPF and ESPPF,identical parameter settings,such as the learning rate,optimizer,and weight decay,were utilized in this study to ensure fair experimental results.As shown in Table 3,HCSP-Net (ESPPF) outperforms HCSP-Net (SPPF) model,with a 3.1% relative improvement in the kappa value(increasing from 95.3%to 98.4%)and a 4.7%relative improvement in the Jaccard value(increasing from 92.8%to 97.5%).

    Table 3: Comparison of Kappa and Jaccard values between CSPDarknet53 and HCSP-Net

    In addition,the impact of SPPF and ESPPF on the convergence of the model was explored in this study.As depicted in Fig.7,the initial 30 training epochs reveal a noteworthy observation: HSCPNet(ESPPF)demonstrates a significantly higher initial accuracy compared with HCSP-Net(SPPF).Furthermore,throughout the entire training process,the accuracy curve of HSCP-Net (ESPPF)steadily ascends,ultimately converging.This indicates that HCSP-Net(ESPPF)is adept at capturing underlying patterns and features in the data during the early stages of model training,effectively mitigating the impact of data noise.

    3.3 SCSP-Net Performance Insights

    The compression of the depth and width of HCSP-Net resulted in an increase in the training time of the model and a significant decrease in the diagnostic accuracy to only 94%,as shown in Fig.8A.This study introduced the HCSP-Net(ESPPF)with an accuracy of 99%as a teacher network to guide the SCSP-Net training to address this issue.Under the supervision of the teacher network,SCSP-Net successfully reduced the confusion between DMD and WMD of this model,and the accuracy was improved to 97%in Dataset(as shown in Fig.8B).Drawing on the knowledge of the teacher network cannot only alleviate the performance degradation caused by HCSP-Net compression but also provide more precise guidance for SCSP-Net to perform better when facing complex disease classification tasks.

    Figure 7: Comparison of the training set accuracy between HSCP-Net (SPPF) and HCSP-Net(ESPPF)

    Figure 8: Confusion matrix.Subfigure(A)illustrates SCSP-Net without knowledge distillation,while subfigure(B)depicts SCSP-Net after knowledge distillation

    According to the statistics in Table 4,the number of parameters and time complexity of different models were comprehensively analyzed in this study.Notably,in HCSP-Net(SPPF),replacing conventional fully connected layers with Transformer modules reduced the number of parameters.However,despite the decrease in the number of parameters,the time complexity did not show a significant declining trend.This indicated that introducing Transformer modules may face particular challenges in improving computational efficiency,requiring further in-depth research and optimization.In addition,after grouped convolutions were introduced,HCSP-Net(ESPPF)showed increased parameters and time complexity compared with the baseline model and HCSP-Net(SPPF).Although the increase was inevitable,this increase was acceptable considering the superior performance of HCSP-Net(ESPPF).It was noteworthy that by compressing the width and depth of the model,the number of parameters and time complexity of SCSP-Net were reduced by 74.90%and 72.97%,respectively,demonstrating its advantages in lightweight design.

    Table 4: Comprehensive analysis of model parameters and computational complexity

    3.4 Ablation Experiment

    In this study,the HCSP-Net model was constructed based on an innovative amalgamation of the enhanced CSPDarknet53 and the Transformer module.This model exhibited substantial distinctions compared with conventional counterparts,such as Resnet50,EfficientNetV2,InceptionV3,and ViT.The underlying rationale for this fusion design lay in the deliberate exploitation of the robust feature extraction capabilities of CSPDarknet53 and the utilization of the Transformer module to capture intricate long-term dependencies within the data,thereby facilitating a more profound exploration of abstract features within images.Regarding feature engineering,incorporating the Transformer module empowered HCSP-Net to discern and capture long-term dependencies inherent in image data more effectively,thus extracting expressive and discriminative features.Furthermore,the ESPPF module played a pivotal role during the initial stages of model training,significantly reducing noise impacts on the model.This aspect was of great significance when addressing challenges associated with data heterogeneity and inhomogeneity prevalent in medical images.To further demonstrate the superior performance of HCSP-Net,the performance of several models was evaluated using binary classification indicators.The experimental results are listed in Table 5.All models in this experiment were trained and validated on Dataset A and tested on Dataset B.The obtained test results were considered the ultimate performance indicators for each model.HCSP-Net (ESPPF) performed optimally in all other metrics compared with other models,achieving a precision of 99.2%,recall of 98.2%,F1-Score of 98.7%,PPV of 99.2%,and NPV of 99.6%.The ablation experiment also confirmed the superior performance of HCSP-Net,as the addition of the transformer and ESPPF module led to an improvement in model performance.

    Table 5: HCSP-Net performance evaluation:Precision,recall,F1-Score,PPV,and NPV

    Moreover,to further validate the generalization capability of HCSP-Net (ESPPF),a validation experiment was performed on a publicly available dataset containing 11 types of retinal fundus images(https://www.kaggle.com/datasets/kssanjaynithish03/retinal-fundus-images/data).NM,DMD,and WMD images were extracted from this dataset to construct an external validation set.Then,HCSPNet(ESPPF)was retrained and tested on this validation set.As shown in Fig.9,HCSP-Net(ESPPF)achieves a classification accuracy of 100% on the test set with 256 images.This validation result demonstrated that HCSP-Net (ESPPF) had a strong generalization ability in distinguishing NM,DMD,and WMD fundus images and can adapt to unseen new datasets,laying a foundation for subsequent clinical applications.

    Figure 9: Confusion matrix.0 for normal macular,1 for dry macular degeneration,2 for wet macular degeneration

    4 Discussion

    Age-related macular degeneration (AMD) is a major cause of irreversible damage to vision in individuals over the age of 50,affecting the health of millions of people worldwide [5,7,10].Early detection and treatment of AMD contribute to slow disease progression.However,with the increasing number of patients with AMD,it is difficult for ophthalmologists to provide a comprehensive diagnostic service for this large patient population;therefore,the use of DL techniques to intelligently analyze AMD is of great clinical importance.

    T-distributed stochastic neighbor embedding (T-SNE) is a nonlinear dimensionality reduction algorithm utilized for visualizing high-dimensional data [37].As there is a lack of interpretability in neural networks,T-SNE was employed in this study to reduce the dimensionality of the last hidden layer in both CSPDarknet53 and HCSP-Net (ESPPF) models.Characteristic probability distribution maps for various types of macular degeneration were generated to aid in interpreting the neural network models,as shown in Fig.10.When applied to the macular degeneration triple classification task,the CSPDarknet53 model exhibited a noticeable overlap between the sample data of DMD and WMD.This suggested that CSPDarknet53 did not effectively learn the distinctive features necessary for differentiating between the two conditions.Conversely,the visualization results of HCSP-Net showed a low percentage of sample data overlap,indicating that this model had an improved ability to distinguish between DMD and WMD.This underscored the enhanced capacity of this model to learn the distinctions between the two conditions,successfully differentiating them.Additionally,it was apparent from the probability distribution plots that the probability plots of DMD and WMD overlapped,further emphasizing the relationship between DMD and WMD,namely that WMD developed in the context of DMD.Clinically,reliably differentiating DMD from WMD is critical for appropriate treatment decisions.Thus,the difficulty of CSPDarknet53 in separating these classes could lead to delays in treating WMD or unnecessary treatment for DMD due to incorrect classification.Meanwhile,the more apparent separation achieved by HCSP-Net demonstrated its potential to support more accurate clinical diagnosis and management between these AMD subtypes.These visualization results complement our analysis,offering a more comprehensive understanding of the model’s behavior and its implications for clinical decision-making.

    Figure 10: t-SNE dimensionality reduction visualization.(A)CSPDarknet53;(B)HCSP-Net(ESPPF)

    Moreover,the HCSP-Net architecture was examined and evaluated experimentally.The Transformer model was tested using various model architectures,such as replacing only one C3 module at a time at a different location from the transformer module and replacing all C3 modules with the transformer module.However,none of these models performed as well as HCSP-Net,which may be explained from two aspects.First,the small amount of data used in this study prevented the transformer model from learning all the characteristics of each macula,which was necessary for it to obtain superiority over the convolutional neural network model[22].Second,after pre-training on the COCO dataset,the improved CSPDarknet53 was loaded with weights and was able to derive general features from the images.In this instance,changing the intermediate CSPDarknet53 architecture at random would nullify the impact of the pre-trained weights.The Transformer module successfully integrated its effect on the backbone feature extraction module’s information without erasing the pretrained weights,thereby enhancing the performance of HCSP-Net and allowing the model to learn new features.This allows the model to detect minute variations between various maculae,thereby improving the model’s performance while preserving its lightweight characteristics.

    HCSP-Net,constructed in this study,had a parameter count of 4.19 MB.Further reduction of model parameters to enhance its deployment on mobile devices represented the challenge addressed in this study.Building upon HCSP-Net (ESPPF),knowledge distillation was employed to guide the training of SCSP-Net in this study,resulting in an accuracy improvement from 94% to 97%,with parameters accounting for less than a quarter of HCSP-Net.In CSPDarknet53 and HCSP-Net,there was no misdiagnosis of DMD as NM,indicating that the model can capture the differences between the two conditions.However,even with the assistance of knowledge distillation,SCSP-Net exhibited misdiagnosis,incorrectly identifying DMD as NM.The experimental results highlighted the significant impact of model depth and width on learning data features.This indicated that larger models were more prone to extracting abstract features from the data.Consequently,striking a balance between model detection accuracy and operational speed while reducing model parameters is a key point for the future optimization of SCSP-Net.

    An innovative approach was also introduced in this study by replacing the conventional fully connected classification method with a Transformer model based on the self-attention mechanism.Additionally,improvements were achieved by incorporating group convolution to expedite model convergence and enhance robustness,leading to the development of HCSP-Net(ESPPF)and achieving favorable diagnostic outcomes.All training images were sourced from individuals diagnosed with AMD,and professional ophthalmologists meticulously labeled the dataset.HCSP-Net (ESPPF)exhibited exceptional classification precision(99.2%),recall(98.2%),F1-Score(98.7%),PPV(99.2%),and NPV(99.6%).Compared with other studies,our research obviated the need for complicated data preprocessing steps,such as single-channel data extraction or vascular localization[17].Furthermore,HCSP-Net avoided the complexity of ensemble learning or the construction of dual models to enhance accuracy[18,19].While multimodal approaches were explored for accuracy improvement,acquiring diverse data types posed significant challenges in the medical domain[20].Therefore,a straightforward and efficient diagnostic method that circumvented the complexities associated with data preprocessing and model assembly was presented in this study.

    Nevertheless,there are still some limitations in this study.Firstly,although DL has been immensely successful in many domains,acquiring large amounts of data is frequently challenged due to ethical and privacy concerns.This study used data augmentation to increase image diversity during the data preprocessing step and data inversion to extend the dataset.However,the experimental results showed that it was still necessary to improve HCSP-Net(SPPF)by increasing the number of training rounds to deal with fluctuations in accuracy during the pre-training period.This ongoing need for improvements highlighted the complexity of working with limited data in the context of DL-based research.Secondly,a three-class classification study was conducted based on the leading criteria.Expanding the dataset with more diverse cases may help further verify the robustness of the proposed methods.Thirdly,after compressing the depth and width of the model,the diagnostic accuracy of SCSP-Net decreased significantly.Further balancing the accuracy and efficiency of the model was another limitation of this study.Finally,it is required to perform the integration of the model into clinical workflows and user studies to assess its acceptance.This will be an essential direction for further research.

    This initial study focuses on model development and evaluation on a limited dataset.An important direction for future work is integrating the model into clinical workflows and evaluating its acceptance by healthcare professionals through user studies.Besides,as more AMD data with refined grading become available,the model can be retrained and modified to enhance accuracy and maintain longterm reliability.In meantime,SCSP-Net can be enhanced by leveraging the success achieved in HCSPNet,thus providing improved support for medical professionals in diagnosing AMD.

    5 Conclusion

    In conclusion,AMD is a prevalent retinal disease that can cause blindness,making prompt diagnosis critical.In this study,the viability of using DL technology to aid in the classification of AMD was demonstrated.Besides,an HCSP-Net,a classification model combining a convolutional neural network and a transformer,was constructed to achieve the AMD diagnosis accuracy of 99%.This model can fulfill functions in the early diagnosis of AMD based on color fundus photography,offering valuable assistance to clinicians.In particular,it can offer strong support for the early diagnosis and screening of AMD in primary care,assist in detecting early AMD,and provide reasonable treatment recommendations,thereby enhancing the visual quality of patients.

    Acknowledgement:We gratefully acknowledge the support of the Shenzhen Fund for Guangdong Provincial High-Level Clinical Key Specialties,the Sanming Project of Medicine in Shenzhen,and the Shenzhen Science and Technology Planning Project.

    Funding Statement:Shenzhen Fund for Guangdong Provincial High-Level Clinical Key Specialties(SZGSP014),Sanming Project of Medicine in Shenzhen (SZSM202311012) and Shenzhen Science and Technology Planning Project(KCXFZ20211020163813019).

    Author Contributions:CW and JZ: Analyzed,discussed the data,and drafted the manuscript.XH:Analyzed,discussed the data,and collected and labeled the data.SZ and WY:Designed the research,collected and labeled the data,and revised the manuscript.

    Availability of Data and Materials:The raw data supporting the conclusions of this article will be made available by the authors,without undue reservation.

    Ethics Approval:This study was approved by the Medical Ethics Committee of Shenzhen Eye Hospital(Approval Code:2023KYPJ015,Approval Date:February 24,2023).

    Conflicts of Interest:The authors declare that this study was conducted without commercial or financial relationships that could be construed as potential conflicts of interest.

    午夜福利视频精品| 午夜视频国产福利| 在线观看三级黄色| 亚洲av成人精品一二三区| 精品久久久久久久久亚洲| 欧美亚洲 丝袜 人妻 在线| 国产女主播在线喷水免费视频网站| 天天操日日干夜夜撸| 三级国产精品欧美在线观看| 在线看a的网站| 在线看a的网站| 男女啪啪激烈高潮av片| 日韩av不卡免费在线播放| 2022亚洲国产成人精品| 欧美精品一区二区免费开放| 99久久精品热视频| 亚州av有码| 亚洲一级一片aⅴ在线观看| 一区二区av电影网| 九九久久精品国产亚洲av麻豆| 少妇被粗大猛烈的视频| 少妇被粗大猛烈的视频| 亚洲综合色惰| 亚洲自偷自拍三级| 国产精品国产三级国产av玫瑰| 18禁裸乳无遮挡动漫免费视频| 亚洲av福利一区| 成人影院久久| 久久97久久精品| 我要看黄色一级片免费的| 亚洲精品第二区| 老司机亚洲免费影院| 久久国内精品自在自线图片| 久久午夜福利片| av国产久精品久网站免费入址| 一本久久精品| av不卡在线播放| 2021少妇久久久久久久久久久| 亚洲四区av| 成年人午夜在线观看视频| 亚洲欧美一区二区三区国产| 熟女人妻精品中文字幕| 国产无遮挡羞羞视频在线观看| 久久久久久久久久久丰满| 尾随美女入室| 久久99一区二区三区| 久久国产乱子免费精品| 日日撸夜夜添| 免费av中文字幕在线| 美女内射精品一级片tv| 日韩一区二区三区影片| 日日啪夜夜爽| 一级毛片aaaaaa免费看小| 伊人久久精品亚洲午夜| av福利片在线观看| av网站免费在线观看视频| 精品国产乱码久久久久久小说| 自拍欧美九色日韩亚洲蝌蚪91 | 久久97久久精品| 亚洲欧洲精品一区二区精品久久久 | 国产黄片视频在线免费观看| 亚洲av不卡在线观看| 99re6热这里在线精品视频| 深夜a级毛片| 青青草视频在线视频观看| 日韩一区二区三区影片| 一个人看视频在线观看www免费| 一区二区三区精品91| 91久久精品国产一区二区三区| 亚洲欧美中文字幕日韩二区| 亚洲精华国产精华液的使用体验| 午夜av观看不卡| 9色porny在线观看| 国产片特级美女逼逼视频| 黄色配什么色好看| 国产精品99久久久久久久久| 十八禁网站网址无遮挡 | 国产成人91sexporn| 亚洲美女视频黄频| 亚洲av国产av综合av卡| 边亲边吃奶的免费视频| 伊人亚洲综合成人网| 2022亚洲国产成人精品| 国产精品秋霞免费鲁丝片| 精品一区在线观看国产| 久久久久久久久久成人| 蜜臀久久99精品久久宅男| 免费大片18禁| 黑人巨大精品欧美一区二区蜜桃 | kizo精华| 国产 一区精品| 综合色丁香网| 久久精品国产a三级三级三级| 久久久国产精品麻豆| 九九久久精品国产亚洲av麻豆| av国产久精品久网站免费入址| 91成人精品电影| 在线天堂最新版资源| 国产精品人妻久久久影院| 日本欧美国产在线视频| 乱人伦中国视频| 亚洲欧美成人精品一区二区| 啦啦啦啦在线视频资源| 少妇的逼好多水| 在线观看国产h片| 成人无遮挡网站| 久久青草综合色| 人体艺术视频欧美日本| 日韩在线高清观看一区二区三区| 免费观看性生交大片5| 18禁在线播放成人免费| 少妇猛男粗大的猛烈进出视频| 久久精品久久久久久久性| 国产免费又黄又爽又色| 久久久久国产网址| 在线看a的网站| 精品久久久久久电影网| 99热网站在线观看| 国产亚洲最大av| 国产成人91sexporn| 久久久久久伊人网av| 伦理电影免费视频| 亚洲高清免费不卡视频| 欧美精品亚洲一区二区| 免费看av在线观看网站| 欧美变态另类bdsm刘玥| 国产成人一区二区在线| 亚洲av欧美aⅴ国产| 精品国产乱码久久久久久小说| 欧美国产精品一级二级三级 | 蜜桃在线观看..| 久久久久国产网址| 日韩欧美 国产精品| 国产极品粉嫩免费观看在线 | 久久免费观看电影| 欧美亚洲 丝袜 人妻 在线| 精品国产乱码久久久久久小说| 一边亲一边摸免费视频| 一区二区三区四区激情视频| 国产精品久久久久久精品电影小说| 国产精品久久久久久精品古装| 新久久久久国产一级毛片| 久久国产精品男人的天堂亚洲 | 婷婷色av中文字幕| 在线观看免费视频网站a站| 人妻制服诱惑在线中文字幕| 精品熟女少妇av免费看| 少妇人妻一区二区三区视频| 中国三级夫妇交换| 久久久久人妻精品一区果冻| 亚洲成人手机| 免费看av在线观看网站| 两个人免费观看高清视频 | 亚洲,一卡二卡三卡| 51国产日韩欧美| 欧美 日韩 精品 国产| 特大巨黑吊av在线直播| 中文字幕久久专区| 99国产精品免费福利视频| 久久人妻熟女aⅴ| 亚洲激情五月婷婷啪啪| 国产一区二区在线观看av| 午夜久久久在线观看| 国产av国产精品国产| 国产成人aa在线观看| 新久久久久国产一级毛片| 99国产精品免费福利视频| 久久热精品热| 97超视频在线观看视频| 久久6这里有精品| 国产片特级美女逼逼视频| 国产日韩欧美亚洲二区| 欧美日本中文国产一区发布| 日韩欧美 国产精品| 一级毛片 在线播放| 亚洲欧美日韩卡通动漫| 亚洲精品第二区| 国产淫语在线视频| 国产在线视频一区二区| 22中文网久久字幕| 视频区图区小说| 多毛熟女@视频| 国产亚洲精品久久久com| 特大巨黑吊av在线直播| 成年女人在线观看亚洲视频| 黑丝袜美女国产一区| 亚洲美女黄色视频免费看| √禁漫天堂资源中文www| 精品卡一卡二卡四卡免费| 午夜激情福利司机影院| 亚洲第一av免费看| 亚洲在久久综合| 国产精品偷伦视频观看了| 国产一区亚洲一区在线观看| 少妇丰满av| 国产亚洲午夜精品一区二区久久| 欧美日韩精品成人综合77777| av播播在线观看一区| 欧美另类一区| 亚洲精品久久午夜乱码| 国产男女内射视频| 最后的刺客免费高清国语| 最近2019中文字幕mv第一页| 日韩一区二区三区影片| 国产高清有码在线观看视频| 免费人妻精品一区二区三区视频| 美女脱内裤让男人舔精品视频| 国产精品99久久99久久久不卡 | 80岁老熟妇乱子伦牲交| 中文字幕制服av| 啦啦啦啦在线视频资源| 另类精品久久| 三上悠亚av全集在线观看 | 欧美精品人与动牲交sv欧美| 日韩av在线免费看完整版不卡| 午夜av观看不卡| 噜噜噜噜噜久久久久久91| 国产精品一区www在线观看| 免费黄频网站在线观看国产| 亚洲电影在线观看av| 日本wwww免费看| 国产黄片视频在线免费观看| av线在线观看网站| 老司机影院毛片| 国产欧美日韩综合在线一区二区 | 久久久久久人妻| 一个人看视频在线观看www免费| 99久久精品热视频| 天堂俺去俺来也www色官网| 国产熟女欧美一区二区| 最近中文字幕高清免费大全6| 亚洲第一av免费看| 亚洲怡红院男人天堂| 精品人妻熟女av久视频| 精品国产露脸久久av麻豆| 亚洲情色 制服丝袜| 国产色爽女视频免费观看| av在线观看视频网站免费| 亚洲婷婷狠狠爱综合网| 精品人妻熟女毛片av久久网站| 自拍欧美九色日韩亚洲蝌蚪91 | 国产男人的电影天堂91| 亚洲图色成人| 美女主播在线视频| 97精品久久久久久久久久精品| 三级经典国产精品| 免费大片18禁| 午夜免费男女啪啪视频观看| 日韩三级伦理在线观看| 老熟女久久久| 少妇被粗大猛烈的视频| 韩国高清视频一区二区三区| av网站免费在线观看视频| 大片免费播放器 马上看| 蜜桃在线观看..| 爱豆传媒免费全集在线观看| 蜜桃久久精品国产亚洲av| 亚洲av国产av综合av卡| 亚洲国产色片| 日本免费在线观看一区| 亚洲色图综合在线观看| kizo精华| 免费av中文字幕在线| 涩涩av久久男人的天堂| 3wmmmm亚洲av在线观看| 三级国产精品片| 亚洲国产精品一区二区三区在线| 一本久久精品| 五月开心婷婷网| 成人特级av手机在线观看| 观看免费一级毛片| 精品一区二区免费观看| 国产精品伦人一区二区| 日本与韩国留学比较| 中文乱码字字幕精品一区二区三区| 免费大片黄手机在线观看| 国产精品免费大片| 国产色婷婷99| 色吧在线观看| 久久精品国产亚洲av天美| 日产精品乱码卡一卡2卡三| 日本欧美国产在线视频| 国产av码专区亚洲av| 亚洲色图综合在线观看| 国产精品99久久久久久久久| 丰满迷人的少妇在线观看| 久久久久久久亚洲中文字幕| 一级毛片 在线播放| 女的被弄到高潮叫床怎么办| 久久99精品国语久久久| 黄色视频在线播放观看不卡| av.在线天堂| 亚洲欧美日韩卡通动漫| 免费看光身美女| √禁漫天堂资源中文www| av线在线观看网站| av网站免费在线观看视频| 日本爱情动作片www.在线观看| 日韩av在线免费看完整版不卡| 亚洲精品乱久久久久久| 王馨瑶露胸无遮挡在线观看| 国产乱人偷精品视频| 少妇高潮的动态图| 国产欧美日韩综合在线一区二区 | 国产乱来视频区| 久久鲁丝午夜福利片| 少妇人妻久久综合中文| 亚洲av免费高清在线观看| 国产精品国产av在线观看| 最黄视频免费看| 日韩免费高清中文字幕av| 亚洲三级黄色毛片| 国产精品福利在线免费观看| av卡一久久| 日韩视频在线欧美| 国产精品福利在线免费观看| 一二三四中文在线观看免费高清| 国产精品99久久久久久久久| 视频中文字幕在线观看| 欧美激情国产日韩精品一区| 欧美性感艳星| 国产精品人妻久久久久久| 精品午夜福利在线看| 久久精品国产a三级三级三级| 69精品国产乱码久久久| 男人狂女人下面高潮的视频| 日本色播在线视频| 另类精品久久| 国产真实伦视频高清在线观看| 麻豆乱淫一区二区| 一区二区三区精品91| 少妇人妻精品综合一区二区| 日韩av不卡免费在线播放| 中文乱码字字幕精品一区二区三区| 边亲边吃奶的免费视频| 简卡轻食公司| 久久久亚洲精品成人影院| av不卡在线播放| 美女脱内裤让男人舔精品视频| 毛片一级片免费看久久久久| h日本视频在线播放| 国产黄片视频在线免费观看| 国产欧美亚洲国产| 黄色怎么调成土黄色| 人人妻人人爽人人添夜夜欢视频 | 亚洲第一av免费看| 一级片'在线观看视频| 中文在线观看免费www的网站| 亚洲美女视频黄频| 伊人久久国产一区二区| 最新中文字幕久久久久| 99热这里只有精品一区| 极品人妻少妇av视频| 精品99又大又爽又粗少妇毛片| 五月伊人婷婷丁香| 亚洲美女搞黄在线观看| 国产成人91sexporn| 99久久人妻综合| 久久久久久久久久久久大奶| 黄色毛片三级朝国网站 | 午夜精品国产一区二区电影| 欧美日韩一区二区视频在线观看视频在线| 成人漫画全彩无遮挡| 中文字幕人妻熟人妻熟丝袜美| 91aial.com中文字幕在线观看| 免费观看性生交大片5| 日韩精品免费视频一区二区三区 | 久久99热6这里只有精品| 日韩欧美 国产精品| 国语对白做爰xxxⅹ性视频网站| 亚洲在久久综合| 午夜激情久久久久久久| 亚洲av不卡在线观看| 插逼视频在线观看| 男女边吃奶边做爰视频| 黑人巨大精品欧美一区二区蜜桃 | 男人添女人高潮全过程视频| 国产精品无大码| 亚洲va在线va天堂va国产| 91精品一卡2卡3卡4卡| 免费播放大片免费观看视频在线观看| 卡戴珊不雅视频在线播放| 精品一品国产午夜福利视频| 日本午夜av视频| 久久久久精品性色| 岛国毛片在线播放| 精品一区在线观看国产| 丁香六月天网| 自拍偷自拍亚洲精品老妇| 亚洲四区av| av有码第一页| 午夜久久久在线观看| 亚洲精品一区蜜桃| 久久99蜜桃精品久久| 日韩亚洲欧美综合| h日本视频在线播放| 我要看日韩黄色一级片| 免费少妇av软件| 亚洲av成人精品一二三区| 少妇熟女欧美另类| av在线观看视频网站免费| 国产69精品久久久久777片| 欧美日韩视频精品一区| 国产爽快片一区二区三区| 久久精品久久久久久噜噜老黄| 亚洲美女搞黄在线观看| 免费不卡的大黄色大毛片视频在线观看| 欧美精品高潮呻吟av久久| 黑人高潮一二区| 久久狼人影院| 成人影院久久| 久久鲁丝午夜福利片| 91精品国产国语对白视频| 成人亚洲精品一区在线观看| 日韩免费高清中文字幕av| 国产成人精品久久久久久| 曰老女人黄片| 在线观看免费日韩欧美大片 | 欧美+日韩+精品| 一区二区三区四区激情视频| 久久久久久久久久人人人人人人| 在现免费观看毛片| 日韩av在线免费看完整版不卡| 国产精品国产三级国产专区5o| 日本av免费视频播放| 男人狂女人下面高潮的视频| 亚洲国产av新网站| 黄色欧美视频在线观看| 精品久久国产蜜桃| 久久这里有精品视频免费| 极品人妻少妇av视频| 亚洲精品456在线播放app| 久久久亚洲精品成人影院| 高清欧美精品videossex| 人妻制服诱惑在线中文字幕| 国产精品女同一区二区软件| 亚洲人成网站在线观看播放| 九色成人免费人妻av| 涩涩av久久男人的天堂| 国产精品秋霞免费鲁丝片| 国产精品久久久久久精品古装| 日韩人妻高清精品专区| 一本一本综合久久| 高清午夜精品一区二区三区| 在线观看免费高清a一片| 伦精品一区二区三区| 黑人猛操日本美女一级片| 99热这里只有是精品50| 亚洲av男天堂| 中文字幕人妻熟人妻熟丝袜美| 久久久久精品久久久久真实原创| 特大巨黑吊av在线直播| 亚洲精品日韩av片在线观看| 另类精品久久| 亚洲美女黄色视频免费看| 亚洲精品第二区| 交换朋友夫妻互换小说| 麻豆成人午夜福利视频| 国产伦理片在线播放av一区| 久久97久久精品| 熟女电影av网| 国产中年淑女户外野战色| 日本午夜av视频| 中文在线观看免费www的网站| 九色成人免费人妻av| 午夜福利,免费看| 嫩草影院入口| 欧美精品人与动牲交sv欧美| 久热久热在线精品观看| 各种免费的搞黄视频| 国产精品国产av在线观看| 啦啦啦中文免费视频观看日本| 欧美日韩综合久久久久久| 欧美日韩av久久| 久久国产乱子免费精品| 国产色爽女视频免费观看| 伦理电影免费视频| 国产黄片美女视频| 欧美日韩av久久| 午夜av观看不卡| 精品亚洲成a人片在线观看| 男女边吃奶边做爰视频| 岛国毛片在线播放| 成年人午夜在线观看视频| 街头女战士在线观看网站| 中文字幕免费在线视频6| 国产精品国产三级国产av玫瑰| 国产午夜精品一二区理论片| 91久久精品国产一区二区成人| 你懂的网址亚洲精品在线观看| 国产精品99久久99久久久不卡 | 六月丁香七月| 精品久久久久久久久av| 欧美日本中文国产一区发布| 日本欧美国产在线视频| 如日韩欧美国产精品一区二区三区 | 熟女人妻精品中文字幕| 久久精品国产自在天天线| 国产日韩欧美在线精品| 中文精品一卡2卡3卡4更新| 久久久久久久亚洲中文字幕| 日韩中文字幕视频在线看片| 丝袜在线中文字幕| 国产在线男女| 又大又黄又爽视频免费| 成年av动漫网址| 少妇人妻久久综合中文| 中文在线观看免费www的网站| 亚洲av男天堂| 又爽又黄a免费视频| 日本av免费视频播放| 国产成人aa在线观看| 亚洲精品,欧美精品| 极品教师在线视频| 人人妻人人澡人人看| 日日啪夜夜爽| 国产男人的电影天堂91| 国产av一区二区精品久久| 91久久精品国产一区二区三区| 中文资源天堂在线| 精品久久久久久久久av| 国产精品麻豆人妻色哟哟久久| 久久久久久久久大av| 国产一区有黄有色的免费视频| 亚洲自偷自拍三级| 亚洲av成人精品一区久久| 亚洲一区二区三区欧美精品| 中文字幕精品免费在线观看视频 | 这个男人来自地球电影免费观看 | 午夜福利,免费看| 少妇人妻一区二区三区视频| 2021少妇久久久久久久久久久| 午夜福利,免费看| 最近最新中文字幕免费大全7| 亚洲国产精品成人久久小说| 亚洲欧美日韩东京热| 亚洲av.av天堂| 女性被躁到高潮视频| 人妻一区二区av| 免费不卡的大黄色大毛片视频在线观看| 99热这里只有是精品在线观看| 欧美区成人在线视频| 丰满迷人的少妇在线观看| 亚洲国产毛片av蜜桃av| 如日韩欧美国产精品一区二区三区 | 热re99久久精品国产66热6| 国产免费福利视频在线观看| 久久国产亚洲av麻豆专区| 97超碰精品成人国产| 日韩大片免费观看网站| 又爽又黄a免费视频| 国产一区亚洲一区在线观看| h视频一区二区三区| 欧美国产精品一级二级三级 | 免费观看性生交大片5| 高清黄色对白视频在线免费看 | av又黄又爽大尺度在线免费看| 内射极品少妇av片p| 老司机影院成人| 欧美bdsm另类| 日韩欧美一区视频在线观看 | 成人毛片a级毛片在线播放| 亚洲人成网站在线播| 男女国产视频网站| 国产av码专区亚洲av| 亚洲av综合色区一区| 久久人人爽人人爽人人片va| 性高湖久久久久久久久免费观看| 久久精品夜色国产| 亚洲国产精品专区欧美| 91精品一卡2卡3卡4卡| 夜夜骑夜夜射夜夜干| 午夜av观看不卡| 国产淫语在线视频| 大片免费播放器 马上看| 2022亚洲国产成人精品| tube8黄色片| 国产高清有码在线观看视频| 天天躁夜夜躁狠狠久久av| 十八禁高潮呻吟视频 | 最近中文字幕2019免费版| 久久精品久久久久久久性| 自拍欧美九色日韩亚洲蝌蚪91 | 男人爽女人下面视频在线观看| 久久精品久久久久久噜噜老黄| 少妇的逼水好多| 午夜福利影视在线免费观看| 亚洲精品日韩av片在线观看| 亚洲丝袜综合中文字幕| 久久午夜福利片| 91aial.com中文字幕在线观看| 精品亚洲成国产av| 欧美另类一区| 久久久久久久久久成人| 有码 亚洲区| 麻豆成人午夜福利视频| 精品酒店卫生间| 高清欧美精品videossex| 在线看a的网站| a级毛色黄片| 日韩精品有码人妻一区| 乱人伦中国视频| 各种免费的搞黄视频| 亚洲精品一二三| 18禁在线无遮挡免费观看视频| 交换朋友夫妻互换小说| 国产一级毛片在线| 我的女老师完整版在线观看| 久热这里只有精品99| 国精品久久久久久国模美| 永久网站在线| 亚洲精品中文字幕在线视频 | 亚洲高清免费不卡视频| 男女啪啪激烈高潮av片| 欧美成人精品欧美一级黄|