• 
    

    
    

      99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

      Optimized Deep Learning Approach for Efficient Diabetic Retinopathy Classification Combining VGG16-CNN

      2023-12-15 03:58:40HebaElHosenyHebaElsepaeWaelMohamedandAymanSelmy
      Computers Materials&Continua 2023年11期

      Heba M.El-Hoseny,Heba F.Elsepae,Wael A.Mohamed and Ayman S.Selmy

      1Department of Computer Science,the Higher Future Institute for Specialized Technological Studies,Obour,11828,Egypt 2Department of Electrical Engineering,Benha Faculty of Engineering,Benha University,Benha,13511,Egypt

      ABSTRACT Diabetic retinopathy is a critical eye condition that,if not treated,can lead to vision loss.Traditional methods of diagnosing and treating the disease are time-consuming and expensive.However,machine learning and deep transfer learning(DTL)techniques have shown promise in medical applications,including detecting,classifying,and segmenting diabetic retinopathy.These advanced techniques offer higher accuracy and performance.Computer-Aided Diagnosis(CAD)is crucial in speeding up classification and providing accurate disease diagnoses.Overall,these technological advancements hold great potential for improving the management of diabetic retinopathy.The study’s objective was to differentiate between different classes of diabetes and verify the model’s capability to distinguish between these classes.The robustness of the model was evaluated using other metrics such as accuracy (ACC),precision (PRE),recall (REC),and area under the curve (AUC).In this particular study,the researchers utilized data cleansing techniques,transfer learning(TL),and convolutional neural network(CNN)methods to effectively identify and categorize the various diseases associated with diabetic retinopathy (DR).They employed the VGG-16CNN model,incorporating intelligent parameters that enhanced its robustness.The outcomes surpassed the results obtained by the auto enhancement(AE)filter,which had an ACC of over 98%.The manuscript provides visual aids such as graphs,tables,and techniques and frameworks to enhance understanding.This study highlights the significance of optimized deep TL in improving the metrics of the classification of the four separate classes of DR.The manuscript emphasizes the importance of using the VGG16CNN classification technique in this context.

      KEYWORDS No diabetic retinopathy(NDR);convolution layers(CNV layers);transfer learning;data cleansing;convolutional neural networks;a visual geometry group(VGG16)

      1 Introduction

      Eye disorders cause different infirmities,such as glaucoma [1],cataracts,DR,and NDR,so early diagnosis and treatment prevent blindness.Glaucoma was expected to affect 64.3 million people worldwide in 2013,rising to 76.0 million in 2020 and 111.8 million in 2040.The diagnosis of glaucoma is frequently put off because it does not manifest symptoms until a relatively advanced state.According to population-level surveys,only 10 to 50 percent of glaucoma patients are conscious that they have the condition.To prevent these diseases,diabetic blood pressure must be kept consistent,and the eye must undergo routine examinations at least twice a year [2].The second class is Cataracts are one of the most common visual diseases,where the irises appear cloudy.People who suffer from cataracts[3]find problems in reading activities,driving,and recognizing the faces of others.The World Health Organization (WHO) estimates that there are approximately 285 million visually impaired people globally,of whom 39 million are blind,and 246 million have moderate to severe blindness[4].Better cataract surgery has emerged in recent years than it did in the preceding.In patients without ocular complications like macular degeneration(DR)or glaucoma[5],85%-90%of cataract surgery patients will have 6-12 best-corrected vision[6].The third class is DR[7],categorized into two retinal disease cases:proliferative DR(PDR)and non-proliferative DR(NPDR).A low-risk form of DR is called NPDR,where the Blood vessel membranes in the retina are compromised.The retina’s tissues could expand,causing white patches to appear.On the other hand,the high risk,called PDR,High intraocular pressure may cause the blood tissues to have difficulty transferring fluid to the eye,which destroys the cells responsible for transmitting images from the retina to the cerebrum[8].The fourth class is NDR or typical case.

      The paper’s main contributions are distinguished from others by employing data cleansing techniques,TL,and intelligent parameter techniques.Various cleansing techniques are implemented to compare their performance in different applications,involving replacing,modifying,or removing incorrect or coarse data.This enhances data quality by identifying and eliminating errors and inconsistencies,leading to improved learning processes and higher efficiency.Two enhancement filters,auto enhancement(AE)and contrast-limited adaptive histogram equalization(CLAHE),are applied in steps such as augmentation and optimization using adaptive learning rates with different rates for each layer.

      This work’s primary contribution lies in analyzing and balancing datasets as a cleansing step to achieve high-quality feature extraction.This is accomplished through the application of AE and CLAHE filters.Subsequently,the VGG16 and CNN algorithms are employed with varying dropout values.Two evaluation stages are conducted:the first measures the performance of cleansed images using a MATLAB program,indicating that AE outperforms CLAHE in metrics.The second evaluation involves training the algorithm and compiling metrics such as AUC,ACC,PRE,REC,confusion matrix(CM),and loss curves,reaffirming the findings of the first evaluation.AE is identified as the superior enhanced filter in this manuscript,achieving metrics not previously obtained with the same database.

      The paper is structured as follows: The first section presents related work,followed by an explanation of deep TL,CNN,and the employed cleansing filters with corresponding mathematical equations.Section 3 discusses the VGG16 and CNN algorithms used to divide the original DR images into four groups.Simulation results and visual representations are included.Finally,the paper concludes and outlines potential future work.

      2 Related Works

      The classification of ophthalmological diseases has been the subject of numerous research proposals.We conducted a literature review to determine the primary methods for glaucoma (GL),cataract (CT),DR,and NDR diagnoses based on images.For a deeper understanding of the issue and to brainstorm workable solutions for raising the ACC of our TL model.We looked at recent journals and publications.After these steps,we reached the goal of using an open-source and freely downloaded dataset(s) and examined a model to compare our efforts with previous experiments.This study introduced an ensemble-based DR system,the CNN in our ensemble(based on VGG16),for different classes of DR tasks.We gained evaluation metrics that are confined in ACC that were achieved with the help of the enhancement filters(98.79%CNN with dropout 0.02 and the used filter auto-enhancement filter with loss 0.0390) and outcomes of CLAHE (96.86% TL without dropout with loss 0.0775).In Ahmed et al.’s research on cataracts,CNN with VGG-19 was applied,where the ACC was 97.47%,with a PRE of 97.47%and a loss of 5.27%[9].Huang et al.implemented a semisupervised classification based on two procedures named Graph Convolutional Network(GCN)and CNN,where They scored the best ACC compared to other conventional algorithms with an ACC of 87.86%and a PRE of 87.87%[10].The Deep Convolutional Neural Network(DCNN)was set up by Gulshan et al.[11]to recognize DR in retinal fundus images,and a deep learning algorithm was applied to develop an algorithm that autonomously detects diabetic macular edema and DR in retinal fundus images.The main decision made by the ophthalmologist team affected the specificity and sensitivity of the algorithm used to determine whether DR was moderate,worse,or both.DCNN,with a vast amount of data in various grades per image,was used to create the algorithm with 96.5%sensitivity and 92.4%specificity.Kashyap et al.[12],using two CNNs,implemented a TL technique.Li et al.[13]created another TL model for categorizing DR into four classes:normal,mild,moderate,and severe.This was in addition to applying TL in two ways while using baseline approaches such as Alex Net,Google Net,VGGNet-16,and VGGNet-19[14].The TL model classified the optical coherence tomography (OCT) images for the diseases resulting from the diabetic retina.Kermany et al.[15],who also produced Inception-v3[16],carried out this novel.Their approach was trained,tested,and validated using OCT images from four different categories: Choroidal Neovascularization,Diabetic macular edema,diabetic drusen,and NORMAL.Additionally,they tested the effectiveness of their strategy using 1,000 randomly chosen training samples from each category.Lu et al.[17]also described the TL technique for diagnosing DR using OCT images.They classified five classes from the OCT data sets[18].Kamal et al.[19]used five algorithms(standard CNN,VGG19,ResNet50,Dense Net,and Google Net).They concluded that the best metric measurements from the VGG19 algorithms with fine-tuning were 94% AUC.Sensitivity and specificity were 87.01% and 89.01%,respectively.Ahn et al.[20] presented CNN (consisting of three CNVs and Maxpooling in every layer and two fully connected layers in the classifier).The researchers and their colleagues used a private dataset of 1,542 images.They achieved ACC and an AUC of 87.9%and 0.94 on the test data,respectively.The authors in [21-24] used the algorithms (VGG16,VGG16 with augmentation (they used techniques like mirroring and rotating),VGG16 with dropout added to the architecture,and two cascaded VGG16)to get the highest ACC.Pratt et al.[25]added a CNN to this architecture of preprocessing methods for classifying micro-aneurysms,exudates,and hemorrhages on the retina.They fulfilled Sensitivity(95%)and ACC(75%)on 5000 validation images.Islam et al.[26]conducted experiments on eight retinal diseases using CLAHE as a pre-processing step and CNN for feature extraction.Sarki et al.[27] submitted the CNN-based architecture for the dual classification of diabetic eye illness.Diabetic eye disease of several classes’levels: With VGG16,the maximum ACC for multiclassification is 88.3%,and likewise,for mild multi-classification,it is 85.95%.Raghavendra et al.[28]built a CNN(which included four CNV layers and applied batch normalization,one ReLU,and one Max-pooling at each layer before adding fully connected and Soft-max layers).They can hit an ACC of around 98.13%using 1426 fundus photos from a specific dataset.Recently,presented classification and segmentation approaches for a retinal disease were used to increase the classification ACC.A novel tactic has been posited to boost the fineness of the retinal images (enhancement techniques)before the steps of the classifier.Simulation outcomes for that task reached 84%ACC without fuzzy enhancement,but when applying fuzzy,the ACC reached 100%.This meant that fuzzy was important to discriminate between the different types of retinal diseases[29].This paper proposed a model for Ocular Disease Intelligent Recognition (ODIR),using augmentation techniques to achieve balance in different datasets.As a result,the ACC of each disease in multi-labeled classification tasks was improved by making better images and working with different TL algorithms [30].In the paper by Sultan A.Aljahdali,Inception V3,Inception Res-Net V2,Xception,and DenseNet 121 were used as a few examples of pre-trained models that were applied and provided CAD ways of CT diagnosis.The Inception ResNetV2 model had a test ACC of 98.17%,a true positive rate of 97%,and a false positive rate of 100% for detecting eye disease [31].Nine models (ResNet50,ResNet152,VGG16,VGG19,AlexNet,Google Net,DenseNet20,Inception v3,and Inception v4) served as the foundation for Kamal et al.’s survey paper,which had the best ACC of 0.80.In this article,the dataset is categorized into three ocular illnesses(Strabismus,DR,and GL),where various techniques are used,such as TL,DL,and ML approaches[19].This review study is based on five retina classes sorted by severity(No DR,Mild,Moderate,Severe,and Proliferate).The distribution of that paper is presented in terms of three model-supervised,self-supervised,and transformer models that achieved percentages of 91%,67%,and 50%,respectively [32].Li et al.presented the enhancement methods through two filters.The first filter,Adaptive Histogram Equalization(AHE),enhances the contrast between four classes of images (about 7935 images).The second filter was a nonlocal mean,eliminating the noise.The results of these two filters were used as input measurements of metrics performance:ACC,specificity,and sensitivity (94.25%,94.22,and 98.11).Consecutively [33],different databases of more than two diseases were used and gave acceptable validation ACC for the training process.This process was checked by five different versions of the AlexNet CNN(Net transfer I,Net transfer II,Net transfer III,Net transfer IV,and Net transfer V): 94.30%,91.8%,89.7%,93.1%,and 92.10% successively[34].Sharif A.Kamran presented the generative adversarial network(VTGAN).Vision transformers depend on a constructive adversarial network (GAN) made up of transformer encoder blocks for discriminators and generators,as well as residual and spatial feature fusion blocks with numerous losses,to provide brilliant fluorescein angiography images from both normal and abnormal fundus photographs for training.When network metrics were measured on a vision transformer using the three common metrics criteria ACC,sensitivity,and specificity,the scores were 85.7%,83.3%,and 90%,respectively[35].

      3 Materials and Methods

      3.1 Data Cleansing(Retinal Image Enhancement)

      The diagnosis of DR can be performed manually by an ophthalmologist or through CAD systems.Retinal images are typically captured using a fundus camera.However,several parameters can influence the quality of these diagnostic images,such as eye movements,lighting conditions,and glare.Image quality is critical in the classification,segmentation,and detection processes.Any abnormalities or malformations in the fundus images can have a negative impact on the ACC of the diagnosis.The presence of noise in the images can lead to a decrease in the evaluation metrics used to assess the performance of the diagnostic model.To address these issues,it is crucial to cleanse the datasets of images by removing any malformations or artifacts.The present study uses AE and CLAHE filters to clean data.This cleansing process aims to enhance the images’quality and improve the diagnostic model’s ACC.

      3.1.1 CLAHE

      The filter employs a clipping level in the histogram to determine the intensity of the local histogram mapping function.This helps reduce undesired noise in the retinal image[36].Two essential variables in CLAHE are the block size and clip limit,which are used to adjust the image quality.The clip limit is calculated using the following Eq.(1):

      β: the clip limit;Smax: maximum of the new distribution;N: dimensions of the retinal image;L:level of the image;M represents the area size,N represents the grey-level value(256),and represents the clip factor,which expresses the addition of a histogram limit with a value of 1 to 100[37].

      3.1.2 AE Filter

      This is the second filter to be used in our proposal.Which increases the brightness overall in the image[38].It is simply lightness without any light.The luminosity scale is from(0 to 100),where 0 is no light(black),and 100 is white.The arithmetic means of red,green,and blue can be said to be bright.We can recognize the AE function in the MATLAB code by changing the following three parameters to get the highest quality of enhanced images:In the brightness of various objects or areas.Adjust the relative amounts of dark and light regions of Fundus photos using If you increase the contrast,then the light-colored object will be brighter,and the dark-colored object will be darker.You can say it is the difference between the lightest color object and the darkest color object as an arithmetic Eq.(2):

      3.2 Enhanced Data Evaluation

      After applying the MATLAB code to enhance the quality of the DR classes using both the CLAHE and AE filters,we conducted further analysis to measure the performance of these filters.A custom MATLAB code function was developed to calculate evaluation metrics based on the outcomes of the filters.Table 1 presents the results of these evaluation measurements,including entropy fused,average gradient,edge intensity,and local contrast.From the table,it can be observed that the AE filter achieved higher values for these metrics compared to the CLAHE filter.The higher values obtained for the evaluation measurements suggest that the AE filter was more successful in improving the quality of the DR classes.

      Table 1: Comparison of enhanced outcomes metrics by CLAHE and AE filter

      3.3 Convolutional Neural Network

      CNNs have recently gained significant popularity,particularly for image classification tasks.These deep learning algorithms utilize learnable weights and biases to analyze fundus images and differentiate between them.CNNs[39,40]have multiple input,output,and hidden layers.They have been successfully applied in various computer vision applications[41],such as semantic segmentation,object detection,and image classification.One notable advantage of CNNs is their ability to recognize essential features in images without human supervision.Additionally,the concept of weight sharing in CNNs contributes to their high ACC in image classification and recognition.CNNs can effectively reduce the number of parameters that must be trained while maintaining performance.However,CNNs do have some limitations.They require a large amount of training data to achieve optimal performance.Furthermore,the training process of CNNs can be time-consuming,especially without a powerful GPU,which can impact their efficiency.CNNs are designed as feed-forward neural networks[42]and incorporate filters and pooling layers for image and video processing[43].

      3.3.1 Convolution Layers

      CNV layers are the main building blocks of CNN,which include output vectors like a feature map,filters like a feature detector,and input vectors like an image.After going through a CNV layer,the image is abstracted to a feature map,sometimes called an activation map.Convolution occurs in CNNs when two matrices consisting of rows and columns are merged to create a third matrix.This process repeats with an accurate stride(the step for the filter to move).Doing so decreases the system’s parameters,and the calculation is completed more quickly[44].The output filter size can be calculated as a mathematical Eq.(3).This layer has several results.The first is an increase in dimensionality accompanied by padding,which we can write in the code as“same padding”.The second decreases the dimensionality,and in this process,it happens without padding and is expressed as“valid padding”.Each pixel in the new image differs from the previous one depending on the feature map[45].

      Feature Map=Input Image×Feature Detector

      W:the size of the Input image,f:the size of the CNV layer filters,p:padding of the output matrix,S:stride.

      3.3.2 Activation Function

      Activation functions in a CNN model are crucial in determining whether a neuron should be activated based on its input.They use mathematical processes to assess the significance of the information.In the hidden layers of the model,the Rectified Linear Unit(ReLU)activation function is commonly employed.ReLU helps address the vanishing gradients issue by ensuring that the weights do not become extremely small[46].Compared to other activation functions like tanh and sigmoid,ReLU is computationally less expensive and faster.The main objective of an activation function is to introduce nonlinearity into the output of a neuron.In the suggested framework,the SoftMax function is used for making decisions.Softmax is a straightforward activation function that produces outcomes from 0 to 1,as illustrated in Eq.(4).This activation function is often used for classification tasks[47].

      Max(0,x);if x is positive,output x,otherwise 0;+∞Range:0 to+∞.

      3.3.3 Pooling Layer

      This strategy is applied to reduce the dimensions of the outcomes of previous layers.There are two types of pooling:maximum pooling and average pooling.Max pooling removes the image noise,but average pooling suppresses noise,so max pooling performs better than average pooling [48].In our model,we introduced max pooling with dropout to prevent overfitting during model training(When a neural network overfits,it excels on training data but fails when exposed to fresh data from the issue domain).

      3.3.4 Fully Connected Layer(FC)

      The input picture from the preceding layers is flattened and supplied to the FC layer.The flattened vector then proceeds through a few additional FC levels,where the standard processes for mathematical functions happen.The classification process starts to take place at this level.The Softmax activation function is applied in FC to decide on the classification technique[49].

      3.4 Transfer Learning

      TL is a crucial component of deep learning[22,50],concentrating on storing information obtained while solving an issue and applying it to other closely related problems.TL increases the effectiveness of new training models by doing away with the requirement for a sizable collection of labeled training data for each new model.Further benefits of TL include faster training times,fewer dataset requirements,and improved performance for classification and segmentation detection problems.VGG16 is the most widely utilized transfer network that we employed in our article(VGG),which proved its effectiveness in many tasks,including image classification and object detection in many different tasks.It was based on a study of how to make these networks deeper.

      3.5 Implemented Data Sets

      Our proposal utilized datasets divided into four categories:C,GL DR,and NDR.These datasets,consisting of 4271 images,were obtained from the Kaggle dataset[51].The images in the dataset are in colored format,red,green,blue(RGB),include both left-and right-eye images.The purpose of using these datasets was to classify the images into the categories mentioned above,and the distribution of the images among the classes is summarized in Table 2.

      Table 2: The implemented data sets

      3.6 The Proposed Algorithm of TL and CNN Architecture

      The algorithm used is the VGG16 model;the current work suggests modifying the VGG[52,53]model to get better outcomes and achieve better results.In VGG16,only the ImageNet dataset was used for pre-training the model.VGG16 has fixed input tensor dimensions of 224 ×224 with RGB channels.This model is passed through many convolutional neural networks(CNV)layers,where the most miniature used filters were 3 × 3.The most important thing that distinguishes the algorithm of TL is that it does not need many hyperparameters.They used 3 × 3 CNV layers with stride one and Maxpooling (2 × 2) with stride 2.They consistently employed the same padding.Convolution and Maxpooling are organized in the model with block-CNV layers having 64 filters,block-2 CNV layers having 128 filters,block-3 CNV layers having 256 filters,block-4 CNV,and block-5 CNV5 having 512 filters.This task starts with identifying the input RGB images,whose dimensions are 224×224,but the images in the database have different sizes.256×256 for CT with a size of 8.84 KB,DR with 46.9 KB,GL with 10.5 KB,and 224 × 224 with a size of 63.8 KB for NDR.The VGG16 models were scaled down to 200 × 200 pixels.The four classes of datasets were used to perform classification procedures by CNN models based on deep VGG16.To prepare for the classification,we should balance the datasets into groups of approximately 1000 for four classes to prevent overfitting or underfitting.After balancing the data,two enhancement filters—the AE filter and the CLAHE filter—performed data cleansing steps.Data cleansing removes noise from original images using a MATLAB program after enhancing,resizing,and reshaping the original images.Image resizing is a necessary process because of the scale difference between images.The image augmentation approach increases the training dataset’s size and improves the model’s capacity.Augmentation in this algorithm occurs in preprocessing datasets;various augmentation process types are used in three steps: Zoom range=0.15,rotation range=20,and fill_mode=nearest[54].The augmentation procedure aims to prevent or minimize overfitting on a small quantity of data[55].

      The first model used CNN with a dropout of 0.02 and without TL,which helped improve the classification network and prevent overfitting [56].A common problem occurs when training data in CNN is insufficient;this technique is presented in [57].The second is CNN.The third is CNN,with a dropout of 0.2.These three architectures apply in both cases of data division(80%,20%)and(90%,10%).It is the same for VGG16;these three cases apply to our model.For the training process,Adaptive Moment Estimation(Adam optimizer)gives the highest results[58],in which the network parameters were optimized.Compared to other optimizers,Descent with Momentum and Root Mean Square Propagation (SGDM and MSPROP) in terms of ACC and loss,Adam is the best optimizer presented in [59].Datasets were used for training in this scenario with the following parameters:training epoch of 20,batch size of 32,and learning rate (set by the Adam optimizer according to a model designed and changed automatically in the program to fit the training model).Loss in the form of categorical cross entropy;these parameters are shown in Table 3.The final process is to classify the test data and predict the output.Classes initialize the VGG16 fit model and extract the model’s statistical evaluations,such as the CM,ACC,PRE,REC,AUC,and test loss.The block diagram of the proposed algorithm is introduced in Fig.1.

      Figure 1:The proposed algorithm

      Table 3: Optimized parameters of proposed models

      Table 5: Highlights the evaluation metrics for CNN models for 90%training and 10%testing

      Table 6: Performance analysis of the three case architectures of VGG16

      Table 7:Performance analysis of the three cases architectures of VGG16 for the 2nd section of datasets

      3.7 Evaluation Metrics

      According to our models,we observed that the AE filter had excellent ACC,and the incorrectly categorized instances for every class were small.The metrics evaluation depends on four essential measurements ACC,PRE,REC,and AUC.We want to accomplish these objectives with our methodology;however,false predictions must be avoided.Our study’s measurement performance can benefit from the CM because it makes it easy to compare the values of four indexes: True Positive(TP),True Negative(TN),False Positive(FP),and False Negative(FN)[60].

      ? True Positive(TP):When the predicted value and actual value are the same and the predicted values are positive,the expected model values are also positive.

      ? True Negative (TN): The expected and actual values are identical;besides the real value is negative,the model’s predicted value is negative.

      ? False Positive (FP): The predicted value is false;the actual value is negative,and the model’s expected value is positive.

      ? False Negative(FN):The predicted value is false;the actual value is positive,and the model’s expected value is negative.

      ACCplays a pivotal role in evaluating the metrics;it is the ratio of the sum of true positives and true negatives to the total number of samples.It can be determined from the following Eq.(5):

      PREis the total number of positive predictions (total number of true positives) divided by the total number of expected positives of class values(total number of true positives and false positives).Eq.(6)serves as an example of this.

      RECis the number of True Positives (TP) divided by the number of True Positives and False Negatives (FN),and another name for REC is sensitivity.It can be measured from the arithmetic Eq.(7):

      4 Results and Simulation Graphs

      This section examines how well the classification model works by looking at how CNN and VGG16 work when the training parameters for intelligence are changed.We evaluated the results of the model using the Kaggle dataset.The implementation utilized Tensor Flow and Keras,with Keras serving as a deep machine learning package and Tensor Flow acting as the backend for machine learning operations.A CNN model was employed for classification experiments.The model consisted of four CNV layers followed by a max-pooling process.Additionally,for the fully connected layer,the original Dense layers are removed and replaced with two Dense Layers,two with 1024 nodes,and the final one with 4 for classification.The ReLU was applied to all layers.The experiments were conducted in two scenarios:with or without dropouts.In addition,the change in the enhanced dataset division Tables 4 and 5 illustrate the metrics outcomes for CNN models,and Fig.2 illustrates CNN architecture.The formation of the VGG16 architecture is five blocks of 3×3 convolutions,followed by a max pool layer used during the training phase.To further mitigate overfitting,a dropout of 0.2 and 0.02 was applied to the output of the last block.Dropout is a regularization technique that randomly sets a fraction of the input units to zero during training,which helps prevent over-fitting on specific features.After the dropout layer,a dense layer consisting of 1024 neurons was added.The dense layer is fully connected,allowing for more complex interactions between the features extracted by the convolutional layers.Finally,the output layer is dense with four outputs,each corresponding to a specific category of the DR images.This configuration enables the model to classify input images into DR categories.This algorithm is illustrated in Fig.3,and the results are in Tables 6 and 7.

      Figure 2:CNN architecture

      Figure 3:VGG16 architecture

      4.1 Concluded Results

      After analyzing different cases of CNN architecture and VGG16,it was observed that the results obtained using the AE filter were the most favorable in terms of metrics.Table 8 presents the ACC of 98.7%achieved when employing the AE filter.

      Table 8:Concluded the best performance of the proposed algorithm and with VGG16 of dropout 0.02 for AE,80%;20%dataset

      4.2 CM

      The enhanced database with the AE filter exhibited the highest performance improvement among the cases studied.Furthermore,the CM in Fig.4 and the AUC in the figure provided additional insights into the classification outcomes.

      4.3 Receiver Operating Characteristic(ROC)Curve

      Plotting the ROC Curve is a trustworthy way to evaluate a classifier’s classification ACC.The True Positive Rate(TPR)and False Positive Rate(FPR)charts allow us to observe how the classifier responds to various thresholds.The closer the ROC curve comes to touching the upper left corner of the picture,the better the model performs in categorizing the data.We may compute the AUC,which shows how much of the graphic is below the curve[61,62].The model becomes more accurate as the AUC approaches the value of 1.The figure below shows the combined dataset’s AUC score and ROC curve after being tested on four classes of DR.The black,diagonally dashed line shows the 50%area.According to the graphic,the combined model with the VGG16 and CNN of AE filters and AUC performs better at classifying DR and typical retinal pictures.VGG16CNN has AUC curves,as shown in the below Fig.5.

      Figure 4:CM of the proposed algorithms

      Figure 5:The ROC curve of the VGG16 and CNN evaluated on four datasets.(0)CT,(1)GL,(2)DR,(3)NDR

      4.4 Comparisons

      By comparing the obtained results from the proposed framework with other research[29,61],It can be concluded that the proposed model in this paper has achieved better ACC than the others.This is presented in Table 9.

      Table 9: Comparison between the proposed framework and other algorithms

      5 Conclusion

      This study evaluated the performance of four distinct datasets related to eye conditions:GL,CT,DR,and NDR.This article started by performing data cleansing to ensure the quality of the datasets.Afterward,they prepared the classes for initializing algorithms based on the VGG16CNN architecture.In this work,various parameters were adjusted and experimented with.The datasets were divided into training,validation,and testing sets using different ratios,such as 90%for training,5%for validation,and 5%for testing,or 80%for training,10%for validation,and 10%for testing.Dropout values,which help prevent overfitting,were set to 0.02 and 0.2.Multiple architectures were implemented and tested,leading to variations in the experimental setup.

      The dropout was applied both with and without the CNN architecture or VGG16.To train and test the network for classifying the enhanced classes,a DTL approach was employed in this study.We used TL techniques to leverage pre-trained models and improve classification performance.The proposed model showed promising results when using AE-enhanced classes in combination with TL and CNN models.The achieved metrics included an ACC of 98.62%,an SPE of 98.65%,and a REC of 98.59%.The authors suggest several improvements further to enhance the model’s ACC for future work.One recommendation is to expand the dataset by adding new distinctive classes related to eye conditions.Increasing the diversity and size of the dataset can help the model generalize better and improve its performance.Additionally,incorporating new TL techniques beyond the ones used in this study may enhance the model’s capabilities and overall performance.

      Acknowledgement:The authors thank the Department of Electrical Engineering,Faculty of Engineering,Benha University,for providing intellectual assistance.

      Funding Statement:The authors received no specific funding for this study.

      Author Contributions:The authors confirm their contribution to the paper as follows:study conception and design: Heba F.Elsepae and Heba M.El-Hoseny;data collection: Ayman S.Selmy and Heba F.Elsepae;analysis and interpretation of results: Heba F.Elsepae,Heba M.El-Hoseny and Wael A.Mohamed;draft manuscript preparation: Ayman S.Selmy and Wael A.Mohamed.All authors reviewed the results and approved the final version of the manuscript.

      Availability of Data and Materials:The datasets generated during and/or analyzed during the current study are available from the corresponding author upon reasonable request.

      Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

      贺兰县| 枣阳市| 正安县| 平乡县| 东城区| 乐业县| 阿尔山市| 雷州市| 皮山县| 乌拉特后旗| 仪陇县| 门头沟区| 蒙自县| 房山区| 华阴市| 和田县| 仁化县| 蕉岭县| 增城市| 宜阳县| 许昌市| 潢川县| 长丰县| 木里| 青冈县| 大丰市| 札达县| 阳江市| 织金县| 上高县| 财经| 丹棱县| 奉化市| 通渭县| 邹平县| 大城县| 阳信县| 怀化市| 平顺县| 恩施市| 青川县|