• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Convolution-Based Heterogeneous Activation Facility for Effective Machine Learning of ECG Signals

    2023-12-12 15:49:18PremanandandSathiyaNarayanan
    Computers Materials&Continua 2023年10期

    Premanand S.and Sathiya Narayanan

    School of Electronics Engineering,Vellore Institute of Technology,Chennai,600127,India

    ABSTRACT Machine Learning(ML)and Deep Learning(DL)technologies are revolutionizing the medical domain,especially with Electrocardiogram(ECG),by providing new tools and techniques for diagnosing,treating,and preventing diseases.However,DL architectures are computationally more demanding.In recent years,researchers have focused on combining the computationally less intensive portion of the DL architectures with ML approaches,say for example,combining the convolutional layer blocks of Convolution Neural Networks(CNNs)into ML algorithms such as Extreme Gradient Boosting(XGBoost)and K-Nearest Neighbor(KNN)resulting in CNN-XGBoost and CNN-KNN,respectively.However,these approaches are homogenous in the sense that they use a fixed Activation Function(AFs)in the sequence of convolution and pooling layers,thereby limiting the ability to capture unique features.Since various AFs are readily available and each could capture unique features,we propose a Convolutionbased Heterogeneous Activation Facility(CHAF)which uses multiple AFs in the convolution layer blocks,one for each block,with a motivation of extracting features in a better manner to improve the accuracy.The proposed CHAF approach is validated on PTB and shown to outperform the homogeneous approaches such as CNN-KNN and CNN-XGBoost.For PTB dataset,proposed CHAF-KNN has an accuracy of 99.55%and an F1 score of 99.68%in just 0.008 s,outperforming the state-of-the-art CNN-XGBoost which has an accuracy of 99.38%and an F1 score of 99.32% in 1.23 s.To validate the generality of the proposed CHAF,experiments were repeated on MIT-BIH dataset,and the proposed CHAF-KNN is shown to outperform CNN-KNN and CNN-XGBoost.

    KEYWORDS Electrocardiogram;convolution neural network;machine learning;activation function

    1 Introduction

    It is well known that many diseases affect our health in the human ecosystem.According to World Health Organization [1],there are some shocking statistics: Due to heart-related problems,around 17.9 million people die each year,which accounts for 32% of deaths globally.In this,more than 75% of death occurs in low and middle-income countries like India,Malaysia,Pakistan,Sri Lanka,Thailand,Indonesia,and many more.An Electrocardiogram(ECG)is a non-invasive medical diagnostic device that cardiologists and medical practitioners predominately use to record the heart’s physiological activity throughout time.

    The morphological characteristics of the ECG signal are shown in Fig.1 [2],which are pivotal in heart issue identification.The data from the device may be in a signal or image format,so the consultant can see the heart’s electrical activity and check for irregularity in the interval and any pattern change with morphology.Any changes give information about the severity condition of a particular patient.It should note that since heart-related disorders account for nearly one-third of all fatalities worldwide,accurate diagnosis is crucial.

    Figure 1:Morphological features of an ECG waveform

    In recent years,there has been a rapid advancement in portable ECG monitors for the medical field,including the Holter monitor[3],event monitor,and wearable technology in several healthcare areas.As a result,the growth of the studied ECG data has outpaced the capacity of human cardiologists.Moreover,solid mathematical foundations provide greater accuracy than manual calculation,where Artificial Intelligence(AI)technologies like[2,4,5]come into play.Automation and accuracy in analyzing ECG data has become fascinating recently.

    Machine Learning(ML)[6–9]has become increasingly important in various fields due to its many advantages over manual calculations.ML algorithms can process large amounts of data much faster than humans,allowing for more efficient and speedy decision-making.They can also scale to handle large volumes of data and analyze it in real-time,making it possible to analyze data from multiple sources with varying levels of granularity.Additionally,ML algorithms are less prone to errors and can provide consistent results over time,leading to improved accuracy and reliability of data analysis.These algorithms can also detect complex patterns and relationships in data that may be difficult or impossible for humans to see,enabling businesses and organizations to identify new opportunities,optimize processes,and make more informed decisions.Overall,ML has the potential to automate and optimize many aspects of data analysis,saving time and resources,improve accuracy and consistency,and enable more effective decision-making.

    Deep Learning (DL) [10–12] techniques are increasingly important in healthcare and medical signal analysis.DL is a subset of ML that uses Neural Networks(NN)to learn from large amounts of data.In medical signal analysis,DL architectures can be trained on vast amounts of medical data to automatically detect patterns and abnormalities that may indicate certain health conditions.DL can help doctors identify patients at high risk of developing a particular state,allowing for early intervention and preventative treatment.DL architectures can also personalize treatment based on a patient’s medical signals,predicting which treatments are most likely effective for a particular patient.DL can improve patient outcomes and reduce the risk of adverse side effects.DL architectures in healthcare and medical signal analysis can significantly improve patient outcomes,reduce healthcare costs,and revolutionize how we diagnose and treat various health conditions.

    Convolutional Neural Network(CNN),a typical DL architecture,comprises of multiple layers:convolution layers,pooling layers,and fully connected layers[13–15].Convolution layers use filters to capture local patterns and features from the input data.Pooling layers down-sample the filtered feature maps and preserving essential features.Max-pooling is widely used in CNNs.A convolution layer with an Activation Function(AF)followed by a pooling layer is termed as a convolutional layer block.In CNNs,there will be several convolution layer blocks followed by a fully connected NN layer.The output of the last convolutional layer block is flattened into a vector representation and fed into fully connected layers for high-level feature extraction.The fully connected layer is the computationally demanding portion of the CNN,and this leads to high execution time.

    For biomedical applications such as ECG signal classification,ML algorithms are consistent and reliable for fast decision making (i.e.,quick predictions) but they do not match the accuracy levels of DL architectures.Conventional ML algorithms often rely on shallow models with limited ability to learn complicated data patterns.DL architectures on the other hand are highly accurate but time consuming.DL models employ NN with numerous layers that may extract increasingly abstract aspects from data,allowing DL models to learn more complicated data representations,leading to higher performance in many applications.Additionally,DL models can learn directly from raw data,eliminating the need for manual feature engineering that is often required in traditional ML models.To balance these tradeoffs,researchers considered integrating CNN with ML.They focused on combining the computationally less intensive portion of the DL architectures with ML approaches,say for example,combining the convolutional layer blocks of CNN with ML algorithms such as KNN resulting in CNN-KNN[16].

    Although the approaches integrating CNN with ML algorithms are faster compared to the baseline CNN approach,these approaches are homogenous in the sense that they use a fixed AFs in the sequence of convolution and pooling layers,thereby limiting the ability to capture unique features.Since various AFs are readily available and each could capture unique features,in this work,we propose a CHAF which uses multiple AFs in the convolution layer blocks,one for each block,with a motivation of extracting features in a better manner to improve the accuracy.

    The contributions of this manuscript are as follows:

    ? A novel framework for feature extraction,termed as CHAF,is proposed.The CHAF comprises of several convolution layer blocks,each using different AFs.Due to the heterogeneous nature of CHAF,the features are extracted in an efficient manner,thereby give the model a more robust capacity for learning ECG signals.

    ? The proposed approach,CHAF-KNN(i.e.,CHAF integrated with KNN),is validated on two ECG datasets(PTB dataset and MIT-BIH dataset)and shown to outperform the state-of-theart approaches.In the case of PTB,CHAF-KNN has an accuracy of 99.55%and an F1 score of 99.68%in just 0.008 s.In the case of MIT-BIT which is relatively larger than PTB,the proposed CHAF-KNN has an accuracy of 99.08%and an F1 score of 98.99%in just 0.07 s.

    The remainder of this manuscript is organized as follows.Section 2 reports the survey of different architectures related to ML and DL.In Section 3,we expound on the proposed methodology stepby-step.Section 4 presents the experimental validation of the proposed approach and discusses its advantages and limitations.Finally,Section 5 concludes the paper with recommendations for future research.

    2 Literature Survey

    The diagnostic analysis of the PTB ECG dataset,a publicly available dataset containing a comprehensive collection of ECG recordings from individuals with diverse cardiac diseases,makes it ideal for different AI algorithmic models.Many ML and DL architectures are performed on the PTB dataset for better accuracy [2].The KNN algorithm,termed as a lazy learning approach,builds a model by storing feature vectors and class labels from the training dataset.To classify a new data point,KNN calculates distances between the new feature vector and the labeled samples,using Euclidean distance in this case.The K nearest neighbors are identified,and majority voting is employed to assign the new data point to the class label that is most frequent among the K neighbors.Tree-based ML algorithmic models report high accuracy like 97.88% for Random Forest (RF),98.23%for Extreme Gradient Boosting(XGBoost),and 98.03%for Light Gradient Boosting Machine(LightGBM)and outperform other ML algorithms,and their performance is comparable to the highend DL architectures[2].Apart from tree-based algorithms,other ML algorithms like Support Vector Machine(SVM)give better accuracy like 96.67%with morphological&temporal features[17],98.33%with 220 features[18]&100%[19],along with grasshopper optimization algorithm,99.9%for cubic models&99.8%for a quadratic model[20],95.1%for Adaptive boosting(AdaBoost)[17]algorithm,and 94.9%for Logistic Regression[21].

    CNN architecture [13–15] is an advanced architecture type utilized in DL,especially for image and signal processing tasks.CNNs are one example of advanced architectures in DL,which can refer to any NN architecture that goes beyond the fundamental feed-forward architecture.CNNs are highly suited for tasks like image identification and classification since they automatically learn and detect complicated aspects in the input.Convolutions and pooling operations are frequently used in the architecture of a CNN to extract and compress the most crucial elements from the input data.

    Advanced DL architectures can use Recurrent Neural Networks(RNNs),attention mechanisms,and adversarial training,to name just a few.These models are created to address specific issues,such as modeling sequential data,improving model resilience,and providing realistic images or other sorts of data.In general,and particularly for image and signal processing applications,CNNs are a prominent and often-used example of sophisticated DL architectures.However,various cutting-edge designs can be used with varied data or applications.

    Regarding DL architecture,the PTB dataset was the subject of relatively few studies,either by using it as a signal or by using images that were transformed from the PTB signal itself.Classification is performed by DL architecture in a few articles[22–25]by converting PTB signals to images;even though by using sophisticated architecture,we can get some good metrics value but still when we do signal-to-image conversation execution time will be more,this will be one disadvantage even if we are using an advanced combination of layers.Researchers directly translated images from the PTB signal in the article [22] without using signal processing procedures and utilized six layers of CNN architecture by removing goldberger-leads or image compression and achieved 81 ± 4% accuracy compared to physicians with 67±7%accuracy.In another scenario,by applying Continuous Wavelet Transform (CWT),the PTB signal is converted to an image by scalogram technique,which is a frequency coefficient of CWT,then classified by simple CNN [25] architecture and got accuracy by 87.50%,followed by 92.50%by Alexnet[26]architecture,applied in parallel to two graphics processing unit,with five convolutional layers,three pooling layers,two fully connected layers,and an output layer,converted to 227 ?227 shape to proceed for classification,100%by GoogleNet,structured by nine inception layers,then by residual neural network(ResNet)architecture achieved an accuracy of about 99.17%lastly 100%by utilizing EECGNet.

    The modelling in the preceding publications is accomplished by transforming signal-to-image.Using the PTB signal,the imbalanced condition can be handled by assigning class weight for Artificial Neural Network(ANN)[23],achieving an accuracy of about 97.64%.The authors of the paper[24]used CNN architecture,in which the predictor network consists of five residual blocks,each with 32 kernels of size 5,for a maximum pooling size of 5,and stride two,followed by 32 neurons fully connected layers activated by ReLU AF,and achieved 95.9%accuracy.Multiple convolution kernels of various receiving domains extract distinctive features of different scales.The retrieved multiple-scale features were combined using CNN and achieved an F1 score of 99.69%[26].Other DL architectures with long short-term memory and SVM algorithms[27]gained 96.27%accuracy.A scalable framework and transfer learning allow adding more subjects to an existing system without saving previously recorded ECGs [28].The quantum evolutionary algorithm is used to prune techniques that use the redundant filters in the proposed CNN model,and they have a final accuracy of roughly 99%.As a result,from the studies reviewed,we highlight the metrics values and neglect to take execution time into account,which is another crucial notion.The fact that we are dealing with medical data makes this essential.

    DL architectures outperform ML algorithms on average in terms of classification accuracy.In contrast,DL technology demands more processing power.Significant research is being undertaken to utilize a partial CNN architecture for feature extraction.They use various ML algorithms for classification,such as CNN with KNN for plant classification [28],CNN with XGBoost for EEG spectrogram picture analysis [29,30],CNN with SVM for magnetic resonance images [31],weed detection [32],image classification [33],handwritten digit recognition [33,34] and human activity recognition [35].It is not about the applications they use but about the output.Each technology has advantages and weaknesses;when we combine them,we can achieve something new,in our case,execution time and classification accuracy.Our work offers a computationally efficient classification model for the PTB dataset by integrating computationally less demanding components of DL architectures with ML algorithm(KNN).Some research has been done with the same procedure after converting the signal (1D) to an image (2D).As a result,they achieved an accuracy of 99.52% in the ECG application[36]and an F1 score of 84.19%in the Atrial Fibrillation ECG[36].Apart from these works,there are many works in literature reporting fast ML models for ECG based heartbeat classification and arrythmia detection [37].In [38],a comprehensive review of techniques to handle abnormal patterns in ECG signals is presented.Notable among the ECG datasets used in literature for validation of DL and ML algorithms are PTB and MIT-BIH[39–41].There are many important factors to consider when deciding whether to use DL over ML,and one among them is the AF[42–46].AFs are essential in DL architecture because they give the model the required non-linearity,which allows it to recognize intricate patterns and connections in the data and also the output of the AF for each neuron in a layer decides whether or not that neuron is activated,and as a result,whether or not its output signal is sent to the neurons in the next layers.

    3 Proposed Methodology

    It is evident from literature that the approaches integrating CNN with ML algorithms are faster compared to the baseline CNN.However,these approaches are homogenous in the sense that they use a fixed AF in the sequence of convolution layer blocks.In DL architecture,AFs can be used in either a homogeneous manner(i.e.,same AF in all layers)[47–49]or a heterogeneous manner(i.e.,different AF for each layer).To the best of our knowledge,heterogeneous activation facility is not explored or experimented in literature.Heterogeneous approach can give the model a more robust capacity for learning.Homogeneous approach,on the other hand,might not be as good at encapsulating the complete intricacy of the data.There are several AFs available and the accuracy of the DL approaches vary with the choice of AFs used [42–46].This reveals the fact that each AF captures features in a manner unique to them.In this section,details of the datasets,significance and choice of AFs,details of evaluation metrics,and the proposed CHAF architecture are discussed in a detail.

    3.1 Dataset Description and Implementation

    PTB Dataset:Germany’s National Metrology Institute compiled the PTB Diagnostic ECG database[39,40].Twelve leads,including three frank vectorcardiography leads(vx,vy,and vz),three augmented limb leads (aVR,aVL,and aVF),six precordial leads (V1,V2,V3,V4,V5,V6) and limb leads (I,II,and III) were used to compute each subject’s data.Two hundred ninety distinct participants (209 men and 81 women) contributed 549 data to the collection,22 of whom lacked a clinical report.A dataset of healthy people (52) and patients with various heart-related disorders(Myocardial Infraction-148,Heart Failure-18,Bundle Branch Block-15,Dysrhythmia-14,Myocardial Hypertrophy-7,Valvular Heart disease-6,Myocarditis-4,Misellaneous-4)is available in the diagnostic groupings for the remaining 268 people.

    MIT-BIH Dataset:The MIT-BIH arrhythmia database [40,41] comprises 48 half-hour ECG recordings from 47 subjects,collected between 1975 and 1979.Twenty-three recordings were randomly selected from 4000 24-h ambulatory ECGs from inpatients and outpatients at Boston’s Beth Israel Hospital.The remaining 25 recordings were specifically chosen to include less common but clinically significant arrhythmias not well-represented in a small random sample.The recordings were digitized at 360 samples per second per channel with 11-bit resolution over a 10-mV range.Each record includes computer-readable reference annotations for each beat,with approximately 110,000 annotations in total,independently annotated by two or more cardiologists.

    Implementation Details:The PTB diagnostic database and MITBIH dataset can be found at https://www.physionet.org/content/ptbdb/1.0.0/ and https://www.physionet.org/content/mitdb/1.0.0/.The preprocessed signals are obtained as given in [24].First,the continuous ECG signal is divided into 10-s windows,and one window is selected for analysis.Next,the amplitude values of the selected window are normalized to a range between zero and one.Then,the set of all local maxima is found based on the zero crossings of the first derivative.After that,a threshold of 0.9 is applied to the normalized values of the local maxima to identify the ECG R-peak candidates.The median of the R-R time intervals is calculated as the window’s nominal heartbeat period,say T.A signal segment with a length equal to 1.2×T is selected for each R-peak.Each selected segment is padded with zeros to reach a predefined fixed length to ensure uniformity.It is important to note that this beat extraction method is both simple and effective in extracting R-R intervals,irrespective of signal morphology[24].No assumptions are made about the signal’s morphology or spectrum,and all extracted beats have identical lengths,facilitating further processing and analysis.MIT-BIH is a relatively large sized dataset with sample size of 109,444 compared to PTB which has a sample size of 14,550.For our experiments,the train-test split is 70:30.Since the primary focus is on the heterogeneous activation facility,no hyperparameter tuning is introduced.Codes and datasets used for generating the results reported in Sections 3 and 4 are available at https://github.com/anandprems/chaf_tech_science_cmc.

    3.2 Convolution Neural Network and Activation Function

    Convolution and pooling layers,which can map and extract features from input data to expedite learning and minimize over-fitting,are unique features of CNN compared to typical neural networks.In the same way,to extract features from the PTB dataset,we use convolution and max-pooling layers,which are considered to be single block throughout the process.Since we were not sure how many blocks would be required to achieve the ideal feature extraction and classification process combination,we initialized with 10 blocks.To boost neurons from one layer to the next,we used ReLU AF in all levels except the output layer,where we used ‘Sigmoid’AF because our dataset involves binary categorization.In the base level of understanding of CNN architecture,we developed a sequence of blocks consisting of convolution and max-pooling layers employing CNN architecture;since there is no certainty on the best number of blocks to use,we initialized with 10 blocks;the accuracy metrics value increased slightly from 93.34%to 99.18%until the sixth block,after which it began to drop in a steady phase as shown in Table 1.In addition,the number of parameters and execution time from block 1 rapidly increases.It is widely agreed that adding more layers does not guarantee that accuracy would improve further.Instead,there will be a saturation point in every dataset;as a result,for our PTB dataset,6th block is where we will see the best results.

    Table 1:Layer details and performance of CNN architecture with ReLU AF

    3.3 Finding Suitable Activation Function

    While using sequence of blocks (convolution and max-pooling layers) by using PTB dataset,in the initial experimentation from Table 1,we predicted block 6 which gives best result.So,in the experimentation we have utilized ReLU AF(Homogeneous).Most researchers favour ReLU[46]AF for its functionality while using DL architecture in numerous types of study.In our experimentation,we first used ReLU AF for all the layers,then for the final layer,we used‘Sigmoid’AF as the default due to binary categorization.In the next level of experimentation,in order to visualize the significance of AF,we applied different AFs (homogeneously) to the same CNN architecture with six blocks and compared their execution time and accuracy.Among all AFs calculated,for SoftMax AF (in 75.55 s) and Sigmoid AF (in 45.48 s),the performance accuracy is 72.14%,while other AFs like Swish (99.23% in 75.52 s),Exponential Linear Unit (elu) (99.33% in 72.36 s),Parametrized ReLu(99.16%in 77.77 s),Scaled Exponential Linear Unit (selu) (99.04% in 52.84 s),Concatenated ReLu(crelu)(98.92%in 67.36 s),ReLu-6(99.16%in 66.19 s),Leaky ReLu(99.27%in 68.32 s),ReLu(98.94%in 45.59 s),Tanh(99.46%in 46.91 s),Gaussian Error Linear Unit(99.34%in 54.41 s),Linear function(99.49%in 44.59 s),SoftSign(99.31%in 71.74 s),SoftPlus(96.94%in 71.45 s)and Mish AF(99.55%in 60.26 s),which has overall accuracy of above 95%.It can be inferred from that SoftMax and Sigmoid are insignificant in this case.Therefore,it is wise to exercise greater caution by experimenting with various AFs in order to attain high accuracy and a short execution time.

    3.4 Metrics for Model Evaluation

    For ML and DL technologies,we need to justify our model by using metrics value.Metrics provide a clear and concise summary of the model’s performance and can aid in developing trust in the model’s outcomes.Metric values enable us to assess how well a model performs on a specific task and identify the areas where the model may be improved.Among many available metrics,confusion matrix based metrics have more significance,as they provide complete knowledge of a model’s performance.The matrix contains information about the model’s True Positive(TP)which denotes the number of correct positive predictions,True Negative (TN) which denotes the number of correct negative predictions,False Positive(FP)which denotes the number of wrong false predictions(i.e.,false alarms),and False Negative(FN)which denotes the number of wrong negative predictions.These parameters are in turn used to determine a model’s accuracy and related metrics such as precision,recall,and F1 score.

    Accuracy is defined as the fraction of correct predictions and it is mathematically expressed as

    Precision is defined as the ratio of correct positive predictions to the total positive predictions and it is expressed as

    Recall is defined as the ratio of correct positive predictions to the number of actual positives and it is expressed as

    The F1 score considers both precision and recall,and it is defined as follows:

    The F1 score from equation is significant because of its capacity to provide a balanced measure of the model’s performance.In particular,it is useful when both precision and recall are to be balanced.In addition to these accuracy related metrics,execution time is also measured and reported as it is very crucial in biomedical applications.In this manuscript,execution time denotes the computation time in Python 3.9.10 running on an AMD Ryzen 7 4800H with Radeon Graphics 2.90 GHz,RAM 16 GB.For certain models or applications,the accuracy will be high at the cost of execution speed.On the other hand,certain models or applications have lesser execution time at the compromise of accuracy.Therefore,for thorough validation of our proposed approach,five metrics are considered:Accuracy,precision,recall,F1 score and execution time.

    3.5 Convolution-Based Machine Learning Algorithms

    From the second experimentation,with AF,we come to know,not only ReLU AF performs well for our PTB dataset,but by utilizing different AFs in a homogeneously scenario,we got best AFs with good accuracy and also faster process(Execution time).From the experimentation done so far,we have seen,the number of parameters and the complexity of the model rises as the number of layers in CNN architecture increases to certain point but not as exponentially,which leads to greater performance in terms of accuracy to certain block(here till 6thblock).However,there comes expense of higher computing time and memory utilization,making the model difficult to apply in real-world applications.As a result,when constructing any DL architecture,including CNN,it is important to balance metrics values,model complexity,and computational performance.While in the case of ML algorithms,it has some advantages over DL architectures,such as being constructed on mathematical models rather than layers,the complexity and computational performance will be better when compared to DL.ML algorithms,such as KNN,SVM,RF,and XGBoost,are frequently used also straightforward and interpretable than DL architectures and can work well for our data kinds of low to moderate complexity problems.The process can also be easier and faster to train than DL architectures and require fewer computer resources.While in DL architectures,such as CNN and RNN,may tackle more complicated issues with greater accuracy and efficiency,particularly when dealing with huge datasets and high-dimensional characteristics,but execution time is longer as we will be using more layers.That is where ML comes to picture,when ML fuses with DL architecture then we can expect better result.

    3.6 Proposed CHAF-ML Architecture

    We built a customized fully connected CNN architecture at the first level and acquired good accuracy with a longer execution time as the blocks increases with layers and also with homogeneous AF(ReLU).As a result,in the next experimentation we utilized different AFs homogeneously for the same CNN architecture along with ReLU AF and as a result we get different best results with accuracy and execution time.As we selected 6 blocks from CNN architecture(based on accuracy),top 6 AFs for each block we selected for further process.In the next stage of experimentation,we modified our CNN architecture like acquiring feature extraction by CNN and then by removing the dense layer,we included ML algorithms in the classification stage.We witnessed some enhanced accuracy as well as faster execution time when compared to the first stage (CNN architecture).As we are dealing with medical data,for faster response,we implemented CHAF to PTB diagnostic ECG database using a variety of ML algorithms and heterogeneous AFs in each layer for enhanced results as shown in Fig.2.

    Figure 2:Proposed CHAF-ML architecture

    CNNs are powerful feature extractors that can capture spatial and temporal patterns in data.However,they may not always be the most efficient or effective classifiers.By combining CNN with ML algorithms such as SVM,KNN,Decision Tree,and RF,we can leverage the strengths of both approaches to improve classification accuracy.Additionally,utilizing heterogeneous AF between each stage of the CNN can also improve performance.AF plays a critical role in DL architectures,as they introduce non-linearity and enable the model to learn complex relationships between inputs and outputs.However,using the same AF throughout the entire network can lead to a saturation of gradients and slower convergence.By using different AFs at various stages of the network,we can introduce more diversity and flexibility into the model,allowing it to learn more efficiently,and improve the speed of the network.AFs tailored to each layer’s specific characteristics can reduce the required computations,leading to faster execution.

    4 Experimental Results and Discussions

    In this section,the performance of general CNN-ML architectures and CHAF-ML architectures will be reported and analyzed to decide the number of convolution layers for CHAF and the AFs for the proposed model.Then,the proposed CHAF model will be compared with the baseline CNN model and the state-of-the-art models.The train-test split is 70:30 and the implementation information can be accessed at https://github.com/anandprems/chaf_tech_science_cmc.

    As our domain is related to medical data,apart from accuracy and other metrics,we need to take care of different situations,like how effectively we can create a model with less time consumption;as the medical domain needs some timely advice;sometimes it may lead to collateral damage while diagnosis,so classification layer associated with CNN as in Fig.2,is placed in a complicated position as layers get increased,so as the complexity,we thought to reduce the complexity in the classification part by replacing dense layer to any ML algorithms.One advantage of ML algorithm over DL architecture is that ML algorithms are built on the mathematical model rather than the layers concept;hence,if you see Tables 2 and 3 compared to Table 1,we can understand the importance of including ML algorithms.

    Each block combines convolution and max-pooling layers for the feature extraction process.In Table 1,we utilized CNN architecture with homogeneous AF as ReLU for all the blocks except the classification layer;as our dataset comes under the binary category,we used Sigmoid AF.The accuracy of each block gets increasing from block number 1 to 6;in block 6,we get the best result of 99.18%of accuracy,but parallelly execution time at each block gets increases because of increase in parameters.It is not mandatory if our parameters get increased;our metrics value also will increase,and the complexity level by default,will be more if we have more number parameters.Hence execution time also will get increased.Additionally,accuracy will decrease at one point.After 6 blocks,the accuracy started dropping from 99.18%to 98.69%of 8 blocks,and the execution time also increased to 390.109 s,which we need to avoid for the medical domain.In order to bring novelty,we didn’t perform any new architecture,instead we done some modifications with different stage of experiments.In the first experiment stage,we used CNN with homogeneous AF and found the best block from the combination of convolution and max-pooling will be 6 blocks as from above discussion and from Table 1.Beyond that,if we start concatenating the blocks,the accuracy metric values get dropped because of complexity,and execution time starts increasing.So,in the following experiment stage,we started experimenting with different homogeneous AF;by fixing 6 blocks,we performed with different AFs as discussed in Section 3.3.From the experiment,most of the AFs have accuracy of above 96%.Some AFs,like SoftMax(72.14%)and Sigmoid(72.14%),got less accuracy of about 72%,which is very low and should avoid.So,from this we understand that for different AFs the accuracy and execution time varies and AF is also important factor we need to consider.

    In the next stage of experiment,instead of dense layer in CNN for classification process,we modified and utilized ML algorithms like SVM,KNN,RF,Decision Tree,XGBoost and performed with homogenous AF (ReLU) as tabulated in Table 2.So when compared to previous experiment with CNN architecture as in Table 1,here we got better result that also from 6 blocks,it is evident from here,each architecture has some advantage over other,it is not DL which is giving best in all aspects but when we have proper planning we get better result than DL,like we got for RF(99.10%in 3.05 s),XGBoost(99.22%in 2.56 s),Histogram Gradient Boosting(HGB)(99.08%in 11.15 s),KNN(99.31%in 0.006 s),SVM Linear(SVML)(99.26%in 1.92 s),SVM Radial Basis Function(SVMR)(99.12%in 2.52 s)and SVM Polynomial(SVMP)(99.10%in 2.71 s).From the values,we can infer that for the same 6 blocks,compared to Table 1,we get the best result of about 99.31% for CNN-KNN architecture,achieved at 0.006 s,which is very promising by using ReLU homogeneous AF.

    Table 2:Layer details and performance of homogeneous CNN–ML architecture

    In the last experiment,we performed for CNN architecture with different AFs,fixing the number of blocks as 6,we are again experimenting with different AFs for the same architecture CNN and ML algorithms but this time in a heterogeneous way,which means each layer different AF instead of same.After testing with 16 different AFs,it can be seen by using SoftMax and Sigmoid AF,we are getting very low accuracy,and other AFs are getting the accuracy of the above 92%for all the ML algorithms in combination with CNN feature extraction process.So,from 16 different AFs we experimented,we selected top AFs performing well for the combination of CNN and ML algorithms.CNN-ML best combination of AFs are as elu (RF-98.94% in 9.29 s,XGBoost-98.96% in 12.18 s,HGB-99.10% in 20.18 s,KNN-99.22% in 0.008 s,SVML-99.17% in 0.57 s,SVMR-99.08% in 6.26 s,SVMP-99.08%in 1.20 s),tanh (RF-99.06% in 9.02 s,XGBoost-99.17% in 12.27 s,HGB-99.03% in 20.92 s,KNN-99.45%in 0.004 s,SVML-99.19%in 0.84 s,SVMR-99.26%in 5.94 s,SVMP-99.24%in 1.44 s),linear(RF-99.01%in 13.53 s,XGBoost-99.15%in 15.46 s,HGB-99.12%in 28.55 s,KNN-99.40%in 0.008 s,SVML-99.22%in 0.85 s,SVMR-99.22%in 6.09 s,SVMP-99.33%in 4.33 s),softsign(RF-99.10%in 11.54 s,XGBoost-99.15%in 13.71 s,HGB-99.12%in 22.43 s,KNN-99.42%in 0.012 s,SVML-99.31%in 1.35 s,SVMR-99.49%in 1.35 s,SVMP-99.45%in 3.76 s),crelu(RF-99.22%in 12.49 s,XGBoost-99.19%in 19.53 s,HGB-99.17%in 41.68 s,KNN-99.54%in 0.011 s,SVML-99.26%in 1.03 s,SVMR-98.41%in 40.56 s,SVMP-99.22%in 1.42 s)and relu6(RF-98.96%in 4.71 s,XGBoost-99.06%in 8.47 s,HGB-99.08%in 18.26 s,KNN-99.24%in 0.008 s,SVML-98.94%in 1.17 s,SVMR-99.28%in 2.51 s,SVMP-99.06% in 1.52 s).From the values itself,we can assume the importance of CNN-ML over CNN architecture in terms of all aspects.

    The proposed concept CHAF is another modification we do with the AFs.So instead of using AFs homogeneously,we are utilizing heterogeneous AFs,which means from other AFs,we are selecting the top six AFs because we got the best result in 6 blocks.We implemented different combinations with the selected AFs,and the best combinations are tabulated in Table 3;it is evident that we are getting the best result for the variety of tanh,linear,softsign,elu,crelu and relu6 AFs,so it’s not architecture which gives us best metrics value,we need to consider AFs also,here we get the best result for CHAF KNN-99.55%of accuracy at 0.008 s,then CHAF-SVML-99.41%of accuracy at 0.706 s when compared to DL architecture,the model complexity is lesser for KNN and Linear SVM algorithms,that is why we can achieve best accuracy with faster execution.

    Table 3:Layer details and performance of CHAF–ML architecture

    To establish comparison with the state-of-the-art,the proposed CHAF-KNN model and three other models are first validated on PTB dataset.The proposed CHAF-KNN model utilizes the CHAF for extraction of features and KNN for classification.The CHAF in this case comprises of six convolution layer blocks,each using different AFs(i.e.,tanh,linear,softsign,elu,crelu,relu6).This choice is based on the performance reported in Tables 2 and 3.Three models used for comparison are baseline CNN comprising of six convolution layer blocks using ReLu as AF and a fully connected layer[14],CNN-KNN comprising of six convolution layer blocks using ReLu as AF and KNN[16],CNN-XGBoost comprising of six convolution layer blocks using ReLu as AF and XGBoost [50].For KNN,the value of K is fixed as 5.To validate the generality of the proposed approach,another dataset,the MIT-BIH dataset,is considered for evaluation.As mentioned in Section 3.1,MIT-BIH is a large sized dataset compared to PTB.Therefore,the performance of the proposed CHAF-KNN model is benchmarked using both PTB and MIT-BIH to have a clear understanding of the generality and scalability of the approach.Implementation-wise,in the case of MIT-BIH dataset,the number of convolution blocks is five for all four models compared and the five AFs in the case of CHAF-KNN are relu6,elu,softsign,crelu,and tanh.This choice is based on the preliminary experiments conducted as in the case of PTB.Table 4 shows the comparison of the proposed approach with the state-of-the-art,for both PTB and MIT-BIH datasets.Fig.3 shows the performance comparison in terms of accuracy and related metrics for both datasets.Similarly,Fig.4 shows the execution time comparison for both the datasets.

    Figure 3:Performance comparison in terms of accuracy,precision,recall and F1 score-for PTB(left)and MIT-BIH dataset(right)

    In the case of PTB dataset,the proposed CHAF-KNN has an accuracy of 99.55% and an F1 score of 99.68%in 0.008 s,outperforming the state-of-the-art CNN-XGBoost which has an accuracy of 99.38%and an F1 score of 99.32%in 1.23 s.Similarly,in the case of MIT-BIH dataset,proposed CHAF-KNN has an accuracy of 99.08%and an F1 score of 98.99%in just 0.07 s,outperforming the state-of-the-art CNN-XGBoost which has an accuracy of 98.36%and an F1 score of 98.37%in 13.80 s.It can be inferred that,in the proposed CHAF-KNN,the heterogeneous activation facility contributes to improvement in accuracy and F1 score and the KNN expedites the classification process.

    Advantages of CHAF Architecture:Architecture-wise,the heterogeneous nature of CHAF introduces more diversity and flexibility into the model.CHAF is accommodative in the sense that it can include multiple AFs and perform well with most ML models.The convolution layer blocks in CHAF utilize different AFs to introduce different non-linearity,thereby enabling the model to learn complex input-output relationships and extract features in a better manner compared to that of the convolution blocks in CNN.This fact is evident from the performance improvement in CHAF-KNN as compared to CNN-KNN,in the case of both PTB and MIT-BIH datasets.As expected,the CHAF-KNN is much faster that the baseline CNN and the CNN-XGBoost,and has speed comparable to that of the CNN-KNN.

    Figure 4:Execution time comparison–PTB(left)and MIT-BIH dataset(right)

    Limitations of CHAF and Scope for Improvement:The choice of AFs is based on the results obtained from the preliminary experiments.Therefore,AFs tailored according to each layer’s specific characteristics can be deployed to improve the performance further.On comparing the execution times of the proposed CHAF for PTB and MIT-BIH,it is obvious that the execution time is proportional to the size of the dataset.Special focus on handling scalability is required to bring down the execution time.

    5 Conclusion

    CNN-ML architectures used a fixed AF in the convolution and pooling layers sequence,making them homogeneous.Due to the availability of a variety of activation functions available,and their capability to capture features uniquely,we proposed a CHAF which uses multiple AFs in the convolution layer blocks,one for each block,with a motivation of extracting features in a better manner.Experimental results presented in this manuscript show that the CHAF is accommodative(i.e.,it can accommodate many AFs)and effective with most ML algorithms.For PTB and MIT-BIH datasets,the proposed CHAF-KNN outperforms the state-of-the-art approach,the CNN-XGBoost,with improved accuracy and much lesser execution time.In addition,CHAF-KNN outperforms its homogeneous CNN-KNN with improved accuracy and comparable execution time.

    Recommendations for Future Work:(i) Hyperparameter tuning can be introduced to further validate the performance of the proposed approach.(ii)AFs tailored according to each layer’s specific characteristics can be deployed to improve the performance further.

    Acknowledgement:The authors would like to express their gratitude to Vellore Institute of Technology(VIT),Chennai,India,for providing valuable support to carry out this work at VIT.We thank Dr.Balasundaram A,Assistant Professor (Senior Grade 2),Centre for Cyber Physical Systems,VIT,Chennai,for his valuable suggestions for the improving the quality of this manuscript.

    Funding Statement:The authors received no specific funding for this study.

    Author Contributions:Study conception and design: Premanand S,Sathiya Narayanan;data collection: Premanand S;analysis and interpretation of results: Premanand S,Sathiya Narayanan;draft manuscript preparation:Premanand S.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:Code formats and datasets used for generating the results reported in Sections 3 and 4 are available at https://github.com/anandprems/chaf_tech_science_cmc.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    国产黄色免费在线视频| 男人舔奶头视频| 免费av中文字幕在线| 伦精品一区二区三区| 久久精品国产亚洲网站| 人妻少妇偷人精品九色| 九九久久精品国产亚洲av麻豆| 黄色视频在线播放观看不卡| 婷婷色综合www| 嫩草影院入口| 99热国产这里只有精品6| 久久久亚洲精品成人影院| 亚洲国产日韩一区二区| 国产一区二区三区综合在线观看 | 久久鲁丝午夜福利片| 欧美97在线视频| 高清毛片免费看| 一区二区三区免费毛片| 七月丁香在线播放| 精品久久久久久电影网| 观看av在线不卡| 国产精品国产三级专区第一集| 亚洲精品视频女| 精品人妻偷拍中文字幕| 国产熟女欧美一区二区| 国产精品麻豆人妻色哟哟久久| 国产欧美日韩一区二区三区在线 | 69精品国产乱码久久久| 久久99精品国语久久久| 亚洲欧美成人精品一区二区| 自线自在国产av| 能在线免费看毛片的网站| 婷婷色麻豆天堂久久| 岛国毛片在线播放| 精品人妻一区二区三区麻豆| 纯流量卡能插随身wifi吗| 少妇 在线观看| 在线观看国产h片| 少妇人妻一区二区三区视频| 久久鲁丝午夜福利片| 国产乱来视频区| 婷婷色综合www| 热99国产精品久久久久久7| 我要看日韩黄色一级片| 菩萨蛮人人尽说江南好唐韦庄| 在线看a的网站| 国产免费福利视频在线观看| 在线播放无遮挡| 男人添女人高潮全过程视频| 免费看日本二区| 91在线精品国自产拍蜜月| 午夜免费鲁丝| 成人18禁高潮啪啪吃奶动态图 | 熟女电影av网| 亚洲精品一区蜜桃| 久久人人爽人人片av| 国产黄片视频在线免费观看| 大码成人一级视频| 一级黄片播放器| 亚州av有码| 亚洲精品aⅴ在线观看| 免费av中文字幕在线| 亚洲av福利一区| 九九久久精品国产亚洲av麻豆| 老女人水多毛片| 成人毛片a级毛片在线播放| 观看美女的网站| 精品人妻一区二区三区麻豆| 日本vs欧美在线观看视频 | 天美传媒精品一区二区| 好男人视频免费观看在线| 伊人久久国产一区二区| 一级毛片 在线播放| 黑人高潮一二区| a级片在线免费高清观看视频| 国产免费又黄又爽又色| 国产精品蜜桃在线观看| 色婷婷久久久亚洲欧美| 久久6这里有精品| 亚洲国产精品一区三区| 日韩av免费高清视频| 97超视频在线观看视频| 亚洲精品色激情综合| 亚洲综合色惰| 伊人久久精品亚洲午夜| 欧美成人午夜免费资源| 曰老女人黄片| 啦啦啦视频在线资源免费观看| 国产精品秋霞免费鲁丝片| 国产精品不卡视频一区二区| 中文字幕精品免费在线观看视频 | 最后的刺客免费高清国语| 成人漫画全彩无遮挡| 男女边摸边吃奶| 国产片特级美女逼逼视频| 一级片'在线观看视频| 精品人妻一区二区三区麻豆| 男人和女人高潮做爰伦理| 国产精品国产三级国产av玫瑰| 免费人妻精品一区二区三区视频| 久久久久网色| 日韩亚洲欧美综合| 伦理电影大哥的女人| 亚洲婷婷狠狠爱综合网| 亚洲国产日韩一区二区| 国产av一区二区精品久久| 国产 精品1| 色5月婷婷丁香| 免费黄色在线免费观看| 午夜福利网站1000一区二区三区| 欧美+日韩+精品| 99热这里只有是精品在线观看| 最黄视频免费看| 免费大片黄手机在线观看| 精品亚洲成a人片在线观看| 狠狠精品人妻久久久久久综合| 精品一区在线观看国产| 只有这里有精品99| 精品国产一区二区三区久久久樱花| 人人妻人人爽人人添夜夜欢视频 | 亚洲人成网站在线观看播放| 一个人免费看片子| 欧美高清成人免费视频www| 一级爰片在线观看| av在线app专区| 最近中文字幕高清免费大全6| 久久久久精品久久久久真实原创| 国产精品久久久久久久电影| 两个人免费观看高清视频 | 久久午夜综合久久蜜桃| 国产伦精品一区二区三区视频9| 国产永久视频网站| 久久久午夜欧美精品| 一本一本综合久久| 伦理电影大哥的女人| 一级片'在线观看视频| 中文在线观看免费www的网站| 精华霜和精华液先用哪个| 精品酒店卫生间| 亚洲国产色片| 六月丁香七月| 人妻少妇偷人精品九色| 精品亚洲乱码少妇综合久久| 亚洲av在线观看美女高潮| 亚洲欧美中文字幕日韩二区| 在现免费观看毛片| 精品久久久久久电影网| 欧美老熟妇乱子伦牲交| a级一级毛片免费在线观看| 在线观看一区二区三区激情| 国产日韩欧美亚洲二区| 日日啪夜夜爽| 丝瓜视频免费看黄片| 国产在线一区二区三区精| 亚洲精品一区蜜桃| 老熟女久久久| 国产色婷婷99| 国产探花极品一区二区| 色94色欧美一区二区| 久久精品国产a三级三级三级| 色婷婷久久久亚洲欧美| 国产亚洲午夜精品一区二区久久| 搡女人真爽免费视频火全软件| 天堂中文最新版在线下载| 亚洲精品一二三| 国产成人一区二区在线| 少妇的逼水好多| 国产亚洲精品久久久com| 久久99精品国语久久久| 久久热精品热| 午夜免费男女啪啪视频观看| 久久久国产欧美日韩av| 久久99精品国语久久久| 久久鲁丝午夜福利片| 国产亚洲av片在线观看秒播厂| 插阴视频在线观看视频| www.av在线官网国产| 久久亚洲国产成人精品v| 伦理电影大哥的女人| 久久久久久久久久人人人人人人| 国内少妇人妻偷人精品xxx网站| 大话2 男鬼变身卡| 久久久久久久久久久免费av| 黄色配什么色好看| 少妇的逼水好多| 国产高清三级在线| 亚洲av在线观看美女高潮| 狂野欧美激情性bbbbbb| 国产精品久久久久久精品电影小说| a 毛片基地| 在线观看美女被高潮喷水网站| 一级片'在线观看视频| 国产男人的电影天堂91| 午夜久久久在线观看| av一本久久久久| 亚洲av电影在线观看一区二区三区| 男女免费视频国产| 国产伦精品一区二区三区四那| 亚洲va在线va天堂va国产| 国国产精品蜜臀av免费| 日本vs欧美在线观看视频 | 国产伦精品一区二区三区四那| 免费黄频网站在线观看国产| 成人二区视频| 在线观看www视频免费| 一区二区三区乱码不卡18| 夫妻午夜视频| 中文精品一卡2卡3卡4更新| 国产欧美日韩综合在线一区二区 | 极品教师在线视频| 天美传媒精品一区二区| 久久精品久久久久久久性| 精品一区在线观看国产| 亚洲国产精品一区三区| 日产精品乱码卡一卡2卡三| 精品国产一区二区三区久久久樱花| 国产精品一区二区在线观看99| 亚洲欧美成人精品一区二区| 美女中出高潮动态图| 久久久久久久亚洲中文字幕| 久久久久久久国产电影| 婷婷色av中文字幕| 亚洲精品中文字幕在线视频 | 女性生殖器流出的白浆| kizo精华| 成人漫画全彩无遮挡| 国产精品三级大全| 免费高清在线观看视频在线观看| 成人黄色视频免费在线看| 麻豆成人av视频| 国产精品不卡视频一区二区| 成年人免费黄色播放视频 | 我要看日韩黄色一级片| 久久女婷五月综合色啪小说| 亚洲第一av免费看| 国产乱人偷精品视频| 老女人水多毛片| 黄色日韩在线| 少妇人妻 视频| 王馨瑶露胸无遮挡在线观看| 最近的中文字幕免费完整| a级毛色黄片| 日本91视频免费播放| 国产男女超爽视频在线观看| 欧美bdsm另类| 91在线精品国自产拍蜜月| 欧美丝袜亚洲另类| 九九爱精品视频在线观看| 久久精品国产亚洲av涩爱| 中文欧美无线码| 天美传媒精品一区二区| 国产美女午夜福利| 新久久久久国产一级毛片| 十八禁网站网址无遮挡 | 国产亚洲最大av| 国内精品宾馆在线| 成年av动漫网址| 国产在线视频一区二区| 国产精品偷伦视频观看了| 男女国产视频网站| 亚洲国产色片| 亚洲三级黄色毛片| 亚洲欧美一区二区三区黑人 | 一区二区三区四区激情视频| 麻豆乱淫一区二区| 久久国产精品男人的天堂亚洲 | 国产在视频线精品| 伊人久久国产一区二区| 成人亚洲精品一区在线观看| 久久精品国产自在天天线| 免费大片18禁| 国产中年淑女户外野战色| 日韩欧美一区视频在线观看 | 一本一本综合久久| 丝袜喷水一区| 夜夜看夜夜爽夜夜摸| av免费在线看不卡| 午夜老司机福利剧场| 久久久久久人妻| 久久鲁丝午夜福利片| 美女主播在线视频| 亚洲va在线va天堂va国产| 免费黄频网站在线观看国产| 国产男女超爽视频在线观看| 久久久久精品久久久久真实原创| 亚洲,一卡二卡三卡| 国产一级毛片在线| 高清视频免费观看一区二区| 久久热精品热| 亚洲经典国产精华液单| 国产白丝娇喘喷水9色精品| 日韩强制内射视频| 综合色丁香网| 国产精品一二三区在线看| 人妻 亚洲 视频| 精品久久久精品久久久| 亚洲精品国产av蜜桃| 亚洲久久久国产精品| 久久99一区二区三区| 国产一区亚洲一区在线观看| 国产成人a∨麻豆精品| 水蜜桃什么品种好| 国产精品99久久久久久久久| 国产黄频视频在线观看| 亚洲色图综合在线观看| 久久精品国产a三级三级三级| .国产精品久久| 精品午夜福利在线看| 下体分泌物呈黄色| 亚洲丝袜综合中文字幕| 欧美区成人在线视频| 男女边摸边吃奶| 色哟哟·www| 国产日韩一区二区三区精品不卡 | 国产av精品麻豆| 五月玫瑰六月丁香| 一区二区三区乱码不卡18| 欧美丝袜亚洲另类| 国产成人freesex在线| 69精品国产乱码久久久| 日本av手机在线免费观看| 日韩一区二区三区影片| 精品亚洲成国产av| 欧美激情国产日韩精品一区| 日韩人妻高清精品专区| 成人特级av手机在线观看| 色视频www国产| 成人亚洲欧美一区二区av| 亚洲伊人久久精品综合| 在现免费观看毛片| 亚洲精华国产精华液的使用体验| freevideosex欧美| 久久99热这里只频精品6学生| 欧美成人精品欧美一级黄| 在线精品无人区一区二区三| 久久99精品国语久久久| 搡老乐熟女国产| 国产中年淑女户外野战色| 人妻 亚洲 视频| 汤姆久久久久久久影院中文字幕| 大码成人一级视频| 色婷婷久久久亚洲欧美| 日本欧美视频一区| 亚洲第一区二区三区不卡| 桃花免费在线播放| 男人和女人高潮做爰伦理| 亚洲第一av免费看| 菩萨蛮人人尽说江南好唐韦庄| 国产精品久久久久久久久免| 国产精品成人在线| 色视频www国产| 国产精品国产三级专区第一集| av女优亚洲男人天堂| 中文字幕人妻丝袜制服| 国产在视频线精品| 国产精品久久久久久久电影| 一本久久精品| 欧美三级亚洲精品| 亚洲四区av| 欧美日韩综合久久久久久| 日日摸夜夜添夜夜爱| 欧美+日韩+精品| 噜噜噜噜噜久久久久久91| 80岁老熟妇乱子伦牲交| 多毛熟女@视频| 国产色爽女视频免费观看| 日韩人妻高清精品专区| 大香蕉97超碰在线| 成年美女黄网站色视频大全免费 | 搡老乐熟女国产| 久久久精品免费免费高清| 亚洲国产欧美在线一区| 精品久久久噜噜| 亚洲av福利一区| 又黄又爽又刺激的免费视频.| 国产欧美日韩综合在线一区二区 | 亚洲一区二区三区欧美精品| 国产精品国产三级国产专区5o| 日韩人妻高清精品专区| 麻豆精品久久久久久蜜桃| 国内少妇人妻偷人精品xxx网站| 人妻少妇偷人精品九色| 亚洲一区二区三区欧美精品| 建设人人有责人人尽责人人享有的| 亚洲激情五月婷婷啪啪| 韩国av在线不卡| 亚洲av国产av综合av卡| a级毛色黄片| 亚洲精品日韩在线中文字幕| 最近2019中文字幕mv第一页| 国内少妇人妻偷人精品xxx网站| 国产美女午夜福利| 国产黄色免费在线视频| 亚洲精品成人av观看孕妇| 亚洲激情五月婷婷啪啪| 欧美丝袜亚洲另类| 婷婷色综合大香蕉| 精品99又大又爽又粗少妇毛片| av福利片在线| 蜜桃久久精品国产亚洲av| 亚洲av国产av综合av卡| 男人狂女人下面高潮的视频| 日韩不卡一区二区三区视频在线| 高清毛片免费看| 极品人妻少妇av视频| 久久精品国产a三级三级三级| 少妇人妻久久综合中文| 亚洲国产毛片av蜜桃av| 亚洲内射少妇av| 久久这里有精品视频免费| a级毛片免费高清观看在线播放| 国产午夜精品一二区理论片| 多毛熟女@视频| 亚洲欧美成人精品一区二区| 精品少妇久久久久久888优播| 国产欧美亚洲国产| 王馨瑶露胸无遮挡在线观看| 亚洲av二区三区四区| 成人无遮挡网站| 久久国产精品大桥未久av | 国产成人精品久久久久久| 国产精品久久久久成人av| 午夜福利在线观看免费完整高清在| 老司机影院成人| 久久女婷五月综合色啪小说| 一区二区三区乱码不卡18| 国产免费福利视频在线观看| 人妻少妇偷人精品九色| 国产精品国产三级专区第一集| 久久99热6这里只有精品| 日本与韩国留学比较| 婷婷色av中文字幕| 国产欧美日韩综合在线一区二区 | 日本午夜av视频| 国产精品久久久久久精品古装| 日本色播在线视频| 99热这里只有精品一区| 亚洲一级一片aⅴ在线观看| 一个人免费看片子| 国产成人精品婷婷| 亚洲不卡免费看| 亚洲国产av新网站| 精品一区二区三卡| 亚洲国产精品一区三区| 女性生殖器流出的白浆| 亚洲精品国产av成人精品| 欧美xxⅹ黑人| 精品一区二区三区视频在线| 这个男人来自地球电影免费观看 | 99国产精品免费福利视频| 久久 成人 亚洲| 亚洲人成网站在线观看播放| 久久精品久久精品一区二区三区| 久久久久久久久久人人人人人人| 热99国产精品久久久久久7| 天天操日日干夜夜撸| 狂野欧美激情性bbbbbb| 哪个播放器可以免费观看大片| 黄色怎么调成土黄色| 亚洲欧美日韩东京热| 热re99久久国产66热| 全区人妻精品视频| 简卡轻食公司| 中文欧美无线码| 自线自在国产av| 精品午夜福利在线看| 久久久欧美国产精品| 亚洲av电影在线观看一区二区三区| 十八禁网站网址无遮挡 | 欧美精品一区二区大全| 99热这里只有是精品在线观看| 69精品国产乱码久久久| 亚洲国产毛片av蜜桃av| 青春草亚洲视频在线观看| 人妻制服诱惑在线中文字幕| 精品久久久精品久久久| 男女啪啪激烈高潮av片| 国产精品伦人一区二区| 国产成人91sexporn| 亚洲第一av免费看| 国产精品女同一区二区软件| kizo精华| 久久久久人妻精品一区果冻| 日韩视频在线欧美| 色婷婷av一区二区三区视频| 婷婷色av中文字幕| 2022亚洲国产成人精品| 久久精品国产亚洲网站| 精品国产一区二区久久| 久久国产精品男人的天堂亚洲 | 亚洲精品乱久久久久久| a级片在线免费高清观看视频| 久久97久久精品| 最近中文字幕2019免费版| 成年人午夜在线观看视频| 国产精品福利在线免费观看| 老熟女久久久| 欧美少妇被猛烈插入视频| 国产精品99久久久久久久久| 亚洲欧美清纯卡通| 国产黄色免费在线视频| 99久久人妻综合| 国产亚洲91精品色在线| 人妻一区二区av| 亚洲国产最新在线播放| 日韩一本色道免费dvd| 妹子高潮喷水视频| 亚洲av男天堂| 赤兔流量卡办理| 在线观看免费高清a一片| av.在线天堂| 国产视频内射| 精品久久久久久电影网| 在线观看免费视频网站a站| av国产精品久久久久影院| 国产有黄有色有爽视频| a级毛色黄片| 久久久久久久久大av| 一级毛片aaaaaa免费看小| 少妇裸体淫交视频免费看高清| 免费av不卡在线播放| 国产精品久久久久久精品电影小说| 只有这里有精品99| 在线观看人妻少妇| 多毛熟女@视频| 女人久久www免费人成看片| 热99国产精品久久久久久7| 亚洲美女黄色视频免费看| 国产乱人偷精品视频| 有码 亚洲区| 国产伦理片在线播放av一区| 毛片一级片免费看久久久久| 91午夜精品亚洲一区二区三区| 久久久久久人妻| 国产亚洲欧美精品永久| 亚洲国产精品999| 嫩草影院入口| 亚洲av国产av综合av卡| av黄色大香蕉| 伦理电影免费视频| 久久女婷五月综合色啪小说| 在线免费观看不下载黄p国产| 九九久久精品国产亚洲av麻豆| 少妇的逼水好多| 王馨瑶露胸无遮挡在线观看| 国产精品国产三级专区第一集| 97在线人人人人妻| 久久影院123| 欧美最新免费一区二区三区| 下体分泌物呈黄色| 人人妻人人澡人人看| 日韩电影二区| 久久国产亚洲av麻豆专区| 国产永久视频网站| 国产一级毛片在线| 久久久久视频综合| 国产探花极品一区二区| 麻豆成人午夜福利视频| 视频区图区小说| 午夜免费男女啪啪视频观看| 美女内射精品一级片tv| 久久久久久久久久久久大奶| 蜜桃久久精品国产亚洲av| 激情五月婷婷亚洲| 国产精品.久久久| 国产淫片久久久久久久久| 国产熟女午夜一区二区三区 | 伊人亚洲综合成人网| 欧美成人精品欧美一级黄| 国产精品三级大全| av有码第一页| 亚洲人与动物交配视频| av免费在线看不卡| 不卡视频在线观看欧美| 精品一区二区三区视频在线| 少妇人妻 视频| h视频一区二区三区| 卡戴珊不雅视频在线播放| 久久久久久久久久久久大奶| 国产91av在线免费观看| 一级毛片电影观看| 三级经典国产精品| 最近的中文字幕免费完整| 精品久久久久久电影网| 国产伦在线观看视频一区| 大香蕉97超碰在线| 久久久久精品久久久久真实原创| 久久久精品免费免费高清| 精品久久久精品久久久| 欧美日韩视频精品一区| 美女中出高潮动态图| 一二三四中文在线观看免费高清| 在线观看av片永久免费下载| 亚洲精品成人av观看孕妇| 爱豆传媒免费全集在线观看| 人人妻人人爽人人添夜夜欢视频 | 国产av精品麻豆| 男人狂女人下面高潮的视频| √禁漫天堂资源中文www| 久久久国产精品麻豆| 国产乱来视频区| 久久久久视频综合| 性高湖久久久久久久久免费观看| 欧美xxxx性猛交bbbb| 爱豆传媒免费全集在线观看| 伦理电影大哥的女人| 乱系列少妇在线播放| 亚洲av.av天堂| 亚洲,欧美,日韩| 精品一区二区三区视频在线| 亚洲伊人久久精品综合| 亚洲在久久综合| 欧美日韩av久久| 搡女人真爽免费视频火全软件| 欧美老熟妇乱子伦牲交|