• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    SmokerViT:A Transformer-Based Method for Smoker Recognition

    2023-12-12 15:50:02AliKhanSomaiyaKhanBilalHassanRizwanKhanandZhonglongZheng
    Computers Materials&Continua 2023年10期

    Ali Khan,Somaiya Khan,Bilal Hassan,Rizwan Khan and Zhonglong Zheng,?

    1College of Mathematics and Computer Science,Zhejiang Normal University,Jinhua,321004,China

    2School of Electronics Engineering,Beijing University of Posts and Telecommunications,Beijing,100876,China

    3Department of Electrical Engineering and Computer Science,Khalifa University of Science and Technology,Abu Dhabi,127788,United Arab Emirates

    4Key Laboratory of Intelligent Education of Zhejiang Province,Zhejiang Normal University,Jinhua,321004,China

    ABSTRACT Smoking has an economic and environmental impact on society due to the toxic substances it emits.Convolutional Neural Networks(CNNs)need help describing low-level features and can miss important information.Moreover,accurate smoker detection is vital with minimum false alarms.To answer the issue,the researchers of this paper have turned to a self-attention mechanism inspired by the ViT,which has displayed state-of-the-art performance in the classification task.To effectively enforce the smoking prohibition in non-smoking locations,this work presents a Vision Transformer-inspired model called SmokerViT for detecting smokers.Moreover,this research utilizes a locally curated dataset of 1120 images evenly distributed among the two classes (Smoking and NotSmoking).Further,this research performs augmentations on the smoker detection dataset to have many images with various representations to overcome the dataset size limitation.Unlike convolutional operations used in most existing works,the proposed SmokerViT model employs a self-attention mechanism in the Transformer block,making it suitable for the smoker classification problem.Besides,this work integrates the multi-layer perceptron head block in the SmokerViT model,which contains dense layers with rectified linear activation and linear kernel regularizer with L2 for the recognition task.This work presents an exhaustive analysis to prove the efficiency of the proposed SmokerViT model.The performance of the proposed SmokerViT performance is evaluated and compared with the existing methods,where it achieves an overall classification accuracy of 97.77%,with 98.21%recall and 97.35% precision,outperforming the state-of-the-art deep learning models,including convolutional neural networks(CNNs)and other vision transformer-based models.

    KEYWORDS Smoker recognition;SmokerViT;deep learning;transformer for vision

    1 Introduction

    The smoking epidemic is one of the world’s significant public health threats,killing more than 8 million people yearly,including 1.2 million from passive smoking.In 2020 statistics[1],22.3%of the world’s population smoke,and 80%of those 1.3 billion smokers worldwide are in low-and mediumincome countries.According to a report about different causes of death worldwide,smoking is the second most significant risk factor for death [2].Smoking monitoring and preventive policies are included as actions that should be implemented in the World Health Organization(WHO)framework convention on smoking control[3].Therefore,detecting smokers in no-smoking areas is essential for effective surveillance.

    Traditional surveillance methods for smoker detection are inefficient and affected by various factors,thus limiting the development of intelligent surveillance[4,5].Researchers have continuously applied different methods to surveillance systems to answer these problems and benefit from artificial intelligence technology[6,7].Deep learning is the state-of-the-art(SOTA)artificial intelligence method that has become integral to computer vision [8].Compared to traditional image processing and machine learning methods,deep learning does not require complex image pre-processing.The Deep Neural Network (DNN),which employs deep learning techniques [9],significantly improves object detection efficiency by automatically learning the features from raw data.Convolutional Neural Networks(CNNs),a subset of DNNs,have been widely utilized to classify and cluster images based on similarity and recognize objects in scenes.CNNs have encouraged the exponential rise in deep learning as it enables significant advancements in many exciting applications,including surveillance[10,11],medical diagnosis[12],self-driving cars[13],etc.

    Since the first CNN model AlexNet [14],resulted in faster training time efficiency,new CNN models are proposed with improved accuracy with fewer parameters.Early models,such as the Visual Geometry Group (VGG) [15],had many neurons and parameters,which may result in overfitting and involve enormous computational resources.With the application of residual blocks,the training efficiency of CNN models was improved with some widely used SOTA models such as ResNet[16],Inception[17],and DenseNet[18].All the previously proposed CNN methods employed in different applications showed that accuracy is critical for applying deep learning in computer vision.

    CNN models have become a vital tool in computer vision-based surveillance applications.Convolution layers were previously widely used as the fundamental building block;however,current trends of adding attention processes have prompted researchers to rethink this.In addition to assisting CNNs with long-range dependencies,attention may replace convolutions to provide SOTA performance on computer vision tasks [19].Recently,researchers have examined using self-attention in vision-based tasks because of its potential for word-dependency learning abilities[20].Self-attention helps to learn complex relations between neighbours and their further neighbours,which may help with the binary classification problem.

    CNNs need help in describing low-level features and can miss important information.Moreover,accurate detection is vital with minimum false alarms.Considering the need for Artificial intelligence(AI) based surveillance mechanism for smoker recognition in no-smoking indoor and outdoor environments,this study focuses on the interpretation of self-attention and multilayer perceptron(MLP)head for a better understanding of the employed deep learning method.To accurately recognize smokers,this work introduces the method SmokerViT for smoker recognition in smart city indoor and outdoor environments where the Transformer component learns the convolution-like features.In SmokerViT,the patch extractor is 16 × 16 convolution with stride 16.Its output is then multiplied by learnt weights to form q,k,and v embeddings of the self-attention layer.Moreover,the MLP residual block is a linear layer in SmokerViT that raises the embedding dimension by a factor of 4,adds non-linearity then lowers it back to the original.Further linear kernelL2 classifier is used for classification.The proposed SmokerViT model uses these threefold attributes,resulting in a more robust recognition system for Smoking and NotSmoking images.The novelty of this research is to develop a deep learning-based SmokerViT model for Smoker recognition with higher accuracy.The main contributions of this research are:

    ? This research utilizes the smoker detection dataset,which has 1120 images evenly distributed among the two classes(Smoking and NotSmoking).This research performs augmentations on the dataset to have a considerable number of images with various representations to overcome the dataset size limitation.

    ? This paper presents a novel end-to-end deep learning model called SmokerViT,which integrates transformer blocks and MLP head with a fully connected layer to learn complex features and linear kernel 2 regularizer for recognizing smokers.Moreover,SmokerViT,due to its discriminative multi-head self-attention,possesses the intrinsic capabilities to classify images irrespective of the backgrounds,image quality,and noisy artefacts.

    ? This research performs exhaustive analysis to optimize the SmokerViT model to achieve the best performance on the test dataset.It can facilitate future research as a starting point for efficient smoker recognition methods.

    ? The performance of the proposed SmokerViT model is compared with different deep-learning models on the smoker detection dataset.This work uses several evaluation metrics to assess the performance of the SmokerViT model,where it outperforms the existing state-of-the-art solutions in classification accuracy.

    The research paper is organized as follows:Section 2 details the related work associated with this research,Section 3 gives the details of materials and methods adopted for solving the recognition task for the smoker detection problem,and Section 4 offers a detailed performance analysis of the proposed method and comparison with other methods and Section 5 concludes this research.

    2 Related Work

    There has been some research on various applications of surveillance using computer vision based on different proposed CNN methods.These computer vision applications include human activity recognition,pedestrian detection,traffic monitoring,face recognition,vehicle identification,fire detection,motion detection,medical imaging,etc.Authors in[21]compared state-of-the-art machine learning algorithms for insurance fraud detection.The proposed study’s decision tree algorithm performed best for the considered task.Similarly,authors in[22]proposed an improved particle swarm optimization method for data classification.Their proposed method has been tested to optimize the weight of the feed-forward neural network for fifteen datasets.Another research[23]proposed CNN based model for person head detection for counting the crowd in sports videos.Their proposed method solves the multi-scale problem,which is the object detection problem’s core issue.

    The smoker detection problem is relatively new and less explored,possibly due to the unavailability of open-access image/video datasets.Authors in [24] proposed a deep learning method based on YOLOv3-tiny named Improved YOLOv3-tiny to solve the problem of indoor low-precision smoke alarms on their local dataset.The proposed method combined the advantages of YOLOv3 and YOLOv3-tiny in terms of fewer parameters and higher accuracy for the localization task on their local smoker dataset.The proposed method considered the performance metrics of mAP for the localization task.It showed 85% mAP for improved YOLOv3-tiny compared to YOLOv3-tiny,which was 74%.However,their work limitations are the low mAP and the unavailability of the dataset.Another similar method[25]was proposed,named Eye-Smoker,YOLOv3 based transfer learning method for smoker detection on their local dataset.In the proposed method,the smoker is detected based on the cigarette and does not consider other kinds,such as e-cigarettes and smoking pipes.Their proposed method considered the localization task for smoker detection with 90%accuracy and around 94%mAP.Their limitations of work are the low accuracy and unavailability of the dataset.These object detection methods promote fast localization capabilities but lack high accuracy.

    For the classification problem,false alarms should be kept minimal.A significantly high rate of false alarms in one class and a higher number of accurate classifications in another can lead to higher prediction accuracy;however,it might lack solving the desired problem.In[26],the authors proposed a SmokingNet model based on GoogleNet for smoker detection problems on their local dataset.Their work focused on evaluating the performance of smoking and not-smoking image classification with different performance metrics.In their proposed method,the smoking image characteristics are optimized based on the GoogleNet,and the feature extraction ability is enhanced using kernels of non-square convolution.The proposed method achieved 90% accuracy,90% precision and recall,and 90% F1 measure.Their work limitations are using very basic GoogleNet as a base model and the unavailability of the dataset.In previously published work[27],the research proposed Inception-ResNet-V2-based transfer learning,where the pre-trained model was used as a backbone network for the smoker detection problem on the local smoker detection dataset.In the proposed method,the Inception-ResNet-V2 model is used,which is trained on the ImageNet dataset,the weights of the pre-trained Inception-ResNet-V2 are frozen,and new fully connected layers are added with ReLU and sigmoid activation functions.The fully connected layers learn the specific features of the task of smoker detection.The proposed method fed the complete image with an input size of 224×224 to the network.The neural network extracted the features based on the previously learned generic features trained on the ImageNet dataset.The proposed solution has a training accuracy of 95.65%and 96.87%testing accuracy with a recall of 97.32%and precision of 96.46%,discriminating the images of the Smoking and NotSmoking classes.However,the proposed work had high accuracies;still,it lacked training the model from scratch and better learn the low-level features.

    To solve the parallel processing of words by using self-attention in Recurrent Neural Network(RNN)models,a network called Transformer based on attention mechanism and removes recurrence and convolutions was proposed[28],which accomplished great success in natural language processing(NLP).After its success in NLP,an image classification model,Vision Transformer (ViT) [29],was introduced in computer vision,disrupting the traditional CNN model with its competitive performance on large-scale image datasets.With the development of transformers for computer vision in 2021,there has been some research for computer vision applications using vision-based transformers[30–32].Transformers have seen much growth in image classification tasks with accuracy similar to if not more than,CNN models.In [33],the authors proposed a multi-instance vision transformer named MITformer for remote sensing scene classification.In their proposed method,the local feature response was highlighted for the remote sensing scenes.Attention-based MLP was inserted at the end of each encoder to enhance these features.Another work [34] proposed a hybrid CNN and ViT method,CTNet,to classify high-resolution remote sensing (HRRS) images.The proposed method has two modules,T-stream(stream for ViT)and C-stream(stream for CNN).In the T-stream,the flattened patches of the image are sent into the pre-trained ViT for semantic features in HRRS images.At the same time,C-stream is used to extract the local features.Ma et al.[35]proposed a homo-heterogeneous transformer learning (HTTL) for remote sensing scene classification.In the proposed HTTL,a patch generation module is used to design homo-and heterogeneous patches.The feature learning module extracts the feature information of global and local areas.A fusion submodule and metric learning-based classification module are used for the scene classification.

    In other computer vision applications,authors [36] proposed detecting rain and road surface conditions using vision transformers.In their proposed method,a spatial self-attention network is proposed to analyze the relationship between the detection results of adjacent images as a sequenceto-sequence detection task.Dong et al.[37]proposed ViT based representation learning method for polarimetric synthetic aperture radar (PolSAR) image classification.In the proposed method,the ViT learned the global features of the PolSAR images,which improves the classification efficiency.In[38],the authors proposed a multilabel vision transformer ForestViT for multilabel classification of satellite images of deforestation,which adopts a self-attention mechanism,replacing the convolution operations.Wang et al.[39] proposed a double output vision transformer (DOViT) for air quality classification.The tokens are processed with multilabel self-attention (MSA) to extract features for higher accuracy.Authors[40]proposed Transformer based LPViT for classifying and detecting defects in printed circuit boards(PCBs).The proposed method used labels for better model strategy and mask patch prediction to ensure the relationship of different patch extractions.

    However,deep learning-based algorithms were formerly thought of as a black box,and there have been issues with their interpretability for a long time[41–43].CNNs face problems describing the lowlevel features outside the actual area of interest[44].Taking advantage of the context information for feature extraction is not beneficial.This work exploits self-attention blocks and MLP head,which are building blocks of the proposed method for visually interpreting Smoking and NotSmoking images.In the proposed method SmokerViT,which is based on Transformer and MLP head,the smoker recognition problem is considered.It achieves better prediction accuracy without convolutions than the previously proposed CNN methods.

    3 Materials and Methods

    This section details the proposed SmokerViT and the image dataset used for Smoking and NotSmoking classes for recognition problems.The following subsections explain the methodology of this work.

    3.1 Dataset Acquisition

    The dataset for this study is the smoker detection dataset published online as open access,which has different images of people smoking and not smoking indoors and outdoors.To the best of our knowledge,there is no other open-access dataset related to the problem;the smoker detection dataset facilitates future work in proposing new methods.The dataset can be accessed from[27].

    3.2 Dataset Distribution

    The smoker detection problem is considered binary,with two classes named,Smoking and NotSmoking.The NotSmoking class images are labeled 0,while the Smoking images have a class label 1.The smoker detection dataset is balanced and has 1120 images,with 560 images each in the Smoking and NotSmoking classes.This research splits the dataset into training and testing with a ratio of 80:20 with equal distribution from both classes.The training data is further divided into training and validation,with 716 images belonging to training samples and 180 for validation.

    3.3 Proposed Method

    Smoker detection in no-smoking areas is a difficult task where many factors influence the development of an AI-based surveillance system.The smoker recognition problem is solved by using a smoker detection dataset.Higher detection accuracy needs a large dataset for training,although applying deep learning models significantly improves the results.If the dataset is too small,the model is at risk of over-fitting,which means it cannot generalize effectively and will result in poor performance on a new dataset.Therefore,to train a small dataset for deep learning,this work performed data augmentations to have multiple training samples to overcome the dataset size limitation.This research implemented various augmentation processes on the training dataset,as given in Table 1.This work performed various augmentations such as resizing,scaling,flipping,shifting,etc.,as illustrated in Fig.1.All the images in the dataset are resized to a uniform resolution of 224 × 224.After that,augmentations are applied,such as vertical and horizontal shift by a factor of 0.2,50°rotation,zoom by a 0.2 factor,shear transformation,and horizontal flip by 0.2 factor.Some sample augmentation images are depicted in Fig.1.

    Figure 1:Sample data augmentations

    Table 1:Data augmentation

    To efficiently execute the recognition task,this research proposes SmokerViT inspired by Vision Transformer [29].CNNs have been pivotal in solving the problems of computer vision-based applications.In CNN,the pixels of the image are interdependent,and instead of all pixels features being trained on,only extracted features from the image patches using filters are being used for training.However,if complete data of images are used for training,the chances of obtaining the best performance become higher,which is the main work of the Transformer for vision-based applications.In the proposed SmokerViT,the work first converted the image into patches of size 16 × 16.In Transformer,the patches should be of the size that gives the equal rows and columns of the patches in the image.The image size of 224×224 and 16×16 patch size will give 14×14,196 patches per image.After the conversion into patches,it is passed to the Transformer encoder for processing.After that,the output is passed into the Multi-Layer Perceptron(MLP)head,which in the proposed SmokerViT consists of flatten,dense layer with ReLU function and a classification layer with kernel regularizerL2 to output the prediction depicted in Fig.2.

    Figure 2:Working mechanism of SmokerViT

    In SmokerViT,Vision Transformer takes the series of patches of images as input and predicts the class labels for the input image.The transformer differs from traditional CNNs,which do computations using pixel arrays.The Transformer divides the image into patches of fixed size.Then it inputs these patches into a linear projection of flattened patches embedding layer to produce vectors often known as tokens.These tokens precede a series of tokens.Additionally,the location data is provided by the position embedding.The Transformer encoder will receive these tokens as embedded patches and the location data.The Transformer encoder has the same number of outputs as inputs.The output corresponding to the class is then entered into the MLP head to output the prediction and classification.The architecture of SmokerViT is illustrated in Fig.3.

    Figure 3:Architecture of SmokerViT

    To consider the operation of SmokerViT in detail;first,the input imageXwith dimensionh×w×cis divided into several patches ofXpasn×(p2.c),wherehandwrepresent the image resolution of input while(p,p)represents the image patch resolution,cdenotes the number of channels,andrepresents the number of image patches,and this is the input sequence length for the model.These patches are then passed through a linear projection and mapped to theddimension to get the output referred to as patch embedding.The position embeddingEposis added to the patch embeddingEto keep the position information of the input.It is expressed asEpos∈R(n+1)×d,which joins the[class]token=Xclass.Its form at the output of Transformer encoderworks as image representationY.The Transformer encoder contains multi-head self-attention (MSA),layer normalization (LN) and MLP block.

    MSA: This layer linearly integrates the attention output.The encoder receives a sequence of embedding to process,which undergoes three different linear transformations to output the three vectors queryq,keyk,and valuev.The attention output for each embedding is calculated by the dot product of these three vectors.Self-attention is calculated independently and repeatedly in parallel.As a result,it is known as multi-head attention.The attention measures how strongly the patches are connected,subsequently assisting in prediction.The MSA is calculated by the equation given below:

    wheredis the dimension ofk,andwiis the learnable weights.

    LN: Layer normalization balances the mean-variance of each input neuron layer,making it converge faster.Layer normalization is added before each block,as it has no prior image dependencies,so it enhances the performance and decreases the execution time.

    MLP:The MLP in the Transformer encoder consists of two layers with GeLU.

    MLP head:After the Transformer encoder,the output is inserted into the newly added MLP head for the classification of Smoking and NotSmoking images,which consists of flatten layer to flatten the encoder output,dense layers with ReLU activation and linear kernelL2 regularizer as a classifier.

    3.3.1 Activation Function

    The activation function optimizes the processes and learns complex features specific to the task.The proposed method considers ReLU (R) activation function.Ris a piecewise linear function that outputs the input directly if it is positive;otherwise outputs zero and is given by:

    3.3.2 Optimizer

    This work considers RMSProp (Root Mean Squared Propagation) optimizer for the proposed SmokerViT method.RMSprop applies the exponential moving average of the squared gradients to adjust the learning rate.RMSprop only accumulates gradients in a specific fix window instead of letting all the gradients accumulate for momentum.The equation for RMSprop is as follows:

    whereηshows the learning rate,?represents the small term preventing division by zero,E[g2]is the past squared gradients RMSprop running average,andgtis the gradient function.

    3.3.3 Loss Function

    The binary cross entropy loss function is often used for binary classification.It helps to evaluate model accuracy by calculating the prediction probability.Following is the equation for the binary cross entropy loss function:

    wherezrepresents the label,i.e.,1 denotes the Smoking class and 0 denotes NotSmoking class,and ?zis the predicted probability ofz.

    3.3.4 Linear Kernel L2as a Classifier

    This work used the linear kernelL2 algorithm because it can help solve problems with multicollinearity (highly correlated independent variables) by limiting the coefficient and maintaining all the variables.Linear kernel,basic kernel,is the best in case of many features and given byf(whereX,Xjis the data to classify.Linear kernelL2 predicts based on the mean of data to avoid overfitting,unlikeL1,which takes the median of data for estimation.L2 adds the penalty to the cost function as the squared value of the weights and learns complex patterns.L2 is computationally efficient,and predictions are more accurate when the output is a function of all input variables.L2 regularization is calculated by:

    wherewiis the weight andλrepresents the regularization parameter.Ifλis 0,this acts as Ordinary Least Square(OLS),where it will make the weight coefficient 0 and result in underfitting,while ifλis very large,it will increase the weight and result in underfitting.

    4 Performance Evaluation

    The performance of the proposed SmokerViT for smoker recognition is evaluated and compared with other methods on the smoker detection dataset.The system configurations for simulation are i7-11800H,16 GB DDR4,NVIDIA RTX3060 6 GB,and the simulation setup is Anaconda Python 3.8 with Tensorflow 2.6 and Keras 2.3 libraries.The proposed SmokerViT is tested with various hyperparameters values for the best results.Table 2 depicts the hyper-parameters for the simulations.

    Table 2:Simulation parameters

    4.1 Evaluation Metrics

    This section presents the evaluation metrics for analyzing the performance of the proposed SmokerViT method.This work evaluated the methods on the following metrics:

    TheTnis the true negative,accurately classified as NotSmoking images,whileTpis the true positive,accurately classified as Smoking images by the proposed solution.Fnis the false negative where the Smoking image is categorized as NotSmoking,andFpis the false positive where NotSmoking images are labelled as Smoking.Precisionis the percentage of correctly positive outcomes to those the classifier predicted correctly,whereas the ratio of correctly positive results to all the relevant samples that should be positive is known as theRecallorSensitivityof the proposed method.The ratio of correct negative predictions to the results that the classifier predicts as negative is known asSpecificityor true negative rate.FDris the total number of false positive classifications to the total positive classifications.The F1 score is the harmonic mean of precision and recall,which shows how the classifier predicts correctly.FPris the ratio of negatives falsely categorized as positives and the total number of actual negatives,whileFNris the ratio of positives being falsely classified as negatives and the total number of actual positives.Eris the error rate,the ratio of all the incorrect predictions to the total number of test samples.

    4.2 Attention Maps of Learned Features

    This subsection interprets the visualization of the proposed method to understand the smoker recognition mechanism better.This research visualized the attention maps of some sample images predicted for the smoker recognition tasks from the self-attention block,illustrated in Fig.4.Selfattention is the main reason the Transformer integrates data across the complete image,including in the base layers.The attention maps show how well the method utilizes this capacity for the considered task.Some attention heads already focus on the desired representations on most images in the initial layers,demonstrating that the proposed method employs the capability to integrate information globally.The attention to the desired features increases with the model depth,and it becomes clearer what features the model pays attention to for the desired task.The original input images are converted into pseudocolour images to highlight the attention mask applied to the input image.It can be seen from the sample of images considered for both the Smoking and NotSmoking classes that the brighter part represents the attention mapping of the proposed method.Globally this research discovers that the method pays attention to image areas that are significant for classification in terms of semantics.From the maps,it can be noted that the attention of the method is on the cigarettes and their smoke for the detection of smokers.Similarly,the absence of cigarettes and smoke around the person’s mouth or hand is predicted to be a NotSmoking image.

    Figure 4:Sample images of smoking and NotSmoking(a,d)original and(b,c)attention maps

    4.3 Performance Analysis of SmokerViT

    This section presents the performance analysis of the proposed SmokerViT,a method based on the Transformer and MLP head.This work was analyzed by using different regularizers as a classifier to prove the effectiveness of linear kernel regularizer(L2)over Gaussian kernels.Moreover,this work proves the efficiency of the proposed method with and without the proposed MLP head and with and without data augmentation with simulation settings presented in Table 2.

    Table 3 presents the performance of SmokerViT in terms of prediction accuracy using different kernels.The result shows that using linear kernelL2 SmokerViT obtained the best result with 97.77%overall prediction accuracy,while the Gaussian kernel displayed overall prediction accuracy of 94.64%.It can be noted that linear kernelL2 showed the best accuracy for the Smoking and NotSmoking classes with 98.21%and 97.32%,respectively,whereas the Gaussian kernel showed 93.75%Smoking and 95.54%accuracy for NotSmoking class.

    Table 3:Performance of SmokerViT using different kernels

    This work considered the best result obtained on linear kernelL2 for the SmokerViT.This research performed further analysis by removing the MLP head and replaced with a sigmoid as a classifier with a single output to demonstrate the efficiency of the proposed model.Table 4 presents the performance of the proposed method with and without MLP head and augmentation and without augmentation block.It can be observed that the proposed model with MLP head and augmentation has the best results,whereas without augmentation performed worst for both the classes and overall prediction accuracy.The performance of SmokerVit with the proposed MLP head is improved using augmentation to 97.77% from 95.54% without the MLP head.This is because the augmentation significantly increased the dataset size with various representations,which helped predict unseen Smoking and NotSmoking images in the test dataset.While without augmentation,and proposed MLP head has significantly low performance with 88.39%overall prediction accuracy.

    Table 4:Performance of SmokerViT with and without augmentation and MLP head

    Moreover,this work performed simulations to show the effectiveness of using a Transformer as the main network over other SOTA models with anL2 kernel classifier.This work used ResNet,Inception-ResNet-V2 model to show the efficiency of using Transformer over these models.Table 5 shows that the best results are achieved using Transformer with 97.77% accuracy,while the second best results are achieved using Inception-ResNet-V2 with 96.43%accuracy,followed by InceptionV3 with 87.05% accuracy.ResNet performed worse with 85.71% accuracy.It can be observed from the table that SmokerViT has more parameters than the other models;however,the focus of this study is the higher accuracy.

    Table 5:Performance comparison of using different models for feature extraction

    The time complexity of the proposed SmokerViT model can be expressed asO(N2L),whereNrepresents the number of patches in the input image,andLrepresents the number of self-attention layers in the transformer.

    TheO(N2L)notation arises from each self-attention layer having a quadratic complexity ofO(N2),as it involves computing pairwise dot products between all pairs of patches.Since the SmokerViT model hasLself-attention layers,the total time complexity isO(N2L).

    It is worth noting that the SmokerViT model also has additional computational costs associated with the feedforward network and positional embeddings,but these are typically negligible compared to the self-attention computation.

    The loss performance curves of the proposed SmokerViT in terms of training loss and validation loss are depicted in Fig.5.The training phase of the proposed SmokerViT is carried out through 50 epochs.From the result,it can be noted that the training loss started at 76.24%and achieved less than 19.56%loss at the 10th epoch.After 10 epochs,the loss curve remained steady till the 50 epochs,with a final loss of 8.28%.Similarly,the validation loss started at 30.82% and reached 9.62% at the 10th epoch.After that,the loss curve remained steady till 50 epochs,with a final loss of 5.39%.

    Figure 5:Loss performance of the proposed SmokerViT

    The confusion matrix depicts the predictive analysis of Smoking and NotSmoking image classification.It can be seen from the confusion matrix in Fig.6 that the proposed SmokerViT displayed a prediction accuracy of 97.77% and 2.23% error rate with 109Tnand 110Tpwith 3Fpand 2Fn,respectively.Table 6 shows the performance of the proposed SmokerViT on individual classes.The proposed method achieved 98.21% prediction accuracy,97.35% precision,98.21% recall,and 97.78% F1 score for the Smoking class.Whereas for the NotSmoking class,the proposed method displayed 97.32%prediction accuracy,98.20%precision,97.32%recall,and 97.76%F1.The proposed SmokerViT displayed the overall performance with 97.77%prediction accuracy,98.21%recall,97.35%precision,and 97.78% F1 measure for classifying Smoking and NotSmoking images of the smoker detection dataset.

    Figure 6:Confusion matrix of SmokerViT

    Table 6:Performance of SmokerViT

    Receiver Operating Characteristic(ROC)curve,shown in Fig.7a,is another graphical representation for assessing the performance that shows the proposed method’s ability to predict classification with varying prediction thresholds.The ROC curve is plotted by considering Recall(TPr)on the y-axis againstFPron the x-axis.The Area under the Curve(AUC)depicts how well the method differentiates between the classes.The AUC of 0.9948 by the SmokerViT means that it has a 99.48% chance of accurately classifying the Smoking and NotSmoking classes.This work also analyzed the proposed SmokerViT based on the Precision-Recall (PR) curve,which depicts how well it performed for classifying Smoking images because,unlike ROC,the PR curve does not considerTnfor performance evaluation.PR curve also depicts whenever the class distribution has variation,unlike ROC,which shows no change.It can be noted from Fig.7b that the curve is near the top right corner,showing that the SmokerViT performed well in classifying the Smoking class.SmokerViT achieved 99.51%average precision(AP)for the Smoking and NotSmoking classification.

    Figure 7:(a)Receiver operating characteristic curve with AUC and(b)precision-recall curve with AP

    For the smoker recognition problem,false negatives should be minimal.From Fig.8b,it can be seen that the false negative occurred when the background of the image is similar to the person in the image.In computer vision,spatial resolution is crucial,which has led to the inaccurate classification of the Smoking image as NotSmoking.Better-quality images let the model generalize more accurately.In addition,the neural network had trouble distinguishing between cigarette and the background pixels in the images when the background was blurry.The lack of a considerable number of varying images in the training set might also account for the false alarms.Another possible reason is that some photos in the test set were new to the model and lacked representation of comparable images in the training set.The model performed poorly in generalizing the novel scenarios.

    Figure 8:(a)False positives and(b)false negatives

    Subsequently,a similar problem was seen with false positives.Smoker recognition relies heavily on the accuracy and practicality of the classifier,both of which are affected by the number of false positives.Images of false positives are shown in Fig.8a.The scarcity of diversity in the training set and the variety of datasets may result in the incorrect classification of some NotSmoking images as Smoking.It can also be observed image with the background as a cloud was misclassified as Smoking;moreover,a similar hand gesture to the smoking was also labeled as Smoking.

    4.4 Comparative Study with Other Methods

    For validating the effectiveness of the proposed SmokerViT,this research compared the performance with other methods,both CNN and Transformer based models such as EfficientNetV2[45],ResNest [46],MobileNetV3 [47],ResNetD [48],ViT [29],Levit [49],Davit [50] and Coatlite[51] on the smoker detection dataset.The hyperparameters are listed in Table 7.Table 8 presents the comparative analysis of these methods.It can be observed that SmokerViT displayed superiority over other considered methods in classifying the Smoking and NotSmoking classes.All the methods considered for comparison were used as pre-trained models using transfer learning and added the classification layer with a sigmoid activation function.After SmokerViT,ViT performed better among all the other considered methods for classification tasks on the local dataset for the smoker detection problem,as explained in Table 6.ViT achieved 96.43% accuracy,96.43% sensitivity,and 96.43%specificity,followed by Levit with 94.64%accuracy,Coatlite with 91.07%accuracy,Davit with 90.18%,ResNetD with 88.40%accuracy,ResNest with 86.61%accuracy,EfficientV2 with 85.27%and at last MobileNetV3 with 82.14%accuracy.It can be observed that ResNetD and Davit outperformed ViT in terms of sensitivity at 97.32% compared to ViT at 96.43%.However,MobileNetV3 performed worse in terms of specificity than other models due to the significantly large number of false positives and considerably lower number of false negatives.The proposed method performed best among other considered methods in all evaluation metrics.MobileNetV3 has the lowest accuracy of 82.14%compared to other CNN methods for analyzing the unique smoker classification problem.Fig.9 shows the performance comparison of all the methods on individual classes.SmokerViT achieves the best results on both classes,followed by ResNetD and Davit for Smoking class with 97.32%accuracy while ViT for NotSmoking class with 96.63%accuracy.

    Figure 9:Performance comparison in terms of classes of SmokerViT and other methods

    Table 7:Hyperparameters of the models

    Table 8:Comparative analysis of SmokerViT with other methods on smoker detection dataset

    4.5 Comparative Study with Previous Work

    This work compares the performance of the proposed SmokerViT with our previous work[27].Table 9 shows SmokerViT has a better overall prediction accuracy of 97.77%compared to 96.87%by our previously proposed method InceptionResNetV2.It can be noted that SmokerViT has improved the performance of smoker recognition for both classes,where it displayed 98.21% and 97.32%accuracies in discriminating the Smoking and NotSmoking images,respectively.InceptionResNetV2 achieved 97.32%and 96.43%accuracies for the Smoking and NotSmoking classes.Fig.10 compares the two methods in terms of precision,recall,and F1 measure.It can be observed that SmokerVit performed better in all the considered performance metrics,which is because the self-attention mechanism focuses on the entire image,unlike convolutions in CNN that focus on the interpretation of the high-level features rather than low-level features in classifying Smoking and NotSmoking images.

    Figure 10:Performance of SmokerViT and InceptionResNetV2

    Table 9:Comparative analysis of SmokerViT with previous work on smoker detection dataset

    In this research work,SmokerViT displayed self-attention capability and MLP Head to recognize Smoking and NotSmoking images.The results show that better performance is achieved by SmokerViT compared to the SOTA models,implying that the self-attention mechanism and MLP Head architecture may be more suitable than CNN for the Smoker recognition problem.In contrast to transformers,which can compute the attention of any patch,regardless of its distance,a CNN needs to perform additional convolutions to increase the receptive field to determine the relationship between any neighboring pixels,resulting in difficulty in possessing the ability to perform longrange computation.In SmokerViT,the patch embedding component is used to learn convolution-like features,whereas self-attention is used to learn important features and ignore the noisy ones.Results show that the SmokerViT performed better than CNN and Transformer based models,validating the superiority of using both the self-attention mechanism and MLP Head.

    While looking at the results,it can be observed that the SmokerViT has performed better in both the Smoking and NotSMoking classes.However,CNNs models were better at predicting Smoking images while poorly classifying NotSmoking class compared to Transformer based models,which performed well in classifying both the Smoking and NotSmoking classification.While SmokerViT was equally good in classifying both classes indicating that SmokerViT is more robust than using CNN or transformer-based models in dealing with balanced datasets.

    Many researchers used CNN for the smoker detection problem,but there is not much work on this problem.This is the first time using transformers and MLP Head with Linear kernelL2 classifier for the smoker recognition task.Additionally,previous research results were compared with this proposed work,as depicted in Table 9.It can be observed that SmokerViT outperformed the Inception-ResNet-V2 model for Smoking and NotSmoking image classification using the same dataset.Moreover,the high accuracy displayed by the proposed method can help an AI-based smoker detection system and save time and human resources simultaneously.This research can benefit researchers to improve further the methodology for image segmentation to detect cigarette smoker detection problems.

    5 Conclusion

    This research proposed a transformer-based smoker recognition method.For effective surveillance of the no-smoking areas,this research proposed SmokerViT based on the self-attention mechanism instead of CNN.The dataset for this work has two classes with 560 images each for the Smoking and Notsmoking classes.Further,this work performed augmentations on the smoker detection dataset to have many images with various representations to overcome the dataset size limitation.The proposed SmokerViT is inspired by Vision Transformer and adding our own MLP head block,which has a dense layer with ReLU activation function and linear kernelL2 regularizer as a classifier.SmokerViT extracted features through long-range dependency compared to CNN models,which took advantage of useful global information.Ablations were performed on the proposed SmokerViT to prove the efficiency of the MLP head block and data augmentation.The SmokerViT performance was evaluated and compared with the previously proposed CNN model for the smoker detection problem and other Transformer and CNN-based methods.The SmokerViT achieved a 0.93%higher accuracy of 97.77%,with 0.92%better recall of 98.21%and 0.92%better precision of 97.35%compared to the previous proposed Inception-ResNet-V2 based transfer learning method.Moreover,the results showed that SmokerViT achieved competitive performance compared to other models with considerably higher values of the evaluation metrics.

    For future works,several issues that were not addressed in this research need to be observed.The effect of dataset size on the training performance of the method and data augmentation by various complex models for ensuring further variances of the image representations can be considered for future study.Moreover,the weak point of this research is the higher number of parameters and high execution time.In future work,this point would be considered for designing the model,which is lightweight and,at the same time,yield higher accuracy.Moreover,the hybrid method of convolutions and transformer might help the smoker recognition system to perform better,considering the shortcomings of the proposed method.

    Acknowledgement:Authors are thankful to the reviewers for their valuable comments.

    Funding Statement:The authors received no specific funding for this study.

    Author Contributions:The authors confirm contribution to the paper as follows: study conception and design:A.K.;data collection:A.K.;analysis and interpretation of results:A.K.,S.K.,B.H.and R.K.;draft manuscript preparation:A.K.,S.K.,B.H.and R.K.;writing—review and editing,A.K.,S.K.,B.H.,R.K.and Z.Z.;funding acquisition,A.K.and Z.Z.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The dataset considered in this study can be accessed from https://data.mendeley.com/datasets/j45dj8bgfc/1.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    波多野结衣高清无吗| 青草久久国产| 一区福利在线观看| 亚洲九九香蕉| 最近在线观看免费完整版| 两人在一起打扑克的视频| 在线看三级毛片| 中文资源天堂在线| 成人18禁高潮啪啪吃奶动态图| 操出白浆在线播放| 又大又爽又粗| 中文字幕熟女人妻在线| 日韩欧美国产一区二区入口| 日韩成人在线观看一区二区三区| 中文字幕熟女人妻在线| 久99久视频精品免费| 91老司机精品| 欧美三级亚洲精品| 国产私拍福利视频在线观看| 十八禁人妻一区二区| 国产v大片淫在线免费观看| 一区二区三区国产精品乱码| 手机成人av网站| 国产99久久九九免费精品| 在线观看免费日韩欧美大片| 非洲黑人性xxxx精品又粗又长| 欧美中文日本在线观看视频| 国产人伦9x9x在线观看| 中文字幕最新亚洲高清| 亚洲avbb在线观看| 亚洲国产精品久久男人天堂| 男男h啪啪无遮挡| 久久久久免费精品人妻一区二区| 不卡一级毛片| 一区福利在线观看| 国产精品九九99| 亚洲欧美一区二区三区黑人| 国产成人精品无人区| 日本免费a在线| 两个人看的免费小视频| 法律面前人人平等表现在哪些方面| 国产v大片淫在线免费观看| 久久久久精品国产欧美久久久| 日本一本二区三区精品| 午夜久久久久精精品| 亚洲av成人精品一区久久| 亚洲av成人精品一区久久| 黄色丝袜av网址大全| 国产成人av激情在线播放| 熟女电影av网| 一二三四社区在线视频社区8| 亚洲av熟女| 毛片女人毛片| 国产av不卡久久| 美女黄网站色视频| 国产成年人精品一区二区| 美女黄网站色视频| 久久精品国产亚洲av香蕉五月| 亚洲真实伦在线观看| 亚洲人成网站在线播放欧美日韩| 国产成人系列免费观看| 欧美黑人巨大hd| 美女扒开内裤让男人捅视频| 国产亚洲av高清不卡| 一进一出好大好爽视频| 日本 欧美在线| 99热6这里只有精品| 99热这里只有精品一区 | 久久精品国产清高在天天线| 亚洲专区中文字幕在线| 日韩中文字幕欧美一区二区| 伊人久久大香线蕉亚洲五| 日韩欧美精品v在线| 男人舔女人下体高潮全视频| 两个人免费观看高清视频| 免费av毛片视频| 精品乱码久久久久久99久播| 国产精品免费一区二区三区在线| 国产精品日韩av在线免费观看| 成人av在线播放网站| 国产99久久九九免费精品| 亚洲全国av大片| 一个人免费在线观看的高清视频| 精品欧美一区二区三区在线| 国产av麻豆久久久久久久| 国产精品1区2区在线观看.| 精品高清国产在线一区| www.熟女人妻精品国产| 国产探花在线观看一区二区| 校园春色视频在线观看| 亚洲男人的天堂狠狠| 99热这里只有精品一区 | 久久久国产精品麻豆| 国产成人一区二区三区免费视频网站| 9191精品国产免费久久| 中文字幕人成人乱码亚洲影| cao死你这个sao货| 欧美久久黑人一区二区| 精品电影一区二区在线| 舔av片在线| 欧美中文日本在线观看视频| 国产亚洲av高清不卡| 亚洲国产欧洲综合997久久,| 国产视频内射| 久热爱精品视频在线9| 日韩成人在线观看一区二区三区| 久久久久久久久久黄片| 国产av一区二区精品久久| www日本黄色视频网| www日本黄色视频网| 午夜福利欧美成人| 色哟哟哟哟哟哟| 人妻夜夜爽99麻豆av| 国产成人系列免费观看| 午夜免费激情av| 中文字幕久久专区| 欧美日本亚洲视频在线播放| 精品久久久久久久末码| 久久久精品大字幕| 女人高潮潮喷娇喘18禁视频| 亚洲国产欧美一区二区综合| 怎么达到女性高潮| 久久国产乱子伦精品免费另类| 欧美成人性av电影在线观看| 国内精品久久久久久久电影| 日韩欧美国产在线观看| 又黄又爽又免费观看的视频| 欧美+亚洲+日韩+国产| 黄片小视频在线播放| 天天添夜夜摸| 色播亚洲综合网| 极品教师在线免费播放| 一夜夜www| 国产真实乱freesex| 99久久综合精品五月天人人| 黑人欧美特级aaaaaa片| 成人永久免费在线观看视频| 国产视频一区二区在线看| 欧美又色又爽又黄视频| 亚洲成人中文字幕在线播放| 国产精品一区二区三区四区久久| 精品久久久久久久毛片微露脸| 精品久久蜜臀av无| 不卡一级毛片| 亚洲欧美一区二区三区黑人| 国产又黄又爽又无遮挡在线| 全区人妻精品视频| 俄罗斯特黄特色一大片| 欧美久久黑人一区二区| 亚洲国产精品sss在线观看| 久久久久久国产a免费观看| а√天堂www在线а√下载| 免费看美女性在线毛片视频| cao死你这个sao货| 一级作爱视频免费观看| 亚洲av电影不卡..在线观看| 在线播放国产精品三级| 美女 人体艺术 gogo| 在线观看美女被高潮喷水网站 | 国产单亲对白刺激| 我的老师免费观看完整版| 国产一区二区三区在线臀色熟女| www.精华液| 国产亚洲精品久久久久久毛片| 男人舔女人下体高潮全视频| 国产精品爽爽va在线观看网站| a级毛片a级免费在线| 亚洲国产欧美一区二区综合| www国产在线视频色| 欧美久久黑人一区二区| 一本综合久久免费| 每晚都被弄得嗷嗷叫到高潮| 亚洲乱码一区二区免费版| 男女下面进入的视频免费午夜| 欧美成人一区二区免费高清观看 | 欧美日韩精品网址| 亚洲国产欧美人成| 欧美另类亚洲清纯唯美| 又粗又爽又猛毛片免费看| 搡老妇女老女人老熟妇| 久久久精品大字幕| 中文亚洲av片在线观看爽| 国产欧美日韩一区二区三| 黄色女人牲交| 国产免费av片在线观看野外av| 国产精品九九99| 麻豆国产av国片精品| 法律面前人人平等表现在哪些方面| 人人妻人人看人人澡| 久久精品夜夜夜夜夜久久蜜豆 | 亚洲av五月六月丁香网| 久久久久久久午夜电影| 欧美国产日韩亚洲一区| 日本五十路高清| 一个人观看的视频www高清免费观看 | 日本 欧美在线| 啦啦啦韩国在线观看视频| 2021天堂中文幕一二区在线观| 香蕉丝袜av| 99精品久久久久人妻精品| 在线观看午夜福利视频| 久久久久性生活片| 50天的宝宝边吃奶边哭怎么回事| 夜夜看夜夜爽夜夜摸| 国产蜜桃级精品一区二区三区| 俄罗斯特黄特色一大片| 午夜精品一区二区三区免费看| 国产男靠女视频免费网站| 中国美女看黄片| 男男h啪啪无遮挡| 男女做爰动态图高潮gif福利片| 日本一区二区免费在线视频| 中文亚洲av片在线观看爽| av在线天堂中文字幕| 欧美大码av| 国产三级在线视频| 久久香蕉激情| 日本成人三级电影网站| 91成年电影在线观看| 色尼玛亚洲综合影院| ponron亚洲| 亚洲国产欧美一区二区综合| 人人妻人人看人人澡| 免费av毛片视频| 两人在一起打扑克的视频| 日韩中文字幕欧美一区二区| 亚洲黑人精品在线| 一级毛片高清免费大全| 日本精品一区二区三区蜜桃| 熟女少妇亚洲综合色aaa.| 亚洲全国av大片| 麻豆av在线久日| 亚洲国产日韩欧美精品在线观看 | 国产亚洲精品第一综合不卡| 在线观看免费午夜福利视频| 一区二区三区激情视频| 啪啪无遮挡十八禁网站| 成人永久免费在线观看视频| 美女黄网站色视频| 成年人黄色毛片网站| 99国产精品一区二区三区| 十八禁网站免费在线| 亚洲免费av在线视频| 国产精品久久久久久久电影 | 国产av不卡久久| 亚洲avbb在线观看| 欧美av亚洲av综合av国产av| 国产成人av教育| 女同久久另类99精品国产91| 啪啪无遮挡十八禁网站| 不卡av一区二区三区| 哪里可以看免费的av片| 久久午夜综合久久蜜桃| 欧美av亚洲av综合av国产av| 国内精品久久久久久久电影| 悠悠久久av| 中亚洲国语对白在线视频| 亚洲国产欧美网| 亚洲精品色激情综合| 91麻豆av在线| 男女午夜视频在线观看| 国产成人一区二区三区免费视频网站| 两个人的视频大全免费| 黄片大片在线免费观看| 欧美性长视频在线观看| 久久精品国产清高在天天线| 在线观看午夜福利视频| 午夜精品一区二区三区免费看| 精品国产乱子伦一区二区三区| 欧美激情久久久久久爽电影| 久久欧美精品欧美久久欧美| 天堂√8在线中文| 99久久国产精品久久久| 欧美一级a爱片免费观看看 | 国产精品久久久久久人妻精品电影| 亚洲第一欧美日韩一区二区三区| 免费无遮挡裸体视频| 免费在线观看完整版高清| 一边摸一边抽搐一进一小说| 中文字幕熟女人妻在线| 国产人伦9x9x在线观看| 999久久久精品免费观看国产| 舔av片在线| 制服诱惑二区| 国产成人欧美在线观看| 午夜两性在线视频| 欧美另类亚洲清纯唯美| 欧美乱色亚洲激情| 丰满人妻熟妇乱又伦精品不卡| 老司机深夜福利视频在线观看| 亚洲精品久久成人aⅴ小说| 少妇裸体淫交视频免费看高清 | 亚洲国产欧美一区二区综合| 成人三级做爰电影| 午夜激情av网站| 草草在线视频免费看| 国产精品av视频在线免费观看| 18美女黄网站色大片免费观看| 久久国产精品人妻蜜桃| 亚洲成人久久性| 国产av一区二区精品久久| 成年人黄色毛片网站| 波多野结衣巨乳人妻| 欧美+亚洲+日韩+国产| 母亲3免费完整高清在线观看| 国产在线精品亚洲第一网站| 两个人的视频大全免费| videosex国产| 首页视频小说图片口味搜索| 亚洲欧美日韩东京热| 亚洲真实伦在线观看| 天堂√8在线中文| 亚洲一卡2卡3卡4卡5卡精品中文| 国产精品一区二区免费欧美| 婷婷精品国产亚洲av在线| 99国产精品一区二区蜜桃av| 午夜成年电影在线免费观看| 日韩中文字幕欧美一区二区| 午夜福利免费观看在线| 成年女人毛片免费观看观看9| 又紧又爽又黄一区二区| 亚洲精品色激情综合| 很黄的视频免费| 国产精品免费视频内射| www日本黄色视频网| 精品免费久久久久久久清纯| 老司机午夜福利在线观看视频| 久久久久久久久中文| 色哟哟哟哟哟哟| 一个人免费在线观看的高清视频| 亚洲欧美日韩高清在线视频| 岛国视频午夜一区免费看| 精品久久久久久久久久久久久| 巨乳人妻的诱惑在线观看| 亚洲人成伊人成综合网2020| 婷婷精品国产亚洲av在线| 在线观看免费午夜福利视频| 色哟哟哟哟哟哟| 国产精华一区二区三区| 一区二区三区激情视频| 毛片女人毛片| 身体一侧抽搐| 久久久国产成人免费| 精品一区二区三区视频在线观看免费| 久久九九热精品免费| 亚洲专区国产一区二区| 久久久久国内视频| 精品少妇一区二区三区视频日本电影| 精品久久久久久久末码| 亚洲色图av天堂| 一夜夜www| 91九色精品人成在线观看| 亚洲av日韩精品久久久久久密| 啦啦啦观看免费观看视频高清| 国产私拍福利视频在线观看| 久久伊人香网站| 97人妻精品一区二区三区麻豆| 免费在线观看成人毛片| 99国产极品粉嫩在线观看| 在线观看免费午夜福利视频| 18禁观看日本| 热99re8久久精品国产| 人妻夜夜爽99麻豆av| 给我免费播放毛片高清在线观看| 精品一区二区三区视频在线观看免费| 这个男人来自地球电影免费观看| 日本成人三级电影网站| 最近最新免费中文字幕在线| 久久国产精品影院| 高清毛片免费观看视频网站| 欧美三级亚洲精品| 嫁个100分男人电影在线观看| 日本成人三级电影网站| 十八禁人妻一区二区| 三级男女做爰猛烈吃奶摸视频| 日韩av在线大香蕉| 老司机深夜福利视频在线观看| 丝袜美腿诱惑在线| 搡老熟女国产l中国老女人| or卡值多少钱| 色综合婷婷激情| 久久精品综合一区二区三区| 一级毛片精品| 露出奶头的视频| 精品久久久久久久人妻蜜臀av| 久久香蕉国产精品| 日日摸夜夜添夜夜添小说| 亚洲精品美女久久av网站| 国产一区在线观看成人免费| 欧美日韩福利视频一区二区| 日韩欧美国产一区二区入口| 国产成人欧美在线观看| 精华霜和精华液先用哪个| 久久久国产成人精品二区| 2021天堂中文幕一二区在线观| 最新在线观看一区二区三区| 日韩成人在线观看一区二区三区| 成人亚洲精品av一区二区| 午夜福利欧美成人| 国产高清videossex| 亚洲激情在线av| 黑人欧美特级aaaaaa片| 19禁男女啪啪无遮挡网站| 成人一区二区视频在线观看| 亚洲无线在线观看| 国产午夜精品论理片| 久久久精品欧美日韩精品| 黑人操中国人逼视频| 精品国产乱码久久久久久男人| 国产免费男女视频| 国产精品久久久久久人妻精品电影| 国产精品一区二区精品视频观看| 18禁裸乳无遮挡免费网站照片| 国产乱人伦免费视频| 国产精品 国内视频| 精品不卡国产一区二区三区| 午夜免费激情av| 日本黄大片高清| 亚洲人成电影免费在线| 国产男靠女视频免费网站| 亚洲成a人片在线一区二区| 在线视频色国产色| 欧洲精品卡2卡3卡4卡5卡区| 宅男免费午夜| 亚洲 欧美一区二区三区| av有码第一页| 无限看片的www在线观看| 免费观看人在逋| 欧美性长视频在线观看| 欧美成人一区二区免费高清观看 | 18禁黄网站禁片免费观看直播| 欧美高清成人免费视频www| 久久香蕉国产精品| 国产亚洲精品av在线| 国产伦人伦偷精品视频| 久热爱精品视频在线9| 国产精华一区二区三区| 999久久久国产精品视频| av片东京热男人的天堂| av在线播放免费不卡| 日本在线视频免费播放| 亚洲,欧美精品.| 亚洲在线自拍视频| 90打野战视频偷拍视频| 岛国在线观看网站| 桃红色精品国产亚洲av| 亚洲专区中文字幕在线| 亚洲免费av在线视频| 国产一区在线观看成人免费| 欧美大码av| 黄色视频不卡| 一本大道久久a久久精品| 男女视频在线观看网站免费 | 国产真实乱freesex| av天堂在线播放| 久久草成人影院| 国产高清激情床上av| 日本一本二区三区精品| 露出奶头的视频| 不卡av一区二区三区| 午夜福利在线观看吧| 在线观看66精品国产| 两性夫妻黄色片| 不卡一级毛片| 97超级碰碰碰精品色视频在线观看| 巨乳人妻的诱惑在线观看| 精品久久久久久久久久久久久| 国产成年人精品一区二区| 免费一级毛片在线播放高清视频| 51午夜福利影视在线观看| 亚洲专区国产一区二区| 国产熟女午夜一区二区三区| 精品人妻1区二区| 99久久综合精品五月天人人| 精品乱码久久久久久99久播| 亚洲中文日韩欧美视频| 人成视频在线观看免费观看| 亚洲一区二区三区色噜噜| 无人区码免费观看不卡| www日本黄色视频网| 制服人妻中文乱码| 久久精品综合一区二区三区| 天天一区二区日本电影三级| 黄片大片在线免费观看| 神马国产精品三级电影在线观看 | 亚洲18禁久久av| 在线国产一区二区在线| 69av精品久久久久久| 久久久久久免费高清国产稀缺| 午夜老司机福利片| avwww免费| 50天的宝宝边吃奶边哭怎么回事| 俄罗斯特黄特色一大片| 亚洲精品一卡2卡三卡4卡5卡| 真人一进一出gif抽搐免费| 国产精品影院久久| 欧美国产日韩亚洲一区| 女生性感内裤真人,穿戴方法视频| 欧美 亚洲 国产 日韩一| www.www免费av| 天天躁夜夜躁狠狠躁躁| 亚洲精品国产一区二区精华液| 亚洲精品久久成人aⅴ小说| 美女扒开内裤让男人捅视频| 美女午夜性视频免费| 亚洲一区高清亚洲精品| 久久欧美精品欧美久久欧美| 亚洲欧美日韩高清专用| 三级男女做爰猛烈吃奶摸视频| 日韩av在线大香蕉| 岛国在线观看网站| 51午夜福利影视在线观看| 搞女人的毛片| 男女之事视频高清在线观看| 国产伦一二天堂av在线观看| 在线观看美女被高潮喷水网站 | 午夜福利欧美成人| 久久久精品大字幕| 搡老妇女老女人老熟妇| 精品一区二区三区视频在线观看免费| 国产亚洲欧美98| 婷婷六月久久综合丁香| 看片在线看免费视频| а√天堂www在线а√下载| 亚洲国产高清在线一区二区三| 国产视频内射| 1024香蕉在线观看| 国产熟女xx| 18禁裸乳无遮挡免费网站照片| 亚洲精品在线观看二区| 成人18禁在线播放| 国产一区二区三区在线臀色熟女| 精品不卡国产一区二区三区| aaaaa片日本免费| 午夜免费观看网址| 亚洲国产日韩欧美精品在线观看 | 听说在线观看完整版免费高清| 三级男女做爰猛烈吃奶摸视频| 非洲黑人性xxxx精品又粗又长| 欧美黑人精品巨大| 成人精品一区二区免费| 中亚洲国语对白在线视频| 日韩精品青青久久久久久| 国产精品 欧美亚洲| 国产精品九九99| 欧美不卡视频在线免费观看 | 国产在线观看jvid| 脱女人内裤的视频| 51午夜福利影视在线观看| 日本黄大片高清| 亚洲欧美精品综合久久99| 亚洲精品美女久久久久99蜜臀| 国产伦一二天堂av在线观看| 久久午夜亚洲精品久久| 精华霜和精华液先用哪个| 免费观看精品视频网站| 久久久久久久精品吃奶| 大型黄色视频在线免费观看| 国产精品 欧美亚洲| 中文字幕熟女人妻在线| 欧美成人午夜精品| 国产单亲对白刺激| www.999成人在线观看| 日韩国内少妇激情av| 国产熟女午夜一区二区三区| 在线观看一区二区三区| 精品高清国产在线一区| 久久精品成人免费网站| 婷婷精品国产亚洲av在线| 天堂影院成人在线观看| 少妇裸体淫交视频免费看高清 | av超薄肉色丝袜交足视频| 此物有八面人人有两片| 一本精品99久久精品77| 日韩欧美免费精品| 每晚都被弄得嗷嗷叫到高潮| 美女扒开内裤让男人捅视频| 国产精品爽爽va在线观看网站| 一边摸一边做爽爽视频免费| 狂野欧美激情性xxxx| 亚洲熟妇熟女久久| 黄色成人免费大全| av片东京热男人的天堂| 亚洲精品美女久久久久99蜜臀| 岛国在线观看网站| 亚洲成人久久爱视频| 精品国产美女av久久久久小说| 在线观看舔阴道视频| tocl精华| 亚洲精品久久成人aⅴ小说| 亚洲午夜理论影院| 成人手机av| 男人舔女人下体高潮全视频| 波多野结衣巨乳人妻| 一本久久中文字幕| av国产免费在线观看| 黄色a级毛片大全视频| 色播亚洲综合网| 久久久久亚洲av毛片大全| 国产在线精品亚洲第一网站| 欧美zozozo另类| 日韩欧美国产在线观看| 丝袜美腿诱惑在线| 人妻夜夜爽99麻豆av| 欧洲精品卡2卡3卡4卡5卡区| 国产一区二区三区在线臀色熟女| 在线观看www视频免费| 亚洲欧美精品综合久久99| 九九热线精品视视频播放| 亚洲熟女毛片儿| 亚洲av电影不卡..在线观看| 一个人免费在线观看的高清视频| 999久久久精品免费观看国产| 欧美另类亚洲清纯唯美| 俄罗斯特黄特色一大片|