• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Traffic Sign Recognition for Autonomous Vehicle Using Optimized YOLOv7 and Convolutional Block Attention Module

    2023-12-12 15:50:06KuppusamySanjayDeepashreeandIwendi
    Computers Materials&Continua 2023年10期

    P.Kuppusamy,M.Sanjay,P.V.Deepashree and C.Iwendi

    1School of Computer Science and Engineering,VIT-AP University,Andhra Pradesh,522237,India

    2School of Creative Technology,University of Bolton,Manchester,BL3 5AB,UK

    ABSTRACT The infrastructure and construction of roads are crucial for the economic and social development of a region,but traffic-related challenges like accidents and congestion persist.Artificial Intelligence(AI)and Machine Learning(ML)have been used in road infrastructure and construction,particularly with the Internet of Things(IoT)devices.Object detection in Computer Vision also plays a key role in improving road infrastructure and addressing trafficrelated problems.This study aims to use You Only Look Once version 7(YOLOv7),Convolutional Block Attention Module(CBAM),the most optimized object-detection algorithm,to detect and identify traffic signs,and analyze effective combinations of adaptive optimizers like Adaptive Moment estimation (Adam),Root Mean Squared Propagation(RMSprop)and Stochastic Gradient Descent(SGD)with the YOLOv7.Using a portion of German traffic signs for training,the study investigates the feasibility of adopting smaller datasets while maintaining high accuracy.The model proposed in this study not only improves traffic safety by detecting traffic signs but also has the potential to contribute to the rapid development of autonomous vehicle systems.The study results showed an impressive accuracy of 99.7%when using a batch size of 8 and the Adam optimizer.This high level of accuracy demonstrates the effectiveness of the proposed model for the image classification task of traffic sign recognition.

    KEYWORDS Object detection;traffic sign detection;YOLOv7;convolutional block attention module;road sign detection;Adam

    1 Introduction

    Infrastructure and construction of roads in any geographical area play a pivotal role in the economic and social development of the region,as it connects people to business and allows the movement of locomotives and services.One of the present-day primary challenges relating to road infrastructure is accidents,and other traffic-related concerns like traffic congestion,restricted infrastructure capacity,low maintenance of roads,etc.[1,2].Classically,human conception and past experiences have guided the progress of road infrastructure.However,as technology has become ubiquitous,and owing to advancements in automobile-related technologies such as self-parking systems,self-driving cars,fully autonomous systems,etc.,all of which are essentially categorized under the umbrella of Autonomous Driving Systems(ADS).There has been a significant increase in the usage of AI,and its sub-domains in accomplishing some cardinal tasks in ADS.An evaluative study on Deep Neural Networks(DNN)for Traffic Sign Detection (TSD),throws some light on how the detection of traffic signs is an indispensable study because these detection systems encompass anchor components required for safety and support in ADS [3].The IoT devices are utilized to gather the data from the environment,and ML analyses the data to solve the challenges in traffic management systems.The traffic management system contains three layers such as data acquisition,network transmission,and application.The data acquisition is done via sensors,cameras,video monitoring,and online monitoring.The collected data is transmitted over the network using Bluetooth,Wi-Fi (Wireless Fidelity),mobile network,etc.Finally,AI and ML play a major role in the analysis,visualize the analyzed outputs,and derive the systems based on the outputs like ADS [4].The study using ML showcased the latest and most advanced techniques for monitoring construction progress,including methods for collecting data,retrieving information,estimating progress,and presenting the results visually.Along similar lines,AI and ML are used in many more traffic-related issues[5].A review of traffic congestion prediction using AI described the probabilistic reasoning models like fuzzy logic,Hidden Markov Model(HMM),the Bayesian network,Support Vector Machine(SVM),Artificial Neural Networks(ANN),Decision Trees,etc.,Deep Learning(DL)algorithms like Convolutional Neural Networks(CNN),Long Short-Term Memory(LSTM),are used for short-term traffic congestion prediction[6].AI and ML-based incident detectors in Road Transport Systems(RTS)discussed dire problems and plausible solutions for reducing traffic accidents that enhanced the automatic incident system detectors[7].

    The technology development in computer vision plays a key role with goals revolving around improving road infrastructure like road accidents,traffic congestion,etc.Object detection is a subfield of computer vision that uses various DL architectures for recognizing and classifying objects.A comparative analysis of CNN-based object detection algorithms shows YOLOv3 is the fastest,and performs best overall outperforming Single-Shot Detector(SSD),and Faster Region-based CNN(RCNN) [8].However,it is also highlighted that the choice of the algorithm may be dependent upon the specific situation or problem that needs to be solved.For instance,R-CNN works best for small datasets that do not require real-time video outputs,whereas YOLO works best for object detection in the live environment.YOLOv4 runs twice as fast as EfficientNet,with an Average Precision(AP)of 14%more than YOLOv3.The YOLOv7 algorithm surpasses all the well-known real-time objectdetection algorithms concerning AP at 56.8%,and speed with a maximum range of 160 FPS[9].So far,research on the detection of traffic signs has been done using several versions of YOLO,and other object-detection algorithms.

    This study aims to use the fairly latest version of the most optimized object-detection algorithm YOLOv7 to detect and identify traffic signs.This study also tries to dive deep into analyzing effective combinations of adaptive optimizers like Adam and SGD along with YOLOv7.SGD has solid theoretical and mathematical support,along with an exhibition of enhanced stability and generality[10].In most applications,the Adam optimizer is recommended as the default optimization method because it usually generates better results,is faster to compute,and requires fewer tuning parameters than conventional optimization methods [11].Batch sizes 8 and 16 are used for the task of TSD.A portion of the German traffic signs is used for the training purpose.This study also explores the feasibility of adopting smaller datasets while keeping high accuracy to modify the application domain.Fig.1 shows the various traffic sign classes that are pointed by red arrows for human reference.

    Traffic sign recognition is a primary factor for autonomous cars to make safe travel.However,traffic sign recognition system contains more challenges due to limitations that are shown by recent incidents involving autonomous vehicles[12].Conventional traffic sign recognition encounters numerous challenges,such as occlusion,lighting conditions,and the existence of several neighboring traffic signs[13].

    Figure 1:Input images of each class for prediction

    1.1 Motivation

    The motivation for this research work is as follows:

    ? Traffic sign recognition is essential for autonomous cars to navigate safely and efficiently.But there are severe worries regarding the limitations of traffic sign recognition systems and the methods they employ,as shown by recent incidents involving autonomous vehicles and research connected to recognition system failure.Therefore,it becomes even more crucial to create powerful algorithms that can get beyond these constraints and provide precise and trustworthy traffic sign detection to improve the performance and safety of autonomous cars.Effective traffic sign identification is essential for maximizing traffic flow and raising overall road safety in addition to lowering the likelihood of accidents.

    ? Traditional techniques of traffic sign recognition,on the other hand,encounter various problems,such as occlusion,fluctuating lighting conditions,complicated backdrops,and the presence of multiple signs nearby.Due to these challenges,improved approaches must be proposed to manage these situations and provide precise and dependable traffic sign detection and identification.

    1.2 Contributions

    The contributions of this research work are as follows:

    ? This research intends to improve the accuracy and speed of traffic sign detection by incorporating the CBAM into the YOLOv7 framework.The CBAM’s potent attention mechanism enables the model to effectively acquire,and highlight key spatial and channel-wise information,enabling reliable detection of traffic signs even under difficult conditions like occlusion or complicated backdrops.

    ? Investigate and compare the effectiveness of the proposed model with different optimizers and batch sizes.

    ? The enhanced model proposed in this study exhibits improved feature representation,higher detection accuracy,and resilience by combining the characteristics of YOLOv7 and three CBAM modules in a complementary way,advancing the development of autonomous driving technology.

    ? The significance of this study is to achieve high accuracy on a small-sized real-time dataset thereby applying the model proposed in this study to a larger,and more diverse dataset for real-time applications in autonomous vehicles.

    The remainder of this study is organized as follows: Section 2 gives an overview of existing literature on traffic sign recognition for autonomous vehicles,highlighting strengths,limitations,and extensions of current knowledge.Section 3 focuses on the description of the YOLOv7 with the CBAM framework,including its working principle,architecture,and loss function.The dataset used for training and evaluation is described in Section 4.Section 5 describes the evaluation metrics,hyperparameters,and hardware/software configurations used in the experiments.Section 6 presents a detailed analysis of the results,including performance comparisons and visual representations of the model’s capabilities.Finally,the conclusion and future scope are discussed in Section 7.

    2 Literature Review

    There has been a positive trend toward applications in computer vision resulting in a substantial amount of research on TSD using various object-detection algorithms.Relevant to this study,an indepth inspection and analysis of various machine vision-based traffic detection models divided into 5 categories viz.color,shape,color and shape,ML,and Light Detection and Ranging(LiDAR)based models[14].A TSD system based on novel DL architectures used the YOLOv3,and Xception models along with Adam and RMSprop optimizers.These models are designed using the dataset with 3 classes such as “Yellow,diamond-shaped pedestrian crossing sign”,“Yellow,diamond-shaped other traffic signs”and“others”.However,this study is processed with a lower frame rate of 4.5 fps,which could be increased to improve the processing time and performance accuracy[15].

    A study specifically focused on the detection of Indian traffic signs using YOLOv3,and CNN over 5 classes,and attained an accuracy of 87%.However,the authors have not used a real-time traffic detection system to predict each frame in a video[16].A study proposed a cascaded R-CNN to obtain the multi-scale features for TSD that resulted in an accuracy of 99.7%.Additionally,the study also proposed a multi-scale attention mechanism to improve the detection of true traffic signs and reduce false detections[17].The YOLOv5 model is implemented on 8 classes of datasets viz.“No U-turn”,“Road bump”,“Road works”,“Watch for children crossing”,“Crosswalk ahead”,“Give way”,“Stop”,and “No entry”,along with a thorough comparison between the YOLOv5 and SSD.The own dataset used for the model displayed an accuracy of 97.70%.The future scope of the study is to expand the existing dataset and apply newly developed models like Mask R-CNN,CapsNet,and Siamese Neural Networks[18].An improved YOLOv5 model is implemented for real-time multi-scale TSD over a massive size of 182 classes.Data augmentation,and Adaptive Feature Fusion Pyramid Network(AF-FPN)methods were implemented to increase the performance of the standard YOLOv5 model,which indeed increased the accuracy from 60.18%to 62.67%.The performance of the model is low due to the blurring of images captured by the high-speed motion of a vehicle[19].An indigenous CNN architecture is used for TSD with the dataset having 16 classes viz.“green light”,“speed limit”,“no parking”,“bicycle and pedestrians only”,“crossroad 1”,“red light”,“crosswalk 1”,“straight ahead or left turn permitted”,“crossroads 2”,“traffic division”,“no overtaking”,“no turns”,“stop”,“one-way street”and “yellow light”.This approach outperforms YOLOv2,and Fast R-CNN,with an average accuracy of 90% in all types of weather conditions.However,authors have developed a model with less training data that could be increased to improve performance in more environments[20].A combination of Faster R-CNN,and Extreme Learning Machines(ELM)is used over 3 classes.However,the accuracy and performance of the model are not discussed quantitatively,but qualitatively it is stated that combining CNN with ELM increases the accuracy[21].

    A study on TSD and classification in the wild constructed a benchmark dataset “Tsinghua—Tencent 100K” covering real-world conditions.The study trained two models CNN and Fast RCNN which resulted in an accuracy of 88%,and 50%,respectively.The study had been implemented with a minimum number of traffic sign classes that rarely appear in benchmark datasets [22].Another study presented the YOLOv3 model in detecting temporary traffic control detection for road construction projects.The mentioned study used a dataset containing 8 classes viz.“construction cones”,“l(fā)ooper cones”,“construction barrels”,“construction barricades”,“end construction signs”,“road construction ahead signs”,“right lane reduction signs”,“right lane closed ahead signs”.The training resulted in a mean Average Precision(mAP)of 90.82%.The proposed model in the mentioned study recognized more than 98% of the temporary traffic signs correctly and approximately 81% of temporary traffic control devices correctly [23].A design for real-time TSD was implemented with CNN on 50,000 traffic-sign images and reached an accuracy of 97.3%.This model is designed by considering more traffic sign classes,and possible weather conditions affecting the visibility of the signs[24].The“WAF-LeNet”(an upgraded version of LeNet)is developed to recognize and identify traffic signs for autonomous vehicles.The accuracy attained in the study was 96.4%among 43 classes[25].Though there is a fairly small amount of research studies revolving around TSD using YOLOv7,research work was carried out to collect,and label the road damage data using Google Street View.The YOLOv7 model is trained with the collected data and results in an F1 score of 81.7%[26].A study focused on improving the performance of YOLOv5 for the detection of traffic signs in bad weather conditions made use of the Global Context (GC) block,which combined with YOLOv5’s results in an accuracy of 79.2% [27].A study quantitatively demonstrates that the combination of YOLOv7 with a lightweight convolution-based Spatial Pyramidal Pooling Fusion (SPPF) module leads to a significant improvement in model accuracy.The study reports a precise increase of 6.7%in accuracy when incorporating the SPPF module into the YOLOv7 framework[28].A portable image-based ADS system was developed using the YOLOv5 algorithm and Tesla P100 Graphics Processing Unit(GPU)system.It achieved a remarkable speed of 43.59 frames per second[29].Multiple studies have utilized a pre-trained model for TSD on large datasets,and have fine-tuned the respective models by using various optimizers [30–32].Some studies have implemented multi-task learning to simultaneously detect objects like pedestrians and bicycles [33,34].The use of LiDAR and Radar sensors has come up as one of the ways to increase the accuracy of models for TSD in challenging conditions like low lighting[35].A unique method is described in a study for analyzing Global Positioning System(GPS)trajectory data to detect vehicle turns,which involves converting the data to image-based data,postconversion,a personalized CNN model is designed[36,37].

    Previous approaches to TSD have used models like YOLOv5[27],YOLOv7[28],and CNN[36],which are popular and efficient models.However,these models do not have an attention mechanism,which can limit their performance.The model proposed in this study uses YOLOv7 with CBAM,which is an attention mechanism that helps to improve the model’s performance.Specifically,CBAM helps to focus the model’s attention on the most important features in an image,which can lead to better object detection,especially in cases where the objects in an image are small or have low contrast.

    3 Traffic Sign Detection Using YOLOv7 with CBAM

    YOLOv7 is the latest and state-of-the-art object detection model in the family of YOLO singleshot object detection models.YOLOv7 is currently the fastest and best-performing object detection model.YOLOv7 significantly enhances real-time object detection accuracy while lowering inference costs.By cutting around 40% of the parameters and 50% of the processing speed,YOLOv7 effectively beats other well-known object detectors with faster inference speeds,and higher recognition accuracy[38].

    3.1 Working Principle

    The four components that the YOLO algorithm uses to operate are residual blocks,bounding box regression,Intersection Over Union(IOU),and Non-Maximum Suppression(NMS).The initial component of the residual block divides the original image(A)into N equal-sized grid cells,where N is a hyperparameter.Localizing and determining the object’s class using the probability/confidence value is the responsibility of each grid cell.Bounding box regression is the second element that identifies the bounding boxes that correspond to rectangles highlighting all the objects in the image.There can be as many bounding boxes as there are objects within a given image.YOLO uses a single regression module to compute the characteristics of these bounding boxes.Y is the final vector of each bounding box as given in Eq.(1).

    where,Pcis the grid’s probability score for the cell that contains the object.The bounding box’s center’s x and y coordinates in relation to the surrounding grid cell are represented by bx,by.The height,and the width of the bounding box are represented by bh,bw,respectively.The four classes namely prohibitory,dangerous,mandatory,and others are represented by C1,C2,C3,and C4,respectively.Despite not all of them being significant,a single object in an image might frequently have many grid box possibilities for prediction.Such grid boxes are to be discarded in order to retain the relevant grid boxes using the third component IOU.IOU always ranges from 0 and 1.The IOU selection threshold is initially set at 0.5.Fig.2 shows the intersection area divided by the union area which is then calculated for each grid cell by YOLO.Finally,it considers grid cells with an IOU>threshold rather than those predicted to have an IOU ≤threshold.

    Figure 2:Intersection over union

    The final part NMS algorithm is a post-processing technique to remove duplicate and overlapping detections of the same object.When an object is detected,the YOLO algorithm generates multiple bounding boxes with confidence scores indicating the likelihood of an object being present in each box.However,some of these boxes may overlap or contain the same object,resulting in multiple detections for the same object.To address this issue,NMS is used to suppress all but the most confident detection of each object.The algorithm works by first sorting the detected bounding boxes by their confidence scores.Then for each box,it compares its overlap with all other boxes.If the overlap exceeds a certain threshold,the box with the lower confidence score is suppressed.The process is repeated until all boxes have been considered.The generated output helps to improve the overall performance and accuracy of the object detection algorithm.Establishing an IOU threshold is not always adequate since an item may contain several overlapping boxes.Noise might be included if many boxes are overlapped based on an IOU that exceeds the threshold and all those boxes are left unclosed.NMS can be used in these circumstances to keep only the boxes with the highest likelihood of being identified.Hence,the algorithm is designed by initializing the confidence threshold,and IOU threshold values.Then the bounding boxes are organized according to decreasing confidence.If any bounding box contains a confidence threshold 0 that is eliminated.The rest of the bounding boxes are iterated through in a loop beginning with the greatest confidence,and the IOU of the current box with every remaining box that belongs to the same class is calculated.If the IOU of the 2 boxes>IOU_Threshold,then the box with lower confidence is removed from the list of boxes.This operation is repeated until all the boxes are processed in the list.Here is an outline of the code for YOLO,a popular object detection algorithm.Table 1 shows the pseudocode of the steps involved in implementing YOLO.

    3.2 Architecture

    YOLOv7 can be used in many applications other than object detection,like instance segmentation,pose estimation,etc.In comparison to YOLOv4,YOLOv7 utilizes 36% less processing,reduces the number of parameters by 75%,and generates 1.5%higher AP.When compared to the edge-optimized version,YOLOv4-tiny,and YOLOv7-tiny reduce the number of parameters by 39%and computation by 49% while keeping the same AP.Hence,it can be stated that YOLOv7 is more optimized.A YOLO architecture is made up of various components,including a head,neck,and backbone.For the inference speed,the effectiveness of the YOLO network’s backbone is essential.The full YOLOv7 architecture can be seen in Fig.3.

    Figure 3:Proposed architecture of YOLOv7 with CBAM

    The Extended Efficient Layer Aggregation Network (E-ELAN) helps the model learn better while preserving its original gradient path.To increase the speed and accuracy of the model,EELAN considers several variables,including memory cost,input-output channel ratio,element-wise operation,activations,gradient routes,etc.[39].The CSPDarknet53 serves as the backbone network for the YOLOv7 architecture that makes up the Efficient Layer Aggregation Network(ELAN)model.CSPDarknet53 was created to increase the precision and effectiveness of object detection models.On the other hand,E-ELAN is another YOLOv7 architecture that uses EfficientNet as the backbone network.A series of CNNs called EfficientNet is created to attain cutting-edge accuracy while keeping the model’s computing cost to a minimum.The main difference between these two models is the backbone network,i.e.,ELAN uses CSPDarknet53,and E-ELAN uses EfficientNet.EfficientNet is more computationally efficient,but it may sacrifice some accuracy compared to CSPDarknet53.YOLOv7 uses an optimized compound model scaling approach that modifies the characteristics to produce suitable models for various application requirements.For instance,model scaling can improve the resolution of the model,the size of the input image,the depth,or the number of stages,and the width,or the number of channels.The compound scaling technique can keep the model’s original design characteristics.

    After training,one way to improve the model is by re-parameterizing it.The inference process takes longer,but the outcomes are more substantial.The two forms of ensemble re-parameterizations used to complete models are model level and module level.Model level re-parameterization can be done in two ways.In the first method,distinct sets of data are used to train several models with the same architecture,and then average their weights to get the final model.The second method is to take the average of a model’s weight at different epochs.But recently,module-level re-parameterization has been used in a lot of research works.The YOLOv7 contains several heads,including the Lead Head,which is accountable for all the output,and the Auxiliary Head,which helps with training middle layers.To enhance deep network training,a Label Assigner method was created that assigns soft labels after considering ground truth and network prediction results.Reliable soft labels employ optimization techniques to raise the standard and distribution of prediction output in addition to the accuracy of the prediction.However,conventional label assignment generates hard labels based on predetermined norms by directly referencing reality.The YOLOv7 architecture shown above uses kernel sizes such as 3×3,and 1×1 in all its convolution layers with padding of 1 and 2.

    Fig.4 shows a crucial component Cross-Branch Scalability (CBS).It is designed with a convolution layer,Batch Normalization(BN)layer,and a Sigmoid Linear Unit(SILU)activation function to extract images at various scales.Based on the CBS module,which makes up the upper and lower divisions,the MP1 module adds the max-pooling layer.Using max-pooling and the CBS module,the upper division reduces the image’s length and width in half.The lower division uses the first CBS module to reduce the image channel in half,the second CBS layer reduces the image’s width and length in half,and finally,the Concatenation(CAT)operation is used to combine the features retrieved from the top and lower branches,enhancing the network’s ability for feature extraction.The upsampling and CBS modules make up the UP module.

    Figure 4:CBS,MP1,UP modules

    In Fig.5,the ELAN module is made up of numerous CBS modules that have been piled on top of one another while maintaining the same input and output feature sizes.The learning capacity of the network is increased without deviating from the initial gradient path by directing the computing units of various feature groups to learn more diverse features.The Spatial Pyramid Pooling Concat Spatial Convolutional(SPPCSPC)module shown in Fig.6,ELAN-H(Extreme Low-latency Architecture for Network Heads) module,and UP module makes up the majority of the Path Aggregation Feature Pyramid Network(PAFPN)structure that makes up the Neck component of YOLOv7.The bottomup approach makes it simple to move bottom-level data up to the top level,allowing for the effective fusion of various hierarchical aspects.The CBS module,CAT module,and max-pooling module make up the majority of the SPPCSPC module.SPPCSPC uses different pooling kernel sizes such as 5 × 5,9 × 9,and 13 × 13.These modules obtain various perception fields through max-pooling.To predict confidence,category,and anchor frame,Head uses Re-parameterization Visual Geometry Group Block (RepVGG) structure to adjust the number of image channels for the output of Neck at three distinct scales and then passes through 1×1 convolution.The model proposed in this study addresses the scale problem in TSD by utilizing the SPPCSPC module in the last layer of the proposed model.Spatial Pyramid Pooling(SPP)allows capturing features at different scales without reducing the input resolution,while Cross Stage Partial (CSP) connections reduce the number of parameters in the proposed model.By incorporating these modules,the model can effectively handle the large variations in object scales commonly encountered in TSD tasks,improving the accuracy of predictions and enhancing the overall performance of the proposed model.

    Figure 5:ELAN module

    Figure 6:SPPCSPC module

    CBAM is a module used to enhance the performance of CNNs by incorporating spatial and channel attention mechanisms.It focuses on capturing both local and global context information from input feature maps,allowing the network to prioritize relevant image regions while suppressing irrelevant ones.The module consists of two components such as spatial attention and channel attention.The spatial attention module captures spatial dependencies among different channels by modeling interdependencies between spatial locations.This enables the network to focus on relevant regions and suppress background regions.The channel attention module captures interdependencies among channels by assessing the importance of each channel in conveying discriminative information.It emphasizes informative channels while suppressing less informative ones.The spatial and channel attention maps are combined to generate an attention map that captures both spatial and channel-wise information.This attention map is used to weigh the feature maps,allowing the network to selectively attend to relevant features.The YOLOv7 model is trained using the sum of the squared error between the predicted bounding boxes and the actual boxes,along with the cross-entropy loss for the class predictions.Its combination of a lightweight backbone network,effective neck,and multi-scale head make it a powerful tool for a variety of computer vision applications.The technical contribution of this study lies in the integration of three CBAM units before the three outputs of YOLOv7,a model that already detects objects at three different scales.By incorporating the CBAM module,weights are assigned to channel and spatial features of the feature map,which effectively increases the importance of useful features while suppressing irrelevant ones.This attention mechanism enables the proposed model to focus on target regions containing important information that improves accuracy in detecting objects of various sizes.

    3.3 Loss Function

    The loss function used in YOLOv7 is a mixture of different components,including:

    3.3.1 Localization Loss(LL)

    This component of the loss measures the difference between the predicted bounding box coordinates and the actual bounding box coordinates.It uses the Mean Squared Error(MSE)loss function to calculate the loss.

    3.3.2 Confidence Loss

    This component of the loss measures how confident the model is in its predictions.It calculates the difference between the predicted confidence score and the actual confidence score.The confidence score indicates whether the bounding box contains an object or not.The binary cross-entropy loss function is used to calculate this confidence.

    3.3.3 Classification Loss

    This component of the loss measures the difference between the predicted class probabilities and the actual class probabilities.The cross-entropy loss function is used to calculate this classification loss.

    3.3.4 Total Loss

    The overall loss function is a weighted sum of these three components.The weights are hyperparameters that are tuned during training to balance the contributions of the different components.The loss function’s ultimate goal is to reduce the difference between predicted,and ground truth bounding boxes,confidence scores,and class probabilities.

    4 Dataset Description

    The images in the dataset are part of the famous German traffic sign dataset and were preprocessed to ensure consistency in size,resolution,and color.The dataset consists of 741 images of traffic signs that are divided into three subsets like training set of 592 images(79.8920%),a validation set of 99 images (13.3606%),and a test set of 50 images (6.7476%) in a stratified method,ensuring that each subset had a proportional representation of each class.Table 2 shows the dataset with four classes such as prohibitory,dangerous,mandatory,and others with a total of 1,213 appearances of traffic signs.The prohibitory class(class 0)has 731 appearances of traffic signs accounting for 45.89%of the dataset.This class includes traffic signs that prohibit certain actions such as no trucks,speed limit,no traffic both ways,and no overtaking.The dangerous class(class 1)has 268 appearances of traffic signs accounting for 18.04%of the dataset.This class includes traffic signs that warn drivers of potential hazards or dangers,such as construction,priority at next intersection,bend left,bend right,bend,uneven road,slippery road,the road narrows,traffic signal,pedestrian crossing,school crossing,dangerous,cycles crossing,animals and snow.The mandatory class(class 2)has 211 appearances of traffic signs accounting for 13.44%of the dataset.This class includes traffic signs that indicate actions that drivers must take,such as a roundabout,go straight,go right,go left,go left or straight,go right or straight,keep right,and keep left.The other class (class 3) has 345 appearances of traffic signs accounting for 22.63% of the dataset.This class includes traffic signs that do not fall into the prohibitory,dangerous,or mandatory categories such as no entry,stop,give way,priority road,and restriction ends.

    Table 2:Description of dataset

    5 Experimental Setup

    The objective of this study is to achieve high accuracy while keeping the model size and computational complexity low,making it suitable for deployment on embedded systems.The dataset contains 741 images of traffic signs with varying lighting conditions,occlusions,and backgrounds.The pre-processed version of the dataset is used,where images were cropped and resized to 416×416×3 pixels and annotated properly.The dataset was divided into training(79.89%),validation(13.36%),and test (6.747%) sets.The latest version YOLOv7 object detection model is used.The model has three components that predict the class,location,and confidence of the traffic sign detected.It was first trained using the SGD optimization algorithm with batch sizes 8 and 16.It had the following values of hyperparameter with a learning rate of 0.001,weight decay of 0.0005,and momentum of 0.937.The model was trained for 100 epochs,and the total training time was 1.868,and 1.845 h for batch sizes 8,and 16,respectively.Then the model was trained using the Adam optimization algorithm with batch sizes 8,and 16.It had the following values of hyperparameter with a learning rate of 0.001,weight decay of 0.0005,and momentum of 0.937.The model was trained for 100 epochs,and the total training time was 1.916,and 1.862 h for batch sizes 8,and 16 respectively.At last,the model was trained using the AdamW optimization algorithm with batch sizes 8,and 16.It had the following values of hyperparameter with a learning rate of 0.001,weight decay of 0.0001,and momentum of 0.937.The model was trained for 100 epochs,and the total training time was 1.942,and 1.857 h for batch sizes 8,and 16,respectively.

    The evaluation metric mAP is used to measure the accuracy of the model in detecting traffic signs of different sizes,and aspect ratios.Precision,recall,and F1 score are used as secondary evaluation metrics [50,95].The training was conducted on a single Tesla K80 GPU which is available for free version of Google Colab.The calculation of mAP involves calculating the AP for each class of the detected object and then averaging those AP values across all classes.AP is the area under the precisionvs.recall curve.

    The precision-recall curve shows how the precision and recall of the algorithm vary with the detection threshold.The precision is the fraction of detected objects that are correct.The recall is the fraction of true positive predictions among all the real positive cases.Precision and recall are calculated as follows:

    The F1 score is the harmonic mean of precision and recall,calculated as follows:

    A good model should have a high F1 score,high recall,high accuracy,and high precision.Precision and recall,however,typically trade off against one another.To determine the ideal balance between precision and recall,use the F1 score.The precision-recall curve shows how accuracy and recall are traded off for different thresholds.While a low false negative rate is related to great recall,a low false positive rate is related to superior accuracy.A large area under the precision-recall curve indicates that the precision and recall are high.

    6 Results and Discussion

    The approach used in this study is evaluated using the standard mAP metric which measures the accuracy of object detection by computing the AP over all possible levels of recall.Table 3 shows the YOLOv7 model’s performance under different training configurations using four optimization algorithms such as Adam,SGD,AdamW,and RMSProp with batch sizes of 8,and 16.The outcomes demonstrated that both during training and testing,the YOLOv7 model was able to obtain high precision and recall values.This proved that the model has a low rate of FP and FN when identifying objects in images.The mAP metric measures the model’s ability to detect objects at different IOU thresholds.In particular,mAP50 evaluates the AP at an IOU threshold of.5 or 50 percent,while mAP[50,95]measures the AP across all IOU thresholds from.5 to.95,with increments of.05 or 5 percent.The YOLOv7 model achieves high mAP values for both mAP50 and mAP[50,95]indicating that the model can accurately detect objects at different IOU thresholds.

    Table 3:Results of all the combinations of YOLOv7 for TSD

    The results show that Adam performs consistently better than SGD in terms of precision,recall,and mAP50,regardless of batch size.On the other hand,AdamW is similar to Adam in most cases but shows slightly lower performance in terms of mAP50 for batch size 8.In Table 3,it can be observed that the batch size significantly impacts the performance of the YOLOv7 model.For instance,SGD shows lower precision,recall,and mAP values compared to the other configurations for batch size 16.This suggests that SGD is less effective in handling larger batch sizes,possibly due to its inherent instability in noisy,and high-dimensional optimization spaces.

    Fig.7 shows the predicted images for each class including dangerous,other,prohibitory,and mandatory signs,and each image is accompanied by the corresponding confidence level.Fig.7a inferred that the model accurately predicted dangerous class signs with high confidence levels,indicating its ability to detect,and classify potentially hazardous situations on the road.Fig.7b shows that the model was able to predict other class signs with high confidence levels,indicating its ability to accurately classify signs that do not fit into well-defined categories.Fig.7c tells that the model accurately predicted prohibitory class signs with high confidence levels,indicating its ability to detect and classify signs that restrict certain behaviors on the road.Similarly,Fig.7d indicates that the model was able to predict mandatory class signs with high confidence levels,indicating its ability to accurately detect and classify signs that require specific actions from drivers.

    Figure 7:Predicted image for each class

    Figs.8a and 9a illustrate that the YOLOv7 model performs well as it has high recall at low confidence,and as the confidence increases the recall values decrease,and finally confidence becomes 1 when recall becomes zero.Figs.8b and 9b illustrate that the model is performing well as it has high precision at high confidence,and as the confidence threshold decreases,the precision also decreases.Figs.8c and 9c prove that the curve near the upper right corner indicates the increase in recall,the reduction in precision is not immediately apparent,and the overall performance of the model is better.Figs.8d and 9d show that the model has a high F1 score at high confidence,and as the confidence threshold decreases,the F1 score also decreases.It proves that the model becomes less conservative and makes more predictions.Both the models Adam with batch sizes 8,and 16 are performing well in the dataset.However,the model with batch size 8 is performing slightly better than the one with 16.It can be inferred that the precision and the mAP50 of Adam with batch size 8 is more than the Adam with batch size 16.

    Fig.10a illustrates that the model performs well as it has high recall at low confidence,and as the confidence increases the recall values decrease,and finally confidence becomes 1 at recall is zero.Fig.10b shows that the model is performing well but the performance is less than the Adam models as the class-wise prediction is deviating from the overall prediction.By comparing Figs.10c,8c and 9c exhibited that the Adam optimizer performs better than the SGD optimizer.Fig.10d shows that the model has a high F1 score at high confidence,although not as high as the models with Adam,and as the confidence decreases,the F1 score also decreases.It showed that the model becomes less conservative.

    Figure 8:Adam optimizer with batch-size 8

    Figure 9:Adam optimizer with batch-size 16

    Figure 10:SGD optimizer with batch-size 8

    Fig.11 conveys that the model is performing very poorly compared to the previous Adam models,and SGD with batch size 8.For instance,the precisionvs.recall curve in Fig.11c depicts much below at the top corner indicating the model has not learned properly and is struggling to predict the test images and in Fig.11d the model has a very low F1 score at high confidence,and as the confidence decreases,the F1 score also decreases.

    Figure 11:SGD optimizer with batch-size 16

    Figs.12 and 13 indicate that the models AdamW with batch-size 8 and 16 perform well but not as well as models with Adam.Although these models perform better than the SGD optimizer for this particular dataset.The AdamW with batch size 8 performs better than the AdamW with batch size 16.This results that the model with the less batch size performs better for this dataset.

    Figure 12:AdamW optimizer with batch-size 8

    Figure 13:AdamW optimizer with batch-size 16

    Figs.14 and 15 illustrate that the models RMSProp with batch-size 8 and 16 perform well but not as well as models with other optimizers.In comparison between the different batch sizes,RMSProp with batch size 16 performs better than the RMSProp with batch size 8.This results that the model with the more batch size performs better for this dataset.However,its performance is lesser than other all optimizers.

    Figure 14:RMSProp optimizer with batch-size 8

    Figure 15:RMSProp optimizer with batch-size 16

    The results proved that Adam is a more effective optimization algorithm for training the YOLOv7 model and that the performance of the model varies with batch size.This information can be useful for selecting the best configuration for their specific use case.Table 4 shows the comparison of the proposed model with existing models.

    Table 4:Comparison of results

    Overall,the results illustrate the effectiveness of the YOLOv7 model in object detection tasks and provide valuable insights into its performance under different training configurations.This study concluded that Adam with batch size 8 is the most effective of all the combinations above for the TSD use case.

    Considering the proposed model’s effectiveness,it is important to acknowledge its potential limitations in extreme weather conditions and low-lighting scenarios.Since the model is trained on images captured under specific conditions,its performance may be compromised when exposed to diverse and challenging environments.To enhance the model’s generalizability and improve its performance,a more diverse range of input images can be included that encompass varied conditions during the training process.This will enable the model to learn and adapt to different environmental factors,ultimately enhancing its robustness and reliability in real-world applications.

    7 Conclusion and Future Scope

    The proposed study is implemented with four optimization algorithms,namely Adam,SGD,AdamW,and RMSProp with different batch sizes for TSD using YOLOv7.The integration of CBAM improved the model’s performance by focusing on the spatial and channel regions in the input.The evaluation has been performed on both the training,and testing datasets by considering four different metrics namely Precision,recall,mAP50,and mAP[50,95].The experimental results have shown that the Adam optimizer with batch-size 8 and 16 achieves the highest accuracy in terms of all four metrics for both training and testing datasets.Specifically,the precision and recall rates of the Adam optimizer are very high.It proved that the model can correctly identify traffic signs in the input images with high accuracy.Moreover,the results indicate that the choice of batch size can also have a substantial effect on the accuracy of the model.In general,the smaller batch size can lead to better performance,but it also increases the training time.Hence,the trade-off between accuracy,and training time should be considered while selecting the batch size.

    The results obtained in this research provide a strong basis for future work in the field of traffic sign detection.There are several directions in which this work can be extended.One prominent direction for future work is optimizing the performance of YOLOv7 specifically for real-time traffic sign detection on embedded systems like Raspberry Pi and NVIDIA Jetson.By enhancing the efficiency and speed of the model,its applicability in resource-constrained environments can be significantly expanded,enabling practical implementation in various settings.Another crucial aspect that holds great potential is the interpretability of deep learning models.In recent years,there has been a growing interest in understanding and visualizing the decision-making processes of such models.Hence,future research can delve into exploring the interpretability of YOLOv7 for traffic sign detection.This can involve visualizing saliency maps or activation patterns of the model to gain insights into how it makes predictions.These directions hold immense potential for advancing traffic sign detection systems,paving the way for improved road safety,efficient traffic management,and safer autonomous vehicles in the future.

    Acknowledgement:The authors would like to thank the editors and reviewers.

    Funding Statement:The authors received no specific funding for this study.

    Author Contributions:conceptualization,validation,review and editing: Kuppusamy Pothanaicker;supervision: Kuppusamy Pothanaicker and Celestine Iwendi;methodology,formal analysis,result analysis,writing:Sanjay Mythili and Deepashree Pradeep Vaideeswar.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:Data used in this study is available from Traffic Signs Dataset in YOLO format,Version 1.Retrieved December 26,2022 from https://www.kaggle.com/datasets/valentynsichkar/traffic-signs-dataset-in-yolo-format.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    国产成人91sexporn| a 毛片基地| 美女国产高潮福利片在线看| 久久久国产欧美日韩av| 国产精品一区二区三区四区免费观看| 国产一区二区在线观看日韩| 国产高清国产精品国产三级| 中文乱码字字幕精品一区二区三区| 九九爱精品视频在线观看| 色5月婷婷丁香| 欧美精品人与动牲交sv欧美| 国产国拍精品亚洲av在线观看| 毛片一级片免费看久久久久| 纯流量卡能插随身wifi吗| 免费看不卡的av| 欧美xxⅹ黑人| 一本色道久久久久久精品综合| 亚洲,一卡二卡三卡| 国产精品一区www在线观看| 极品少妇高潮喷水抽搐| 国产精品一区二区在线观看99| 天天躁夜夜躁狠狠久久av| 五月玫瑰六月丁香| 久久国内精品自在自线图片| 国产日韩欧美在线精品| 免费观看在线日韩| 黑人高潮一二区| 久久99热这里只频精品6学生| 少妇熟女欧美另类| 在线播放无遮挡| 男女啪啪激烈高潮av片| 久久ye,这里只有精品| 亚洲国产精品成人久久小说| 精品人妻在线不人妻| 久久人妻熟女aⅴ| 亚洲av成人精品一二三区| 观看美女的网站| 久久精品国产亚洲网站| 久久久欧美国产精品| 亚洲av欧美aⅴ国产| 国产成人精品婷婷| 熟女人妻精品中文字幕| 久久毛片免费看一区二区三区| 91久久精品电影网| 大片电影免费在线观看免费| 老司机影院成人| 日本欧美视频一区| 伊人亚洲综合成人网| 亚洲内射少妇av| 五月天丁香电影| 又黄又爽又刺激的免费视频.| 免费高清在线观看日韩| 一本—道久久a久久精品蜜桃钙片| 精品国产国语对白av| 亚洲不卡免费看| 黄色一级大片看看| 极品人妻少妇av视频| 国产成人a∨麻豆精品| 成年av动漫网址| 久久鲁丝午夜福利片| 成人手机av| 成人18禁高潮啪啪吃奶动态图 | 精品人妻熟女毛片av久久网站| 超色免费av| 中文字幕最新亚洲高清| 亚洲成人一二三区av| 亚洲性久久影院| 日产精品乱码卡一卡2卡三| 精品一区在线观看国产| 热re99久久精品国产66热6| 亚洲国产精品999| 国产精品无大码| 午夜免费鲁丝| 老熟女久久久| 寂寞人妻少妇视频99o| 97超视频在线观看视频| a 毛片基地| 日韩一区二区视频免费看| 成人毛片a级毛片在线播放| 中文精品一卡2卡3卡4更新| 婷婷色综合www| 国产日韩欧美亚洲二区| 成人毛片60女人毛片免费| 又粗又硬又长又爽又黄的视频| 一本大道久久a久久精品| 日日啪夜夜爽| 交换朋友夫妻互换小说| 成人漫画全彩无遮挡| 狂野欧美激情性bbbbbb| 久久久久久久久久成人| 国产极品粉嫩免费观看在线 | 国产av一区二区精品久久| 一级,二级,三级黄色视频| 丰满乱子伦码专区| 久热久热在线精品观看| 99精国产麻豆久久婷婷| 色哟哟·www| 久久精品夜色国产| 亚洲性久久影院| 美女主播在线视频| 欧美亚洲 丝袜 人妻 在线| 亚洲国产精品一区三区| 国产69精品久久久久777片| 免费少妇av软件| 欧美三级亚洲精品| 亚洲av福利一区| a 毛片基地| 亚洲精华国产精华液的使用体验| 国产 一区精品| 中文字幕av电影在线播放| 国产深夜福利视频在线观看| 人妻夜夜爽99麻豆av| 一级a做视频免费观看| 免费黄色在线免费观看| 另类亚洲欧美激情| 中文欧美无线码| 亚洲精品久久午夜乱码| 91成人精品电影| 亚洲欧美一区二区三区黑人 | 国产欧美日韩综合在线一区二区| 免费日韩欧美在线观看| 色5月婷婷丁香| av福利片在线| 一区二区三区乱码不卡18| 国产精品麻豆人妻色哟哟久久| 男女免费视频国产| 丰满迷人的少妇在线观看| 久热久热在线精品观看| 99视频精品全部免费 在线| 黑人高潮一二区| 最近的中文字幕免费完整| 亚洲精品一二三| 91久久精品电影网| 91在线精品国自产拍蜜月| 蜜桃久久精品国产亚洲av| 欧美日韩综合久久久久久| 少妇人妻 视频| 一级毛片我不卡| 中文字幕久久专区| 中文欧美无线码| 男男h啪啪无遮挡| 丁香六月天网| 日产精品乱码卡一卡2卡三| 国产高清有码在线观看视频| 99久久人妻综合| 午夜福利在线观看免费完整高清在| 一边亲一边摸免费视频| 如何舔出高潮| 寂寞人妻少妇视频99o| 国产亚洲最大av| 999精品在线视频| 亚洲av二区三区四区| 日日撸夜夜添| 99热全是精品| 国产av码专区亚洲av| 高清av免费在线| 99热这里只有是精品在线观看| 国产欧美亚洲国产| 亚洲综合色网址| 久久精品国产亚洲网站| 大陆偷拍与自拍| 2018国产大陆天天弄谢| 在线亚洲精品国产二区图片欧美 | 久久精品国产亚洲网站| 国产精品久久久久久久电影| 美女xxoo啪啪120秒动态图| 中文字幕制服av| 肉色欧美久久久久久久蜜桃| 天堂8中文在线网| 久久精品国产亚洲av天美| 免费高清在线观看日韩| 18禁在线播放成人免费| 久久国产精品男人的天堂亚洲 | 观看美女的网站| 极品少妇高潮喷水抽搐| 欧美日韩在线观看h| 欧美日韩国产mv在线观看视频| 国产淫语在线视频| 久久久久精品久久久久真实原创| 日韩 亚洲 欧美在线| 不卡视频在线观看欧美| 国产精品欧美亚洲77777| 亚洲精品456在线播放app| 欧美最新免费一区二区三区| 国产精品一区二区在线不卡| 国产精品麻豆人妻色哟哟久久| 欧美亚洲 丝袜 人妻 在线| 大片免费播放器 马上看| 精品一区二区三卡| 日韩精品有码人妻一区| 久久午夜福利片| 亚洲人成网站在线观看播放| 看非洲黑人一级黄片| 婷婷色综合www| 久久av网站| 成人国语在线视频| 嘟嘟电影网在线观看| 国产成人精品在线电影| 18禁在线播放成人免费| 乱码一卡2卡4卡精品| 在线看a的网站| 欧美性感艳星| 大又大粗又爽又黄少妇毛片口| 制服诱惑二区| 亚洲欧美一区二区三区国产| 久久精品人人爽人人爽视色| 成人18禁高潮啪啪吃奶动态图 | 欧美人与性动交α欧美精品济南到 | 亚洲成色77777| 男女边吃奶边做爰视频| 高清毛片免费看| 美女大奶头黄色视频| 卡戴珊不雅视频在线播放| 日韩精品免费视频一区二区三区 | 亚洲精品乱久久久久久| 香蕉精品网在线| 三级国产精品欧美在线观看| 91在线精品国自产拍蜜月| 国产高清三级在线| 国产精品秋霞免费鲁丝片| a 毛片基地| 成人免费观看视频高清| 午夜久久久在线观看| 国产欧美日韩一区二区三区在线 | 中文字幕制服av| 男的添女的下面高潮视频| 久久 成人 亚洲| 日韩一区二区视频免费看| 精品国产一区二区久久| 51国产日韩欧美| 久久精品人人爽人人爽视色| 日韩成人av中文字幕在线观看| 欧美一级a爱片免费观看看| 亚洲精品日韩av片在线观看| 国产黄色视频一区二区在线观看| 在线天堂最新版资源| 青春草亚洲视频在线观看| 国产日韩欧美亚洲二区| 26uuu在线亚洲综合色| 丝袜脚勾引网站| 成人免费观看视频高清| 久久人人爽av亚洲精品天堂| 另类亚洲欧美激情| 99久久中文字幕三级久久日本| 老女人水多毛片| 国产一区二区三区av在线| 一本大道久久a久久精品| 不卡视频在线观看欧美| 亚洲国产精品一区三区| 亚洲第一av免费看| 精品久久久久久电影网| 中文字幕亚洲精品专区| 超色免费av| 亚洲av免费高清在线观看| 黄色一级大片看看| 黄色怎么调成土黄色| 青春草视频在线免费观看| 看非洲黑人一级黄片| 亚洲五月色婷婷综合| 最近中文字幕2019免费版| 在线观看美女被高潮喷水网站| 亚洲第一av免费看| 日韩大片免费观看网站| 日本黄色片子视频| 女人久久www免费人成看片| 国产国语露脸激情在线看| 免费观看无遮挡的男女| 一级爰片在线观看| 国产精品一国产av| a级毛片黄视频| 欧美最新免费一区二区三区| 卡戴珊不雅视频在线播放| 国产精品不卡视频一区二区| 一级片'在线观看视频| 久久久久久久国产电影| 观看av在线不卡| 水蜜桃什么品种好| 永久网站在线| 最后的刺客免费高清国语| 精品少妇黑人巨大在线播放| 制服人妻中文乱码| 在线观看免费视频网站a站| 日本vs欧美在线观看视频| 妹子高潮喷水视频| 91国产中文字幕| 熟女人妻精品中文字幕| 精品熟女少妇av免费看| 欧美日韩国产mv在线观看视频| 精品人妻熟女毛片av久久网站| 男女国产视频网站| 亚洲丝袜综合中文字幕| 在线观看三级黄色| 久久久国产欧美日韩av| 十分钟在线观看高清视频www| 国产欧美日韩一区二区三区在线 | 亚洲av在线观看美女高潮| 亚洲色图 男人天堂 中文字幕 | 亚洲国产成人一精品久久久| 免费观看a级毛片全部| 亚洲人与动物交配视频| 久久免费观看电影| 亚洲欧美中文字幕日韩二区| 寂寞人妻少妇视频99o| 精品国产国语对白av| 看免费成人av毛片| 一区二区三区四区激情视频| 亚洲国产日韩一区二区| 啦啦啦中文免费视频观看日本| 国产午夜精品一二区理论片| 午夜久久久在线观看| 欧美精品高潮呻吟av久久| 九九久久精品国产亚洲av麻豆| 飞空精品影院首页| 欧美日韩视频精品一区| 丰满饥渴人妻一区二区三| 男的添女的下面高潮视频| 中文字幕精品免费在线观看视频 | 亚洲国产精品专区欧美| 久久久久久久久久久丰满| 亚洲熟女精品中文字幕| 飞空精品影院首页| 精品久久久久久久久亚洲| a级毛色黄片| 国产永久视频网站| a 毛片基地| 高清黄色对白视频在线免费看| videossex国产| 亚洲中文av在线| 国产综合精华液| 赤兔流量卡办理| 日本猛色少妇xxxxx猛交久久| 老女人水多毛片| 色婷婷av一区二区三区视频| 亚洲av在线观看美女高潮| 国产欧美日韩综合在线一区二区| 熟妇人妻不卡中文字幕| 国产国语露脸激情在线看| 国产精品蜜桃在线观看| 哪个播放器可以免费观看大片| 韩国av在线不卡| 男女啪啪激烈高潮av片| 99热全是精品| 国产午夜精品久久久久久一区二区三区| 亚洲精品成人av观看孕妇| 久久久久久久久久久久大奶| 欧美日韩av久久| 两个人的视频大全免费| 国产精品99久久99久久久不卡 | 欧美3d第一页| av免费观看日本| 999精品在线视频| 欧美精品一区二区免费开放| 亚洲精品456在线播放app| 亚洲美女黄色视频免费看| 三级国产精品片| 久久久午夜欧美精品| 在线免费观看不下载黄p国产| 纵有疾风起免费观看全集完整版| 亚洲国产精品成人久久小说| 91午夜精品亚洲一区二区三区| 在线 av 中文字幕| 亚洲精品aⅴ在线观看| 在现免费观看毛片| 久久热精品热| 亚洲精品第二区| 亚洲av国产av综合av卡| 人妻系列 视频| 午夜日本视频在线| 麻豆精品久久久久久蜜桃| 亚洲一区二区三区欧美精品| 麻豆乱淫一区二区| 午夜日本视频在线| 热re99久久精品国产66热6| 成人手机av| 成人二区视频| 少妇被粗大猛烈的视频| 九九爱精品视频在线观看| 亚洲美女视频黄频| 国产亚洲精品第一综合不卡 | 国产精品熟女久久久久浪| 成年美女黄网站色视频大全免费 | 国产精品国产三级专区第一集| 亚洲精品成人av观看孕妇| 汤姆久久久久久久影院中文字幕| 免费黄色在线免费观看| 91aial.com中文字幕在线观看| 国产69精品久久久久777片| 欧美精品一区二区免费开放| 久久久久久久国产电影| 日日撸夜夜添| 美女中出高潮动态图| 亚洲综合色网址| 成年女人在线观看亚洲视频| 国产精品一区二区在线不卡| 不卡视频在线观看欧美| 好男人视频免费观看在线| 最近中文字幕2019免费版| 日韩免费高清中文字幕av| 插逼视频在线观看| 日韩亚洲欧美综合| 母亲3免费完整高清在线观看 | 成人18禁高潮啪啪吃奶动态图 | 久久久精品94久久精品| 亚洲欧美成人精品一区二区| 免费观看性生交大片5| 成年人免费黄色播放视频| 一区二区三区四区激情视频| 欧美日韩综合久久久久久| 美女cb高潮喷水在线观看| 永久网站在线| 亚洲五月色婷婷综合| 日本与韩国留学比较| 欧美丝袜亚洲另类| 精品国产乱码久久久久久小说| 亚洲精品乱码久久久v下载方式| 99热网站在线观看| 久久精品熟女亚洲av麻豆精品| 夫妻午夜视频| 亚洲无线观看免费| 在线观看免费高清a一片| 久久久久久久久久人人人人人人| 日韩亚洲欧美综合| 国产极品天堂在线| 亚洲国产欧美日韩在线播放| 日韩人妻高清精品专区| 天堂俺去俺来也www色官网| 欧美精品一区二区免费开放| 亚洲国产成人一精品久久久| 十分钟在线观看高清视频www| 日韩视频在线欧美| 久久久久人妻精品一区果冻| 看免费成人av毛片| 亚洲精品国产av成人精品| 插逼视频在线观看| 亚洲欧美一区二区三区国产| 成人黄色视频免费在线看| 中文字幕制服av| 久久久久久久久大av| 国产一区有黄有色的免费视频| 免费久久久久久久精品成人欧美视频 | 高清毛片免费看| 99久久中文字幕三级久久日本| 大又大粗又爽又黄少妇毛片口| 伦理电影免费视频| 黄色一级大片看看| 日韩免费高清中文字幕av| 天天操日日干夜夜撸| 亚洲国产色片| 国产黄频视频在线观看| 九九爱精品视频在线观看| 免费不卡的大黄色大毛片视频在线观看| 成年女人在线观看亚洲视频| 欧美日韩视频高清一区二区三区二| 免费大片黄手机在线观看| 一级毛片我不卡| 纵有疾风起免费观看全集完整版| 欧美丝袜亚洲另类| 美女国产视频在线观看| 精品人妻在线不人妻| 18禁动态无遮挡网站| 日韩伦理黄色片| 亚洲少妇的诱惑av| 国产男女内射视频| 亚洲性久久影院| 嫩草影院入口| 天天操日日干夜夜撸| 99热这里只有是精品在线观看| 国产欧美日韩综合在线一区二区| 天天影视国产精品| 黑人欧美特级aaaaaa片| 国产高清不卡午夜福利| 亚洲国产精品成人久久小说| 男人操女人黄网站| 97在线人人人人妻| 免费久久久久久久精品成人欧美视频 | 一级爰片在线观看| 中国国产av一级| 性高湖久久久久久久久免费观看| 老司机亚洲免费影院| 亚洲精品乱码久久久久久按摩| 超碰97精品在线观看| 国产成人91sexporn| 日韩一区二区视频免费看| 校园人妻丝袜中文字幕| 精品国产一区二区三区久久久樱花| 下体分泌物呈黄色| 男人操女人黄网站| 少妇精品久久久久久久| 亚洲欧洲精品一区二区精品久久久 | 午夜av观看不卡| 极品少妇高潮喷水抽搐| 亚洲精品国产色婷婷电影| 大香蕉久久网| 一本一本综合久久| 欧美日韩视频精品一区| tube8黄色片| a级毛片免费高清观看在线播放| 99国产精品免费福利视频| 美女内射精品一级片tv| 中文天堂在线官网| 18禁动态无遮挡网站| 五月玫瑰六月丁香| 久久久久久久久久成人| 中文字幕精品免费在线观看视频 | 免费大片18禁| 一级毛片黄色毛片免费观看视频| 国产乱来视频区| 亚洲国产av影院在线观看| 午夜免费观看性视频| 人人妻人人添人人爽欧美一区卜| 插逼视频在线观看| 国产视频内射| 99久久中文字幕三级久久日本| 热99久久久久精品小说推荐| 丝袜美足系列| 亚洲一区二区三区欧美精品| 丝瓜视频免费看黄片| 亚洲欧洲日产国产| 国产白丝娇喘喷水9色精品| 亚洲国产精品成人久久小说| 欧美bdsm另类| 日本欧美视频一区| 久久99热这里只频精品6学生| 亚洲第一区二区三区不卡| 日本黄色片子视频| 97超视频在线观看视频| 欧美激情国产日韩精品一区| 欧美 亚洲 国产 日韩一| 2022亚洲国产成人精品| av线在线观看网站| 夫妻午夜视频| 色94色欧美一区二区| 亚洲人与动物交配视频| 亚洲国产av影院在线观看| 考比视频在线观看| 亚洲成人手机| 亚洲精品自拍成人| 精品人妻一区二区三区麻豆| 亚洲国产精品一区三区| 日本免费在线观看一区| 国产亚洲一区二区精品| 日韩三级伦理在线观看| 免费看光身美女| 在现免费观看毛片| 美女主播在线视频| 久久婷婷青草| av卡一久久| 亚洲精品日韩在线中文字幕| 亚洲图色成人| 制服人妻中文乱码| 欧美日韩一区二区视频在线观看视频在线| 麻豆精品久久久久久蜜桃| 日韩大片免费观看网站| 亚洲久久久国产精品| 一本一本综合久久| 久久精品国产自在天天线| 一区二区日韩欧美中文字幕 | 国产精品99久久99久久久不卡 | 午夜视频国产福利| 国模一区二区三区四区视频| 国产黄频视频在线观看| 欧美一级a爱片免费观看看| 亚洲色图综合在线观看| 久久热精品热| 伊人久久国产一区二区| 黑人欧美特级aaaaaa片| 亚洲av日韩在线播放| 18禁观看日本| 免费av中文字幕在线| 国产免费现黄频在线看| 青春草国产在线视频| 国产精品人妻久久久影院| 一个人看视频在线观看www免费| 大香蕉97超碰在线| 少妇丰满av| 国产精品国产av在线观看| 插阴视频在线观看视频| 一区在线观看完整版| 成人综合一区亚洲| 91久久精品国产一区二区成人| 美女大奶头黄色视频| 国产乱人偷精品视频| 国产综合精华液| 永久免费av网站大全| 精品99又大又爽又粗少妇毛片| 国国产精品蜜臀av免费| 国产高清不卡午夜福利| 国产无遮挡羞羞视频在线观看| 五月玫瑰六月丁香| 看十八女毛片水多多多| 青春草国产在线视频| 中文精品一卡2卡3卡4更新| 女性生殖器流出的白浆| a级毛色黄片| 国产男人的电影天堂91| 色94色欧美一区二区| 韩国高清视频一区二区三区| 午夜激情久久久久久久| 成人亚洲精品一区在线观看| 性色av一级| 日本猛色少妇xxxxx猛交久久| 久久久久久久久久人人人人人人| 亚洲av综合色区一区| √禁漫天堂资源中文www| 熟女电影av网| 亚洲国产成人一精品久久久| 成人影院久久| 在线看a的网站| 久久久久久久久久久丰满| 色视频在线一区二区三区| 永久免费av网站大全| 伊人久久国产一区二区| 亚洲精品久久午夜乱码| 日韩av在线免费看完整版不卡| 久久久久久久亚洲中文字幕|