• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    A Lightweight Road Scene Semantic Segmentation Algorithm

    2023-12-15 03:57:20JianshengPengQingYangandYaruHou
    Computers Materials&Continua 2023年11期

    Jiansheng Peng,Qing Yang and Yaru Hou

    1College of Automation,Guangxi University of Science and Technology,Liuzhou,545000,China

    2Department of Artificial Intelligence and Manufacturing,Hechi University,Hechi,547000,China

    ABSTRACT In recent years,with the continuous deepening of smart city construction,there have been significant changes and improvements in the field of intelligent transportation.The semantic segmentation of road scenes has important practical significance in the fields of automatic driving,transportation planning,and intelligent transportation systems.However,the current mainstream lightweight semantic segmentation models in road scene segmentation face problems such as poor segmentation performance of small targets and insufficient refinement of segmentation edges.Therefore,this article proposes a lightweight semantic segmentation model based on the LiteSeg model improvement to address these issues.The model uses the lightweight backbone network MobileNet instead of the LiteSeg backbone network to reduce the network parameters and computation,and combines the Coordinate Attention (CA) mechanism to help the network capture long-distance dependencies.At the same time,by combining the dependencies of spatial information and channel information,the Spatial and Channel Network(SCNet)attention mechanism is proposed to improve the feature extraction ability of the model.Finally,a multiscale transposed attention encoding(MTAE)module was proposed to obtain features of different resolutions and perform feature fusion.In this paper,the proposed model is verified on the Cityscapes dataset.The experimental results show that the addition of SCNet and MTAE modules increases the mean Intersection over Union(mIoU)of the original LiteSeg model by 4.69%.On this basis,the backbone network is replaced with MobileNet,and the CA model is added at the same time.At the cost of increasing the minimum model parameters and computing costs,the mIoU of the original LiteSeg model is increased by 2.46%.This article also compares the proposed model with some current lightweight semantic segmentation models,and experiments show that the comprehensive performance of the proposed model is the best,especially in achieving excellent results in small object segmentation.Finally,this article will conduct generalization testing on the KITTI dataset for the proposed model,and the experimental results show that the proposed algorithm has a certain degree of generalization.

    KEYWORDS Semantic segmentation;lightweight;road scenes;multi-scale transposition attention encoding(MTAE)

    1 Introduction

    In today’s society,road scene segmentation has become a technology of great importance as urbanization and traffic demand continue to grow.Road scene segmentation aims to accurately separate and identify individual objects on the road from their surroundings in digital images or videos.This technology has a wide range of promising applications in areas such as autonomous driving,traffic monitoring,and intelligent transportation systems.The goal of road scene segmentation is to achieve a comprehensive understanding and perception of the traffic environment by accurately segmenting the vehicles,pedestrians,traffic signs,and other elements on the road.By effectively separating all objects on the road,road scene segmentation provides autonomous vehicles with the necessary environment awareness to ensure safe driving.At the same time,road scene segmentation can be used in traffic monitoring systems to monitor traffic flow in real-time,detect violations,and optimize traffic signal control,thereby improving road safety and traffic efficiency.

    Semantic segmentation is one of the important tasks in the field of computer vision,aiming to classify each pixel in an image or video into different semantic classes accurately.With the continuous development of deep learning,semantic segmentation models have also evolved.From the original full convolutional network(FCN)[1]to various improved methods later,each generation of networks has introduced new ideas and techniques in feature extraction,contextual information,and multi-scale features to continuously improve the accuracy and efficiency of semantic segmentation.Common semantic segmentation models are based on convolutional neural networks to achieve pixel-level classification.The main feature of these models is the adoption of an encoder-decoder structure and the combination of improved strategies such as skip connections and contextual coding,which improve the accuracy of image semantic segmentation.In the ILSVRC2012 vision competition [2],the AlexNet network [3] achieved state-of-the-art results (SOTA) by improved the accuracy of the model through techniques such as nonlinear activation functions,Dropout layers,data augmentation,and multi-GPU training.These ideas and techniques have since been widely used in deep learning models.In 2014,Long et al.improved the convolutional neural network and proposed FCN,which was the first successful application of convolutional neural network to semantic segmentation tasks.The FCN network achieves pixel-level classification by removing the fully connected layer and mapping the feature map output from the convolutional neural network to the output image of the same size as the input image.In 2015,Ronneberger et al.proposed the U-Net network [4],which uses an encoder-decoder structure.The encoder is used to extract the features,while the decoder is used to gradually recover the position and size of the feature image pixels.This model structure is able to fuse shallow features of images with deep features to obtain highly accurate segmentation results.Chen et al.proposed DeepLab network [5],which mainly uses dilated convolutions and multiscale pyramid pooling to improve segmentation accuracy.Niu et al.proposed the hybrid multi-attention network HMANet [6],which employs a channel attention module and introduces a region random playback attention module to reduce feature redundancy and improve the efficiency of the selfattentive mechanism by representing it in a regional manner.Chen et al.proposed DeepLabv3+network [7],which employs a series of technical improvements,including atrous spatial pyramid pooling(ASPP),encoder-decoder structure,and hybrid loss function.Zhao et al.proposed the PSPNet network[8],which uses a pyramid pooling module to capture the global contextual information of the input image and combine it with local information to obtain better semantic segmentation results.

    However,the complexity of road scenes and the limitations of current semantic segmentation models lead to a series of challenges in road scene segmentation.(1) Road scenes have complex diversity,including different road types,changes in lighting conditions,effects of weather conditions,and vehicle occlusion and overlap.All these factors will lead to a decrease in the accuracy and stability of the semantic segmentation model.(2) There are objects of different scales in the road scene,such as pedestrians,vehicles,and traffic signs.Current semantic segmentation models face difficulties in handing scale variation,and it is difficult to accurately segment objects with different scales,especially small-scale targets.(3) In autonomous driving and traffic monitoring systems,real-time segmentation results of road scenes are highly demanding.However,the current semantic segmentation models cannot be applied on embedded devices to achieve real-time segmentation,considering both the number of parameters and inference time.(4) In road scenes,the boundaries between some semantic categories may be unclear or blurred,and it is challenging for current semantic segmentation techniques to accurately capture and segment these blurred boundary regions.Therefore,how to effectively solve these challenges and improve the accuracy and robustness of road scene segmentation has become a hot spot and focus of current research.

    In this paper,we propose a new lightweight semantic segmentation model based on LiteSeg[9]to address these problems.The work we have done is as follows:

    1.The original model backbone network is replaced by the MobileOne[10]network.MobileOne is an efficient neural network that can effectively reduce the number of parameters and computation of the network.Also,to ensure the feature extraction capability of the model,a lightweight and flexible CA attention mechanism[11]is introduced to obtain more efficient feature information.

    2.A multi-scale transposition attention module is proposed.This module can acquire vector features at different resolutions and fuse the features.A Transformer encoder module is also incorporated to operate between channels with the help of covariance matrices of key and query values,which combine the accuracy of global transform networks with the scalability of convolutional structures.

    3.The SCNet module is proposed.This module processes spatial attention and channel attention in parallel to obtain richer feature information.

    2 Related Work

    As deep learning models continue to evolve,there is an increasing need to apply these models to real-world problems,thus requiring more and more models to be deployed on mobile devices.In order to interact with the real environment in real-time,semantic segmentation models need to have real-time processing capabilities and meet accuracy requirements.In recent years,the state-of-the-art lightweight semantic segmentation models can be divided into three main types:encoder-decoder structures,twobranch structures,and multi-branch structures.The model with an encoder-decoder structure extracts the features of the input image through the encoder and then maps these features back to the original image size through the decoder to achieve pixel-level classification.This structure allows for relatively fast inference while maintaining high accuracy.The model with a two-branch structure divides the network into two branches,one for the extraction of global contextual information and the other for capturing local details.This structure can effectively balance global and local information,improve segmentation accuracy,and enhance inference speed to a certain extent.The model with a multibranch structure processes feature information at different scales separately by using multiple parallel branching networks.These branching networks can extract features in different sense field ranges and fuse them to obtain more comprehensive contextual information,thus improving the accuracy of semantic segmentation.The classification of lightweight semantic segmentation models is shown in Table 1.

    Table 1: Lightweight semantic segmentation model classification

    1.Encoder-decoder architecture.In 2016,Paszke et al.proposed the ENet[12]model,which is the first semantic segmentation model that takes real-time into account,but its segmentation accuracy is low.In the same year,Eduardo et al.improved on ENet by proposing the ERFNet [13] model to obtain more information by interleaving the use of null convolution and ResNet blocks.The RGPNet[14]model consists of an asymmetric encoder-decoder and an adapter,which helps to preserve and refine the distributed representation of multiple levels and facilitates the flow of gradients between different levels.The LiteSeg model proposed by Emara et al.explores a deeper version of the ASPP module and applies short and long residual connections and deeply separable convolution to provide a faster and more efficient model.The MSCFNet [15] model uses decomposed convolution blocks and asymmetric residual blocks of dilated convolution to construct the encoder and uses deconvolution instead of the high computational cost of the FPANet [16] model extracts high-level semantic information by aggregating spatial pyramids with feature pyramids and uses a bidirectional directed feature pyramid network to fuse feature information at different levels.Besides,the LETNet[17]model combines U-shaped Convolutional Neural Networks(CNN)with Transformer,the ELANet[18] model designs an effective lightweight attention-guided network,and the ELUNet [19]model provides an efficient and lightweight U-shaped network architecture.The EACNet[20]model uses convolutional decomposition to enhance the feature representation capability and robustness to rotating objects by using depth-oriented convolutional decomposition as a basic feature layer and point-by-point convolution for fusion.The CFPNet[21]model combines the Inception module and null convolution to extract feature maps and contextual information of various sizes.MobileOne is an ultra-lightweight backbone network for mobile devices that achieves significant improvements in latency and accuracy through the introduction of linear branching.

    2.Double branch structure.In 2018,Yu et al.proposed BiSeNet [22],a bilateral segmentation network containing spatial and contextual paths,and they introduced a feature fusion module and an attention refinement module to further improve accuracy at an acceptable cost.To handle communication between parallel branches,the authors proposed BiseNetV2 [23] by adding an effective fusion layer to the BiSeNet model,which enhances the connection between the two paths.Despite the significant progress in speed and accuracy of BiseNetV2,there are still some redundancies in the initial downsampling phase and the fusion layer,which limit the information exchange between spatial and semantic branches.To address this issue,Faster BiSeNet [24] adopts a cleaner design that reduces redundant network architecture and enhances the relationship between the two branches.Aerial-BiSeNet [25] proposes a feature attention module and a channel attention-based feature fusion module based on the channel attention mechanism,effectively refining and combining features to improve the model’s performance.Additionally,Poudel et al.proposed Fast-SCNN[26],which introduces a learning downsampling module based on the existing two-branch fast segmentation method to compute low-level features of multiple resolution branches simultaneously and combine high-resolution spatial details with low-resolution depth features.This method is suitable for low-memory embedded devices with efficient computational power.

    3.Multi-branch structure.In 2018,Zhao et al.proposed ICNet[27]with multi-scale input,using few convolutions at high resolution and a deeper network at low resolution,and finally features fusion.In 2019,Li et al.proposed DFANet[28]to aggregate discriminative features through a series of subsidiary stages.DFANet is based on multi-scale feature propagation,which reduces the model parameters while maintaining a good perceptual field and enhancing the learning ability of the model.In the same year,Liu et al.proposed FDDWNet [29],which uses decomposition-expanded depth-separable convolution to learn feature representations from different scale receptive fields.The MSFNet [30] model designs a multiscale feature network consisting of an enhanced diverse attention module and an upsampling phase fusion module that uses high-level semantic information to complement low-level detail information to improve prediction.In 2021,Fan et al.proposed the Short-Term Dense Concate (STDC)network [31],which constructs the basic modules of the STDC network by reducing the dimensionality of the feature map and using feature map clustering for image representation.NDNet [32] eliminates redundant information through pruning and is suitable for realtime segmentation tasks with narrow width and large depth.To further optimize the output resolution of the segmentation network,NDNet uses point-by-point convolution to connect feature maps,facilitating the aggregation of information from two different levels.DFFNet[33]proposes a lightweight multiscale component of the semantic pyramid module,which improves the efficiency of context encoding through depth decomposition.

    3 Proposed Method

    LiteSeg is one of the current excellent lightweight semantic segmentation models.It is designed based on the encoder-decoder architecture,ASPP,dilated convolutions,and depth-wise separable convolutions.By employing depth-wise separable convolutions and ASPP,the model reduces parameter count while improving segmentation accuracy.The use of dilated convolutions expands the receptive field,enabling better capture of object information at different scales.On the Cityscapes dataset,LiteSeg achieves an impressive mIoU accuracy of 67.81%.It is a lightweight,efficient real-time semantic segmentation model.However,we noticed that the LiteSeg model does not consider the positional information of features and the semantic correlations of long-distance features,resulting in difficulties in accurately segmenting object boundaries and small objects.Therefore,in this paper,we propose a lightweight semantic segmentation model with higher accuracy based on LiteSeg while maintaining the model volume,and the structure is shown in Fig.1.Building upon LiteSeg,this paper utilizes the MobileOne backbone network module and incorporates the CA attention mechanism and SCNet attention module to extract feature information.Additionally,the multi-scale transposition attention coding module is used to extract long-range global features.

    3.1 Feature Extraction Based on CA Attention Mechanism

    The CA attention mechanism is used to enhance the receptive field of deep neural networks,primarily by weighting the feature maps of different channels in the network.Its structure diagram is shown in Fig.2.The CA attention module encodes channel relationships and long-term dependencies through precise location information,and the specific operations are divided into two parts:Coordinate information embedding and Coordinate Attention generation.

    The CA Attention module encodes channel relationships and long-term dependencies through precise location information,and the specific operations are divided into two parts: Coordinate information embedding and Coordinate Attention generation.

    Figure 1:Improved LitSeg network structure

    Figure 2:CA attention mechanism

    Coordinate information embedding:the global pooling approach is used for the global encoding of spatial information for channel attention encoding,but it makes it difficult to maintain location information because it pushes global spatial information into the channel description.To induce the attention unit to capture distant spatial interactions with accurate location information,the global pooling is decomposed according to the formula in Eq.(1)and transformed into a one-to-one feature encoding operation.

    Specifically,given the input,each channel is encoded along the horizontal or vertical coordinates using pooling kernels of size(H,1)or(1,W),respectively.This channel attention captures long-term dependencies along one spatial direction and preserves precise position information along the other spatial direction,which helps the network to locate the target of interest more accurately.

    Coordinate Attention generation:After passing through the transformations in the information embedding,this part splices the above transformations and then uses the convolutional transform function to process them.

    Eq.(2) represents the concatenation operation along the spatial dimension,encoding spatial information in both the horizontal and vertical directions.It is then decomposed into two separate tensor sums along the spatial dimension.Utilizing two additional convolutional transforms,zhandzware individually transformed into tensors with the same number of channels,as shown in Eqs.(3)and(4).

    Finally,the output of the CA module can be expressed as shown in Eq.(5).

    The global pooling layer in the network pools the input feature maps globally on average to obtain the global average of each channel.The channel weighting layer multiplies the original feature maps with the channel weights to obtain the weighted feature maps.Finally,the feature reconstruction layer reconstructs the weighted feature maps into the final feature maps.The addition of the CA module helps MobileOne extract more feature information at a very small additional computational cost.

    3.2 SCNet Attention Module

    Combined attention modules have achieved wide application in the field of image processing,and in[34],a spatial,temporal,and channel attention module was proposed to achieve the extraction of spatio-temporal features using three attention modules.Since using three attention modules in series would add extra computation,this paper aims to balance the accuracy and speed of the model and thus designs the SCNet attention module.The SCNet attention module uses the spatial and channel attention mechanisms and performs feature extraction on the input separately in parallel,and finally fuses them using one-dimensional convolution.The structure of the SCNet attention module model is shown in Fig.3.

    Figure 3:SCNet attention module

    The SCNet attention module consists of a spatial attention module(SAM)and a channel attention module(CAM).The SAM module is capable of assigning different levels of attention to each region.The equations for this module are shown in Eq.(6).

    wherefindenotes the input features andδ(·) denotes the Sigmoid activation function.The spatial attention mapMsis multiplied with the input feature map to perform adaptive refinement in terms of residuals.

    The channel attention module is used to extract channel features from the region feature map in the image frame.The formula for this module is shown in Eq.(7).

    whereδ(·)denotes the Sigmoid activation function,wheregcdenotes the multilayer perceptron(MLP),andMcdenotes the channel attention map.The channel attention mapMcis multiplied with the input feature map in a residual manner for adaptive refinement.

    After the input features pass through the two attention modules separately,the concat function is used to concatenate the two feature maps in parallel,as shown in Eq.(8),where concat denotes the concatenation operation.

    The improved SCNet attention module features the spatial and channel attention modules in parallel to extract the input features separately.The features of each module are multiplied with the input features in the form of residuals for adaptive refinement.The two features in parallel are then concatenated together and finally input to a 1Dconvolutional network for fusion.

    3.3 Multiscale Transposed Attention Encoding Module

    In this paper,different regions of the feature map are segmented using the MTAE module,which encodes the feature map to extract multiscale features.The encoder model is used to extract long-distance global features.The encoder module is a component of the Transformer model,which can reduce the strict neighborhood constraint of graph convolution and focus on the connection between pixels that are physically far away,while also being able to focus on local information.The feature encoding module can associate more features between regions,establish long-distance feature dependencies,and extract richer features.

    The traditional Transformer uses a self-attentive mechanism to interact with image blocks to model image data.However,the complexity of the self-attentive mechanism itself makes it difficult to handle high-resolution images.In this paper,the encoder module uses a transposed attention mechanism,which operates between channels of features with the help of covariance matrices of key and query values.The transposed attention mechanism combines the accuracy of traditional global transform networks with the scalability of convolutional structures with linear complexity in sequence length,thus allowing efficient processing of high-resolution images.

    The mutual covariance attention mechanism is an improvement on the self-attentive mechanism,and the former is capable of handling high-resolution images.In the self-attentive mechanism,the input vectors can form a matrixX.The query matrixQ,the key matrixK,and the value matrixVare obtained by multiplyingXwith three learnable transformation matricesWq,WkandWv,respectively.

    For each query vectorqi,calculate its dot product scores with all key vectorskjand input these scores into the softmax function to obtain a weight vectorwi,where each element represents the correlation betweenqiand the different key vectorskj,as shown in Eq.(9).

    Due to the limitations of the self-attention mechanism,this paper adopts the mutual covariance attention mechanism to interact with image features.The mutual covariance attention is a transposed form of the self-attention mechanism,which improves the attention mechanism based on the mutual covariance matrix.In the self-attention mechanism,the attention score is first calculated using the query matrixQ,the key matrixK,and the value matrixV.Then,the values are weighted and summed by the attention scores to get the output.In contrast,in the mutual covariance attention,the mutual covariance matrix between features needs to be calculated first,as shown in Eq.(10).

    The formulation of the mutual covariance attention is shown in Eq.(11),where bothQandKare generated by the coding layer,tdenotes the learnable parameter,andTdenotes the transpose operation.The XCA module is shown in Fig.4,and each mutual covariance attention is preceded by a LayerNorm layer,which serves to normalize all the data.

    The multiscale perception module based on the transposed attention mechanism is shown in Fig.5.From the figure,it can be seen that the module inserts residual connection structures with hierarchies in the residual units.The encoder structure is used as a filter in the module,while connecting different filter groups in a hierarchical residual-like manner.After chunking,the input is divided into three subsets:x1,x2,andx3.Each feature has the same scale size,but the channels are 1/3 of the input features.The encoder first extracts features from a set of feature maps,and then the output features from the previous set are sent to the next set of encoder filters and another set of input feature maps.This process is repeated several times until all the input feature maps have been processed.Finally,all the feature maps are concatenated to obtain the fusion information.Due to the combination effect,many equivalent feature scales are obtained.

    Figure 4:Reciprocal covariance attention

    Figure 5:Multi-level transposed attention encoding module

    The module partitions the feature map into 3 subsets:x1,x2,andx3,and then performs the operations using the transpose attention mechanism in the encoder,respectively.The formula is shown in Eq.(12).

    It is specified that each passing encoder is an operationAi,and the output isyi.Meanwhile,the outputyi-1ofAi-1is added to the feature subsetxi,and then input toAito complete the feature extraction.

    4 Experimental Results and Analysis

    4.1 Experimental Environment Configuration

    The software and hardware environments for the experiments in this paper are shown in Table 2.

    Table 2: Experimental environment configuration

    Hyperparameter setting:Optimizer,SGD(stochastic gradient descent);momentum,0.937;weight decay,0.0005;learning rate,0.001;epoch,150.

    4.2 Experimental Data and Evaluation Index

    In this paper,we use Cityscapes dataset as the experimental data.Cityscapes dataset is a large-scale dataset for computer vision,focusing on providing training and performance testing for autonomous driving environment perception models.It covers various street scenes,road scenes,and seasons,with a total of 5000 images.The dataset includes 2975 images in the training set,500 images in the validation set,and 1525 images in the test set.

    Before training,the Cityscapes dataset was divided into 19 classes,namely: road,sidewalk,building,wall,fence,pole,traffic light,traffic sign,vehicle,terrain,sky,person,rider,car,truck,bus,train,motorcycle,and bicycle,In the training process,the size of the images is scaled from 1024*2048 to 321*512 according to the server’s computing power and time efficiency,and the number of images loaded in each batch is 8.The experiments use the pre-training weights of the official LiteSeg model to initialize the model parameters,and the other network parameters are kept constant during the training process.

    The evaluation metric used in the experiment is mIoU,which is the average of the IoU of all categories.In semantic segmentation,the IoU value of a category is calculated as the ratio of the intersection of the set of pixels predicted by the semantic segmentation model as the set of pixels in that category to the real set of pixels in that category and the union set.The IoU is calculated as shown in Eq.(13).The mIoU is calculated as shown in Eq.(14).

    where,kis the number of categories,k+1 includes background categories,idenotes the true value,jdenotes the predicted value,pijdenotes the prediction ofitoj,piidenotes the prediction ofitoi,pjidenotes the prediction ofjtoi.

    Params refers to the number of parameters to be trained in the model,including all the weights and bias terms,and is used in this paper to measure the complexity of the model.Giga Floating Point Operations(GFLOPs)refers to the number of floating-point operations performed by the model when performing a single forward computation,in billions of operations per second,which is used in this paper to measure the computational complexity of the model.

    4.3 Analysis of Experimental Results

    4.3.1 Comparison of Different Algorithms

    To validate the effectiveness of the proposed algorithm in this paper,we compared it with LiteSeg,ICNet,ENet,ERFNet,BiSeNetV2,STDC1-Seg50,and SeaFormer[35]algorithms.The experimental results are shown in Table 3.The proposed algorithm in this paper increases the mIoU value by 2.46%,the Params value by 0.19M,and the GFLOPs value by 0.46 compared to the original LiteSeg,which proves that the improvements in this paper are effective.Compared to ICNet,the mIoU value increased by 11.57%,the Params value decreased by 4.86M,and the GFLOPs value decreased by 22.94.Compared to ENet,the mIoU value increased by 4.95%,the Params value increased by 1.40M,and the GFLOPs value decreased by 3.15.Compared to ERFNet,the mIoU value increased by 5.57%,and the Params value decreased by 0.25.Compared to BiSeNetV2,although the mIoU value decreased by 1.57%,our model size is much smaller,the Params value decreased by 39.43M,and the GFLOPs value decreased by 15.78.Compared to STDC1-Seg50,the mIoU value increased by 0.53%,the Params value increased by 6.58M,and GFLOPs value increased by 4.55.Compared to SeaFormer,although the mIoU value decreased by 0.07%,our model size is much smaller,the Params value decreased by 2.18M,and the GFLOPs value increased by 3.37.It can be seen that the proposed algorithm balances accuracy and model size and has the best overall performance among the eight algorithms.

    Table 3: Comparison of metrics of mainstream algorithm

    To further verify the validity proposed in this paper,the metrics on the 19 categories of datasets in Cityscapes were analyzed and compared,and the experimental results are shown in Table 4.The IoU values for sidewalk,building,wall,fence,pole,traffic light,terrain,car,train,motorcycle,and bicycle categories are higher than the other four models.The IoU values of the proposed algorithm in the road category are slightly lower than those of the ERFNet model.In the traffic sign and rider categories,the IoU values of the proposed algorithm are lower than those of the LiteSeg and ERFNet models.In the vehicle category,the IoU of the proposed algorithm is lower than the other four models.In the sky,bus,and track categories,the IoU value of the proposed algorithm is lower than that of the LiteSeg model.In the person category,the IoU of the proposed algorithm is lower than the LiteSeg and ENet models.The proposed model optimizes the feature extraction results by using attention mechanism and multi-scale structure and can extract higher quality feature information of low-resolution targets compared to the original model and other models.However,the resolution of targets such as vehicle,road,and traffic sign are higher,so the improvement of segmentation effect for them is relatively low.

    Table 4: IoU comparison of dataset categories

    This article also analyzed and compared the Class mIoU metric of five algorithms,with 150 iterations in the experiment.The mIoU change curve is shown in Fig.6,from which it can be seen that the proposed algorithm and LiteSeg present a stable growth trend (smooth curve fluctuation)with the increase of epoch and gradually stabilize when epoch reaches 140.ICNet,ENet and ERFNet show small fluctuations in mIoU values with the increase of epochs,and ENet has stabilized when epoch reaches 100,and ERFNet stabilizes only when epoch reaches 140,while the mIoU value of ICNet is not only the lowest but also always in oscillation during the training process.

    Figure 6:mIoU variation graph

    4.3.2 Ablation Experiments

    To further verify the effectiveness of the added modules,they are compared with the original algorithm,and the experimental results are shown in Table 5.It can be seen that adding SCNet and MTAE can increase the mIoU value of LiteSeg by 4.96%.However,the Params and GFLOPs of the algorithm will also increase by 1.61M and 7.14,respectively.In this paper,considering the model size and computation,we use MobileOne instead of the original backbone network and introduce a lightweight and flexible CA attention mechanism.As a result,we achieve a 2.46% increase in the mIoU of the algorithm without significantly increasing the Params and GFLOPs.

    Table 5: Ablation experiment

    4.3.3 Comparison of Visualization Results

    To directly validate the effectiveness of the proposed model in this paper,we compare the segmentation results of seven different algorithms on the Cityscapes dataset,as shown in Fig.7.The yellow box indicates areas where other models have incomplete segmentation compared to our model.

    Figure 7: (Continued)

    Figure 7:Visualization results of multiple models on Cityscapes validation set.From top to bottom 1-Input RGB image;2-Our model;3-LiteSeg model;4-ICNet model;5-ENet model;6-ERFNet model;7-BiSeNetV2;8-STDC1-Seg50;9-SeaFormer

    To verify the generalization ability of the model proposed in this article,we conducted experiments on the KITTI dataset.We selected three models with similar segmentation accuracy as the proposed model in this article,BiSeNetV2,STDC1-Seg50,and SeaFormer,and visually compared their generalization ability,as shown in Figs.8-10.The yellow box represents the part of the model that has not been segmented.

    Figure 8: Visualization results of KITTI dataset.From top to bottom 1-Input RGB image;2-Our model;3-BiSeNetV2

    Figure 9: Visualization results of KITTI dataset.From top to bottom 1-Input RGB image;2-Our model;3-STDC1-Seg50

    Figure 10: Visualization results of KITTI dataset.From top to bottom 1-Input RGB image;2-Our model;3-SeaFormer

    5 Conclusion

    To achieve the segmentation task of complex road scenes,this paper proposes a lightweight semantic segmentation algorithm based on LiteSeg.To reduce the size of the model,the MobileOne backbone network is used to extract features.A lightweight and efficient CA attention mechanism and SCNet module are used to enhance the feature extraction capability of the network,enabling it to focus on discriminative regions in the image and efficiently distinguish differences between different regions to achieve accurate segmentation.Furthermore,cross-dimensional feature fusion is achieved by adding the MTAE module,introducing the encoder module of Transformer in each scale space,and establishing jump connections in each dimensional space to fuse the features from different dimensional spaces.In this paper,the proposed algorithm is tested on the Cityscapes dataset,and the experimental results show that the algorithm improves the mIoU to 71.03%with only a slight increase in Params and GFLOPs compared to LiteSeg,while the IoU of all 12 categories on Cityscapes is higher than that of the LiteSeg algorithm,with only 7 categories having slightly lower IoU values.At the same time,this article will also test the generalization ability of the proposed model on the KITTI dataset,and the experimental results show that the proposed model has a certain degree of generalization ability.This demonstrates that the proposed algorithm meets the demand for accurate and fast segmentation of road images.However,the algorithm proposed in this paper also has some limitations.For example,for the compactness of the model,we abandon the downsampling process in the final stage,which makes the receptive field of the model insufficient to cover large target objects,resulting in limited improvement in segmentation accuracy for high-resolution objects.Due to the limitation of the computing power of the device,the image is cropped during the model training process,resulting in the loss of spatial details in the image,which leads to unsatisfactory segmentation of the boundary part of the image.Future work will conduct in-depth experiments on the power consumption of the model while improving segmentation accuracy.We will use knowledge distillation to further reduce the computational resources of the model and conduct experiments on different datasets captured on smart cars.

    Acknowledgement:The authors are highly thankful to the National Natural Science Foundation of China,to the Innovation Fund of Chinese Universities Industry-University-Research,to the Research Project for Young and Middle-Aged Teachers in Guangxi Universities,and to the Special Research Project of Hechi University.This research was financially supported by the Project of Outstanding Thousand Young Teachers’Training in Higher Education Institutions of Guangxi,Guangxi Colleges and Universities Key Laboratory of AI and Information Processing (Hechi University),Education Department of Guangxi Zhuang Autonomous Region.

    Funding Statement:The authors are highly thankful to the National Natural Science Foundation of China (No.62063006),the Natural Science Foundation of Guangxi Province (No.2023GXNSFAA026025),to the Innovation Fund of Chinese Universities Industry-University-Research (ID:2021RYC06005),to the Research Project for Young and Middle-Aged Teachers in Guangxi Universities(ID:2020KY15013),and to the Special Research Project of Hechi University(ID:2021GCC028).This research was financially supported by the Project of Outstanding Thousand Young Teachers’Training in Higher Education Institutions of Guangxi,Guangxi Colleges and Universities Key Laboratory of AI and Information Processing(Hechi University),Education Department of Guangxi Zhuang Autonomous Region.

    Author Contributions:The authors confirm contribution to the paper as follows: study conception and design: J.Peng;data collection: Y.Hou;analysis and interpretation of results: Q.Yang;draft manuscript preparation:Q.Yang.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data presented in this study are available upon request from the corresponding author.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    91狼人影院| 欧美3d第一页| 男女边吃奶边做爰视频| 亚洲真实伦在线观看| 肉色欧美久久久久久久蜜桃 | 成人美女网站在线观看视频| 免费观看在线日韩| 99热网站在线观看| 日本猛色少妇xxxxx猛交久久| 日本与韩国留学比较| 免费无遮挡裸体视频| 国产成人a区在线观看| 女人被狂操c到高潮| 毛片女人毛片| 欧美xxxx黑人xx丫x性爽| 淫秽高清视频在线观看| 丰满人妻一区二区三区视频av| 久久久久久国产a免费观看| 国产 亚洲一区二区三区 | 亚洲精品亚洲一区二区| 国产精品.久久久| 国产成人a区在线观看| 国产69精品久久久久777片| 97精品久久久久久久久久精品| 亚洲18禁久久av| 亚洲怡红院男人天堂| 亚洲美女视频黄频| 五月玫瑰六月丁香| 3wmmmm亚洲av在线观看| 一个人看的www免费观看视频| 国产伦精品一区二区三区四那| 中文字幕免费在线视频6| 日韩成人av中文字幕在线观看| 日韩精品青青久久久久久| 中文字幕av成人在线电影| 成人毛片a级毛片在线播放| 99热这里只有是精品50| 国产黄色视频一区二区在线观看| 毛片一级片免费看久久久久| 国产免费视频播放在线视频 | 国产高清国产精品国产三级 | 午夜福利在线观看吧| 黄片wwwwww| 免费看a级黄色片| 国产 一区精品| 国产男女超爽视频在线观看| 国精品久久久久久国模美| 激情五月婷婷亚洲| 一级毛片久久久久久久久女| 91久久精品国产一区二区三区| 可以在线观看毛片的网站| 中文字幕久久专区| 日日啪夜夜爽| 国产视频内射| 日韩 亚洲 欧美在线| 国产精品久久久久久精品电影小说 | 国内揄拍国产精品人妻在线| 国产男女超爽视频在线观看| 岛国毛片在线播放| 亚洲,欧美,日韩| 亚洲性久久影院| 十八禁网站网址无遮挡 | 欧美成人a在线观看| 免费播放大片免费观看视频在线观看| 老司机影院毛片| 成年女人看的毛片在线观看| 日韩国内少妇激情av| 久久这里有精品视频免费| 国产欧美日韩精品一区二区| 国产午夜精品论理片| 尤物成人国产欧美一区二区三区| 精品一区二区三区视频在线| 在现免费观看毛片| 亚洲精品日本国产第一区| 中国国产av一级| 久久精品久久精品一区二区三区| 国产亚洲91精品色在线| 天堂√8在线中文| 九色成人免费人妻av| 自拍偷自拍亚洲精品老妇| 亚洲电影在线观看av| 国产成人精品久久久久久| 精品国产露脸久久av麻豆 | 亚洲人与动物交配视频| 亚洲精品中文字幕在线视频 | 国产精品国产三级专区第一集| 夫妻性生交免费视频一级片| 国产av码专区亚洲av| 久久久久久久久中文| 亚洲在线自拍视频| 美女被艹到高潮喷水动态| 欧美97在线视频| 久久久久久久国产电影| 国产视频内射| 亚洲精品一区蜜桃| 高清在线视频一区二区三区| 啦啦啦韩国在线观看视频| 小蜜桃在线观看免费完整版高清| 99热这里只有精品一区| 成人鲁丝片一二三区免费| 内地一区二区视频在线| a级一级毛片免费在线观看| 免费看光身美女| 三级男女做爰猛烈吃奶摸视频| 天天一区二区日本电影三级| 亚洲,欧美,日韩| 国产熟女欧美一区二区| 夜夜看夜夜爽夜夜摸| 久久国产乱子免费精品| 亚洲真实伦在线观看| 国产真实伦视频高清在线观看| 精品不卡国产一区二区三区| 成人毛片a级毛片在线播放| 美女脱内裤让男人舔精品视频| av福利片在线观看| 国产精品一区二区性色av| 人人妻人人澡人人爽人人夜夜 | 黄色配什么色好看| 老司机影院毛片| 免费av不卡在线播放| 永久网站在线| 国产一区二区亚洲精品在线观看| a级毛片免费高清观看在线播放| 欧美另类一区| 日本猛色少妇xxxxx猛交久久| eeuss影院久久| 日韩不卡一区二区三区视频在线| 国产黄色免费在线视频| 五月伊人婷婷丁香| 亚洲欧洲国产日韩| 国产成人91sexporn| 两个人视频免费观看高清| 丰满少妇做爰视频| 国产亚洲午夜精品一区二区久久 | av在线蜜桃| 亚洲精品一区蜜桃| 国产综合懂色| 高清视频免费观看一区二区 | 人妻少妇偷人精品九色| 尤物成人国产欧美一区二区三区| 国产淫语在线视频| 成年版毛片免费区| 国产伦理片在线播放av一区| 久久久欧美国产精品| 日日摸夜夜添夜夜添av毛片| 午夜视频国产福利| 国产又色又爽无遮挡免| 亚洲,欧美,日韩| 欧美精品一区二区大全| 超碰av人人做人人爽久久| 成人亚洲精品av一区二区| 日韩一区二区视频免费看| 亚洲欧美日韩东京热| 亚洲精品亚洲一区二区| 国产免费福利视频在线观看| 亚洲欧美一区二区三区黑人 | 观看美女的网站| 成年免费大片在线观看| 国产男女超爽视频在线观看| 日本与韩国留学比较| 国产成人91sexporn| 亚洲av一区综合| 日本av手机在线免费观看| 亚洲欧美日韩卡通动漫| 秋霞在线观看毛片| 久久久久久久国产电影| 亚洲无线观看免费| 一个人看的www免费观看视频| 黄色一级大片看看| 国产成人精品婷婷| 亚洲av电影不卡..在线观看| 人妻一区二区av| 日本免费在线观看一区| 国内揄拍国产精品人妻在线| 欧美日韩在线观看h| 日韩中字成人| 亚洲国产av新网站| 国产高清国产精品国产三级 | 欧美成人午夜免费资源| 久久久久久久久久成人| 天天躁夜夜躁狠狠久久av| 亚洲一级一片aⅴ在线观看| 欧美性感艳星| 国产成人a区在线观看| 欧美97在线视频| 欧美日韩综合久久久久久| av网站免费在线观看视频 | 亚洲精品成人久久久久久| 丰满少妇做爰视频| av在线老鸭窝| 久久亚洲国产成人精品v| 亚洲av二区三区四区| 国产探花极品一区二区| 亚洲av二区三区四区| 国产乱人视频| eeuss影院久久| 色综合站精品国产| 亚洲最大成人av| 六月丁香七月| 晚上一个人看的免费电影| 欧美变态另类bdsm刘玥| 精品不卡国产一区二区三区| 免费看光身美女| 美女脱内裤让男人舔精品视频| 寂寞人妻少妇视频99o| 久久久午夜欧美精品| 一级片'在线观看视频| 欧美性感艳星| 最近中文字幕高清免费大全6| 午夜福利视频精品| 中文字幕亚洲精品专区| 一级爰片在线观看| 国产成人精品一,二区| 伦理电影大哥的女人| 日韩一本色道免费dvd| 777米奇影视久久| 观看美女的网站| 五月天丁香电影| 亚洲不卡免费看| 免费黄频网站在线观看国产| 国产男女超爽视频在线观看| 日本-黄色视频高清免费观看| 十八禁网站网址无遮挡 | 久久97久久精品| 能在线免费观看的黄片| 欧美xxxx黑人xx丫x性爽| 日韩av在线大香蕉| 国产在视频线精品| 秋霞在线观看毛片| 亚洲av成人精品一区久久| 久久久久免费精品人妻一区二区| 看免费成人av毛片| 亚洲国产av新网站| 日日干狠狠操夜夜爽| 精品少妇黑人巨大在线播放| 久久99蜜桃精品久久| 美女国产视频在线观看| 丝袜美腿在线中文| 一级黄片播放器| 亚洲av成人精品一区久久| 人体艺术视频欧美日本| 国产一区二区在线观看日韩| 国产精品国产三级国产av玫瑰| or卡值多少钱| 日韩亚洲欧美综合| 97热精品久久久久久| 亚洲国产最新在线播放| 一个人观看的视频www高清免费观看| 国模一区二区三区四区视频| 免费黄色在线免费观看| 午夜精品一区二区三区免费看| 黄色配什么色好看| 亚洲精品久久久久久婷婷小说| 亚洲怡红院男人天堂| 欧美激情在线99| 国产视频内射| 欧美丝袜亚洲另类| 国产久久久一区二区三区| 亚洲在久久综合| 亚洲真实伦在线观看| 国产精品人妻久久久影院| 校园人妻丝袜中文字幕| 熟女人妻精品中文字幕| 99视频精品全部免费 在线| 国产黄色小视频在线观看| 色尼玛亚洲综合影院| 黄色配什么色好看| 精华霜和精华液先用哪个| 亚洲内射少妇av| 亚洲va在线va天堂va国产| 亚洲不卡免费看| 波野结衣二区三区在线| 亚洲精品视频女| 国产高清国产精品国产三级 | 夫妻性生交免费视频一级片| 久久99热这里只频精品6学生| 黄色欧美视频在线观看| 久久久成人免费电影| 美女cb高潮喷水在线观看| 日本一本二区三区精品| 熟女电影av网| 中文欧美无线码| 亚州av有码| 欧美zozozo另类| 直男gayav资源| 亚洲av二区三区四区| 草草在线视频免费看| 噜噜噜噜噜久久久久久91| av女优亚洲男人天堂| 亚洲在线自拍视频| 国产精品国产三级国产av玫瑰| 欧美xxxx性猛交bbbb| 久久亚洲国产成人精品v| 国产精品无大码| 一区二区三区四区激情视频| 国产黄色视频一区二区在线观看| 日韩一区二区三区影片| 美女被艹到高潮喷水动态| av在线观看视频网站免费| 久久精品综合一区二区三区| 久久久久精品性色| 乱码一卡2卡4卡精品| a级一级毛片免费在线观看| 亚洲国产精品sss在线观看| 国产精品久久视频播放| 欧美高清性xxxxhd video| 免费大片黄手机在线观看| 成人毛片a级毛片在线播放| 国产一区亚洲一区在线观看| 最近2019中文字幕mv第一页| av专区在线播放| 精品熟女少妇av免费看| 国产精品人妻久久久影院| 日日干狠狠操夜夜爽| 国产乱来视频区| 久久久久国产网址| 精品一区二区免费观看| 久久97久久精品| 久久精品夜色国产| 午夜视频国产福利| 亚洲欧美精品专区久久| 超碰av人人做人人爽久久| 国产亚洲一区二区精品| 久久久久久久久久成人| 只有这里有精品99| 青春草亚洲视频在线观看| av专区在线播放| 亚洲熟妇中文字幕五十中出| 精品一区二区三卡| 26uuu在线亚洲综合色| 91在线精品国自产拍蜜月| 精品久久久久久久久久久久久| 国产成人精品一,二区| 日本熟妇午夜| 人妻夜夜爽99麻豆av| 18禁在线播放成人免费| 国产精品久久久久久久久免| 九九在线视频观看精品| 丰满人妻一区二区三区视频av| 嫩草影院新地址| 免费黄色在线免费观看| 内射极品少妇av片p| 草草在线视频免费看| 国内精品宾馆在线| 国产日韩欧美在线精品| 欧美最新免费一区二区三区| 精品国产三级普通话版| 久久国内精品自在自线图片| 日韩一区二区视频免费看| 国产爱豆传媒在线观看| 欧美激情在线99| 内地一区二区视频在线| 国内精品美女久久久久久| 一个人免费在线观看电影| 最近最新中文字幕大全电影3| 欧美区成人在线视频| 99久久中文字幕三级久久日本| 亚洲精品色激情综合| 麻豆成人午夜福利视频| 国产午夜精品一二区理论片| 久久久久久九九精品二区国产| 精品久久久久久久久av| 在线免费观看不下载黄p国产| 精品久久久久久成人av| 乱人视频在线观看| 国产 一区精品| av国产免费在线观看| 日韩欧美精品免费久久| 欧美97在线视频| 日本与韩国留学比较| 大片免费播放器 马上看| 亚洲无线观看免费| 亚洲熟妇中文字幕五十中出| 亚洲精品成人av观看孕妇| 国内精品一区二区在线观看| 看非洲黑人一级黄片| 91在线精品国自产拍蜜月| 别揉我奶头 嗯啊视频| 91久久精品电影网| 男女国产视频网站| 青春草国产在线视频| 久久久久国产网址| 97人妻精品一区二区三区麻豆| 国产精品精品国产色婷婷| 欧美极品一区二区三区四区| av在线蜜桃| 视频中文字幕在线观看| 久久精品国产亚洲av天美| 亚洲,欧美,日韩| 91av网一区二区| 亚洲va在线va天堂va国产| 午夜免费观看性视频| 婷婷色综合大香蕉| 99久国产av精品国产电影| 欧美高清成人免费视频www| 亚洲av成人精品一二三区| 免费观看的影片在线观看| 精品酒店卫生间| 亚洲综合色惰| 少妇人妻精品综合一区二区| 高清视频免费观看一区二区 | 国产成人免费观看mmmm| 亚洲人与动物交配视频| av专区在线播放| 国产午夜精品久久久久久一区二区三区| 久久久久久国产a免费观看| 久99久视频精品免费| 国产一区二区三区综合在线观看 | 国产伦理片在线播放av一区| 国产成人免费观看mmmm| 免费在线观看成人毛片| 成人亚洲精品一区在线观看 | 成人漫画全彩无遮挡| 91精品国产九色| 亚洲欧美一区二区三区黑人 | 亚洲欧美日韩东京热| 免费看美女性在线毛片视频| 午夜精品一区二区三区免费看| 亚洲成人精品中文字幕电影| 2021少妇久久久久久久久久久| 欧美变态另类bdsm刘玥| 亚洲美女视频黄频| 精品一区二区三区视频在线| 99re6热这里在线精品视频| 久久99热这里只有精品18| 久99久视频精品免费| 搞女人的毛片| 麻豆国产97在线/欧美| 日日啪夜夜爽| 亚洲av中文av极速乱| 午夜免费激情av| 国产在线男女| 亚洲怡红院男人天堂| 国产精品一区www在线观看| 黄色日韩在线| 亚洲国产欧美人成| 日本三级黄在线观看| 亚洲乱码一区二区免费版| 亚洲精品影视一区二区三区av| 国产精品麻豆人妻色哟哟久久 | 免费看日本二区| 91av网一区二区| 麻豆精品久久久久久蜜桃| 身体一侧抽搐| 18禁在线无遮挡免费观看视频| 中文字幕久久专区| 免费高清在线观看视频在线观看| 日韩av在线大香蕉| 午夜精品一区二区三区免费看| 哪个播放器可以免费观看大片| 午夜福利在线观看吧| 亚洲无线观看免费| 国国产精品蜜臀av免费| 又爽又黄a免费视频| 午夜福利在线在线| 国产亚洲91精品色在线| 亚洲精品第二区| 国产免费福利视频在线观看| 日本爱情动作片www.在线观看| 中文字幕久久专区| 美女国产视频在线观看| 精品99又大又爽又粗少妇毛片| 日产精品乱码卡一卡2卡三| 精品人妻熟女av久视频| 亚洲av日韩在线播放| 亚洲高清免费不卡视频| 中文字幕av在线有码专区| 日韩欧美三级三区| 精品久久久久久成人av| av在线观看视频网站免费| 国产淫片久久久久久久久| .国产精品久久| 日韩av在线大香蕉| kizo精华| 久久久久久国产a免费观看| 亚洲av成人精品一区久久| 久久精品久久久久久噜噜老黄| 国产在视频线精品| 日韩,欧美,国产一区二区三区| 我的女老师完整版在线观看| 超碰97精品在线观看| 三级国产精品欧美在线观看| 亚洲va在线va天堂va国产| 日本熟妇午夜| 日韩强制内射视频| 午夜视频国产福利| 又粗又硬又长又爽又黄的视频| 啦啦啦啦在线视频资源| 久久久久久九九精品二区国产| 日日摸夜夜添夜夜爱| 国产激情偷乱视频一区二区| www.色视频.com| 建设人人有责人人尽责人人享有的 | 亚洲国产精品成人久久小说| 国产精品人妻久久久影院| 成人亚洲精品av一区二区| 欧美精品一区二区大全| 久久国内精品自在自线图片| 精品不卡国产一区二区三区| 欧美zozozo另类| 特级一级黄色大片| 少妇的逼好多水| 成人美女网站在线观看视频| 3wmmmm亚洲av在线观看| 久久人人爽人人片av| 我的老师免费观看完整版| 日韩av在线大香蕉| 国产日韩欧美在线精品| 精品国产三级普通话版| 国产一级毛片在线| 神马国产精品三级电影在线观看| 丰满乱子伦码专区| 亚洲人成网站高清观看| 亚洲综合色惰| 中文字幕免费在线视频6| 看非洲黑人一级黄片| 欧美xxxx性猛交bbbb| 欧美三级亚洲精品| 日本免费a在线| 成人一区二区视频在线观看| 免费不卡的大黄色大毛片视频在线观看 | 青春草亚洲视频在线观看| 少妇的逼水好多| 中文精品一卡2卡3卡4更新| 极品少妇高潮喷水抽搐| 亚洲国产av新网站| 高清在线视频一区二区三区| 搡女人真爽免费视频火全软件| 一级毛片我不卡| 国产精品.久久久| 日本免费a在线| 一级毛片aaaaaa免费看小| 在线免费观看的www视频| 内地一区二区视频在线| 久久精品人妻少妇| 久久久久网色| 日本黄大片高清| 免费观看的影片在线观看| 人人妻人人澡人人爽人人夜夜 | 91精品伊人久久大香线蕉| 黄色配什么色好看| 久久97久久精品| 三级经典国产精品| 日韩av在线大香蕉| 伦精品一区二区三区| 激情 狠狠 欧美| 赤兔流量卡办理| 国国产精品蜜臀av免费| 午夜免费男女啪啪视频观看| 亚洲国产欧美在线一区| 又爽又黄a免费视频| 日韩三级伦理在线观看| videos熟女内射| av又黄又爽大尺度在线免费看| 亚洲精品国产av蜜桃| www.色视频.com| 中文资源天堂在线| 亚洲欧美清纯卡通| av免费观看日本| 国产欧美日韩精品一区二区| 精品久久久久久久末码| 人人妻人人澡欧美一区二区| 国产成人a区在线观看| 91在线精品国自产拍蜜月| 国产精品综合久久久久久久免费| 蜜桃亚洲精品一区二区三区| 国产欧美另类精品又又久久亚洲欧美| 亚洲在线自拍视频| 夫妻午夜视频| 国产淫语在线视频| 成人特级av手机在线观看| 中文资源天堂在线| 91久久精品国产一区二区三区| 日韩视频在线欧美| 日本猛色少妇xxxxx猛交久久| 十八禁国产超污无遮挡网站| 国产成年人精品一区二区| 美女被艹到高潮喷水动态| 国国产精品蜜臀av免费| 亚洲人成网站在线观看播放| a级毛片免费高清观看在线播放| 久久精品熟女亚洲av麻豆精品 | 大片免费播放器 马上看| 你懂的网址亚洲精品在线观看| 久久人人爽人人片av| 爱豆传媒免费全集在线观看| 国内精品一区二区在线观看| 日韩 亚洲 欧美在线| 亚洲av男天堂| 成年女人在线观看亚洲视频 | 激情 狠狠 欧美| 一个人看视频在线观看www免费| 欧美一区二区亚洲| 天天一区二区日本电影三级| 女人十人毛片免费观看3o分钟| av在线亚洲专区| 一夜夜www| 九九在线视频观看精品| 亚洲四区av| av网站免费在线观看视频 | 别揉我奶头 嗯啊视频| 高清日韩中文字幕在线| 亚洲在线自拍视频| 我要看日韩黄色一级片| 亚洲成色77777| 亚洲一区高清亚洲精品| 国产成人福利小说| 人体艺术视频欧美日本| 欧美 日韩 精品 国产| 狂野欧美激情性xxxx在线观看| 不卡视频在线观看欧美| 一区二区三区乱码不卡18| 天天躁夜夜躁狠狠久久av| 超碰97精品在线观看| 国产伦精品一区二区三区视频9|