• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Rail Surface Defect Detection Based on Improved UPerNet and Connected Component Analysis

    2023-12-12 15:50:54YongzhiMinJiafengLiandYaxingLi
    Computers Materials&Continua 2023年10期

    Yongzhi Min,Jiafeng Li and Yaxing Li

    1School of Automation and Electrical Engineering,Lanzhou Jiaotong University,Lanzhou,730070,China

    2Gansu Provincial Engineering Research Center for Artificial Intelligence and Graphics&Image,Lanzhou Jiaotong University,Lanzhou,730070,China

    3State Grid Gansu Electric Power,Baiyin Power Supply Company,Baiyin,730900,China

    ABSTRACT To guarantee the safety of railway operations,the swiftdetection of rail surface defects becomes imperative.Traditional methods of manual inspection and conventional nondestructive testing prove inefficient,especially when scaling to extensive railway networks.Moreover,the unpredictable and intricate nature of defect edge shapes further complicates detection efforts.Addressing these challenges,this paper introduces an enhanced Unified Perceptual Parsing for Scene Understanding Network(UPerNet)tailored for rail surface defect detection.Notably,the Swin Transformer Tiny version (Swin-T) network,underpinned by the Transformer architecture,is employed for adept feature extraction.This approach capitalizes on the global information present in the image and sidesteps the issue of inductive preference.The model’s efficiency is further amplified by the windowbased self-attention,which minimizes the model’s parameter count.We implement the cross-GPU synchronized batch normalization(SyncBN)for gradient optimization and integrate the Lovász-hinge loss function to leverage pixel dependency relationships.Experimental evaluations underscore the efficacy of our improved UPerNet,with results demonstrating Pixel Accuracy(PA)scores of 91.39%and 93.35%,Intersection over Union(IoU)values of 83.69%and 87.58%,Dice Coefficients of 91.12%and 93.38%,and Precision metrics of 90.85%and 93.41%across two distinct datasets.An increment in detection accuracy was discernible.For further practical applicability,we deploy semantic segmentation of rail surface defects,leveraging connected component processing techniques to distinguish varied defects within the same frame.By computing the actual defect length and area,our deep learning methodology presents results that offer intuitive insights for railway maintenance professionals.

    KEYWORDS Rail surface defects;connected component analysis;Transformer;UPerNet

    1 Introduction

    In the intricate wheel-rail system,the rail serves dual pivotal roles:it bears the vehicular vertical load and facilitates guidance.The rail head,consequent to this,encounters longitudinal and transverse forces,engendering vibrations throughout the rail system.Such perturbations can lead to detrimental surface defects on the rail.Notably,these imperfections intensify the rail’s vertical vibrations,transmitting these oscillations to the locomotive.This amplification can heighten the potential for resonance between the vehicle and its assorted elastic components.Concurrently,an increase in the rail’s torque can be observed,undermining the wheel’s directional function and jeopardizing the vehicle’s lateral stability.Hence,prompt detection of these rail surface defects becomes imperative[1].

    The vast expanse of the railway network,coupled with a multifarious natural environment and dense departures,renders manual inspection an inefficacious solution.However,the advancement in machine vision coupled with the burgeoning field of deep learning has catapulted non-contact,imagebased defect detection methods into a focal research area.Traditional visual techniques predominantly utilize texture and rudimentary grayscale attributes to discern and pinpoint target zones in unlabeled rail samples.Classic methodologies employ a gradient histogram in tandem with support vector machines,realizing rail surface defect detection.This can be further amalgamated with multiple detectors within a multi-task learning paradigm to enhance detection prowess[2].A notable study by Li et al.[3]has accentuated defect contrasts via local normalization and defect proportion limitation filtering.Their approach employed a grayscale projection contour algorithm,displaying robustness to noise,with the added advantage of swift detection rates.Techniques encompassing the internal point hollowing algorithm,coupled with the chain code tracking algorithm,have demonstrated potential in capturing defect contour data,thereby enabling automated detection of rail surface defect regions [4].In another significant research [5],differentiation of the rail image post-inverse Perona-Malik(P-M)diffusion from the prototypical image was undertaken to minimize environmental interferences.Following this,the defect region was segmented through an edge feature-based filtering algorithm.To further refine the segmentation threshold of rail defects,a novel Otsu method integrating target variance weighting was introduced in [6].Some researchers have applied Laplace low-pass filtering to rail imagery,subsequently differentiating it from the original image.This was followed by morphological feature extraction for defect detection.On average,the accuracy of this methodology was found to be 85.3% [7].Yaman et al.[8] adopted a distinctive approach by extracting feature types from the variance signal of rail imagery,subsequently ascertaining the fault type via fuzzy logic.In light of the insensitivity of defect detection to the rail’s transverse luminosity alterations,a defect detection framework was formulated.This system,underpinned by morphological operators,employed elongated,slender structural elements aligned parallel to the rail’s longitudinal axis for image processing,effectively curtailing false positives[9].Confronting the challenges inherent in rail detection,such as intricate conditions and unequal rail luminescence,a region extraction algorithm hinged on vertical projection and grayscale differentiation was conceptualized.Furthermore,an enhanced rapid robust Gaussian mixture model anchored on the Markov random field was designed,aiming for swift and precise surface defect segmentation[10].Academic discourse has also illuminated a maximum entropy threshold segmentation technique.This method,rooted in the principle where the background entropy remains invariant in low gray domains,while the target entropy fluctuates in accordance with the gray scale,has shown promise in attaining optimal thresholds for demarcating rail surface defects[11].It is crucial to note,however,that while these methods have advanced defect object detection considerably,they demand scene-specific modifications.Thus,their adaptability and real-time performance warrant further enhancement.

    Deep learning methodologies employ convolutional neural networks(CNNs)to represent defects in rail images as abstract digital features for enhanced detection.The prevailing research predominantly capitalizes on object detection and semantic segmentation networks,epitomized by their rapid processing capabilities and obviation of manual feature setting.Certain approaches harness CNNs directly to categorize rail surface defects for detection intentions.For instance,a study delineated in [12] contrived a CNN comprised of 2 convolution layers,2 pooling layers,and a fully connected layer,expressly for rail surface defect detection;a regularization technique was incorporated during the training phase to augment the recognition rate.Faghih-Roohi et al.[13],in their exploration,classified rail defect images into six distinct categories and embarked on training three disparate structures of deep CNNs via the small batch gradient descent approach.Their findings underscored the superior classification performance of more profound network structures,with heightened accuracy when the rectified linear unit(ReLU)was integrated into training.Turning to detection methodologies based on semantic segmentation networks,a study referenced in[14]conceived a 59-layer semantic segmentation network rooted in the SegNet architecture to pinpoint rail surface anomalies.When juxtaposed against the Otsu segmentation and manual threshold segmentation techniques,this semantic segmentation network triumphed with the preeminent detection rate.Further,a study in[15]introduced a lightweight rail surface defect detection network that synergized MobileNetV2 with the You Only Look Once version3(YOLOv3)mechanism.By supplanting Darknet53 with MobileNetV2 for backbone extraction,and concurrently designing a multi-scale predictor inspired by target regression,an inference speed culminating at 87.40%mean average precision(mAP)and 60 frames per second(FPS)was realized.Jin et al.[16] amalgamated semantic segmentation with object detection techniques within their detection framework,integrating Dropout with Xception—a backbone network of DeepLabV3+.To counteract the pixel imbalance between the foreground and background,YOLO was employed to pinpoint defects and deduce foreground and background weighting coefficients,thereby furnishing attention loss for the segmentation model.Dilated convolution emerges as a prevalent strategy to amplify the receptive field.A paper cited in[17]devised a lightweight semantic segmentation model rooted in a cascading self-encoding network for defect detection in rails and crafted a compact CNN specifically for defect classification.This methodology flourished both in detection precision and temporal efficiency.Luo et al.’s works [18,19] incorporated dilated convolution into the Faster Regions with CNN Features(Faster R-CNN)and established parallel convolution channels,each with unique expansion coefficients catering to distinct defect types.This innovation bolstered the detection precision of nuanced defects.Subsequent investigations leveraged the Cascade Regions with CNN Features (Cascade R-CNN) for rail surface defect detection.Addressing challenges like the skewed distribution of sample IoU and misalignment between regions of interest and feature maps engendered by rounding quantization,and the inaccurate prognostication of regression boundaries,various tools such as IoU balanced sampling,region of interest aligns(RoIAlign),and complete intersection over union(CIoU)loss were assimilated to hone the average precision(AP)to a staggering 98.75%.While unsupervised methodologies have also been ventured into for rail surface defect detection,their prevalence remains limited.A study in[20],for example,employed scale normalized cross fast flow for probability distribution estimation,which facilitated the derivation of an image’s standard Gaussian distribution.Anomaly scores were subsequently discerned based on the proximity to the distribution’s epicenter,serving the dual purpose of detection and defect localization.

    Currently,the rail surface defect detection employing image methodologies confronts several challenges: (a) The discernible information present within rail images is somewhat circumscribed,leading to diminished efficiency of methods grounded on texture and grayscale features.(b) The proactive maintenance interventions by the concerned departments hinder the effective accumulation of a voluminous dataset of rail defect images,thereby compromising the efficacy of supervised deep learning.(c)Although a plethora of rail inspection vehicles are competent in detecting fastener defects,the image data they procure for rail damage differs significantly from conventional images,making them unsuitable for direct utilization.(d) Rail surface anomalies predominantly manifest in two distinct forms:ripple and discrete defects.The emergence of discrete defects is characterized by their unpredictability,arbitrariness,and multi-scale dimensions,devoid of any recurring patterns,which intensifies the challenges associated with defect detection [21].(e) The intricate edge configurations of rail surface defects pose considerable challenges to semantic segmentation algorithms.(f) The detection outcomes yielded by deep learning approaches are arduous to correlate directly with the specific parameters delineating the defects.

    In light of these challenges,an enhanced UPerNet semantic segmentation network was introduced specifically for rail defect detection.Concurrently,a Swin-T network integrating the Transformer architecture was deployed for the extraction of defect features,which capitalized on the holistic information embedded within the images while circumventing the issue of inductive bias.The rule window self-attention inherent in Swin-T mitigates the model’s parameter load.Additionally,the cross-GPU synchronized batch normalization technique was harnessed for gradient refinement,with Lovász-hinge being employed as the loss function,facilitating the leveraging of inter-pixel dependency relationships.Conclusively,grounded on the semantic segmentation imagery of rail surface anomalies,the distinct defect regions were identified via connected component analysis,enabling the computation of the genuine length and area of the defects.

    The contributions of this research can be dichotomized as:

    (a)The proposition of a rail surface defect detection algorithm founded on semantic segmentation.Within this framework,the Swin-T,encompassing the Transformer architecture,is incorporated,amplifying the capability of the segmentation network in the extraction of rail surface defect features,and bolstering the defect segmentation quality.Notwithstanding the pervasive application of Transformers across a myriad of visual tasks—attributed to their prowess in assimilating global context information—their deployment for this specific endeavor remains an exception.

    (b)Pivoting from the semantic segmentation representations of rail surface anomalies,connected component analysis is integrated,which adeptly bridges the detection outcomes of the deep learning modus operandi with the intricate parameters defining the defects.Such an approach aids railway maintenance personnel in intuitively grasping the defect metrics,enabling them to orchestrate targeted remediation measures.

    2 Related Work

    2.1 Transformer

    The Transformer architecture,a manifestation of the encoder-decoder network paradigm,is predominantly anchored on the attention mechanism[22].Dispensing with recursion and convolution,it rectifies the inherent sluggishness that plagues the training of Recurrent Neural Networks(RNN)and has demonstrated its proficiency in machine translation endeavors.Models grounded in the Transformer architecture,such as Bidirectional Encoder Representation from Transformers (BERT) [23]and Generative Pre-train Transformer(GPT)[24],have been recognized for their stellar performance in the realm of natural language processing.For myriad computer vision tasks,the prototypical Transformer encoder model is employed,serving as a novel feature extraction apparatus.The Vision Transformer (ViT) [25],designed for image classification,segments images into standardized image blocks.These embedded blocks are subsequently input into the Transformer encoder,and post Multilayer Perceptron(MLP)classification processing,it garners commendable results,rivaling toptier convolutional networks.The DEtection TRansformer(DETR)[26]amalgamates both CNN and Transformer architectures to present an end-to-end object detection framework,reframing object detection as a images-to-sets prediction conundrum.By leveraging the self-attention mechanism of the Transformer,it simulates all pairwise interactions amidst elements in a sequence,rendering it more conducive for the intricate constraints intrinsic to set prediction.Chen et al.[27] devised the Image Processing Transformer (IPT),a model sculpted on the Transformer architecture,adept at addressing image processing challenges,excelling in tasks like image super-resolution,restoration,and denoising across a plethora of marred images from ImageNet.As per Reference [28],a semisupervised video segmentation blueprint was proffered,which incorporated the Transformer at its bottleneck layer to further distill contextual data,thereby enhancing the precision of endoscopic video segmentation in scenarios characterized by scarce annotated data.Despite the widespread adoption and commendation of the Transformer structure across a variety of visual tasks,attributed to its prowess in gleaning global contextual features,its deployment in the domain of rail surface defect detection remains limited.

    2.2 Attention Mechanism

    The genesis of the attention mechanism can be traced back to studies delving into human visual cognition.Given the constraints imposed by information processing bottlenecks,humans innately zero in on certain segments of the information spectrum,while disregarding the remainder.Emulating this intrinsic human trait,the attention mechanism in deep learning serves as an intuitive resource distribution conduit,reallocating weights contingent on the perceived significance of entities.

    The Transformer intricately weaves multiple self-attention layers,culminating in multi-head selfattention(MSA),with an aim to enhance network efficiency,as depicted in Fig.1.The self-attention layer transmutes the input vector into three distinct entities:the query vector(q),the key vector(k),and the value vector(v),subsequently coalescing them into three matrices:Q,K,andV.Attention weights are ascertained through a liaison betweenQandK,and subsequently,the comprehensive weight and output are procured via their interaction withV.Vectors deemed of higher significance are accorded heightened attention in subsequent layers.The self-attention dynamic across disparate input vectors is delineated in formula(1),wheredkrepresents the dimension pertinent toK.

    Figure 1:The principle of multi-head self-attention mechanism

    The concept of multi-head self-attention gives rise to the simultaneous execution of multiple selfattention operations on features within the network.These individual operations are subsequently amalgamated to produce the final output,enabling the network to encapsulate a richer spectrum of feature information.This process is mathematically articulated in formula(2).

    where,headiis the result of each self-attention action,as shown in formula(3),WQi,WKiandWViare all projection weight matrices.

    3 Proposed Methodology

    The UPerNet[29],a semantic segmentation network,is predicated upon both the Pyramid Pooling Module(PPM)and the Feature Pyramid Networks(FPN).Drawing upon four distinct feature layers as elucidated by ResNet,it meticulously weaves together multi-scale information derived from various stages,ensuring a minimized loss of boundary information.This is a marked improvement over the Pyramid Scene Parsing Network(PSPNet)[30],another network subscribing to the FPN framework.Yet,the UPerNet’s reliance on CNN for feature extraction manifested an inherent inductive bias,hindering the holistic utilization of global information.Moreover,when training is undertaken on a solitary GPU with a comparably smaller batch,the application of BN (Batch Normalization)tends to compromise the model’s convergence efficacy.Further,the employment of the cross-entropy loss function overlooks the intricate dependencies present amongst adjacent pixels.In light of these challenges,enhancements were made to the UPerNet,with its overarching structure delineated in Fig.2.The encoder network,seeking refuge in Swin-T [31] for feature extraction,collaborates with the decoder network,which harnesses both PPM and FPN.This ensures a seamless projection of discernible features into the pixel space,post-upsampling.The final ensemble of feature maps,stemming from each stage of the feature extraction network,is encapsulated as {S2,S3,S4,S5}.Concurrently,the feature mapping borne out of the FPN output is represented as {P2,P3,P4,P5},with P5 also succeeding PPM directly.

    Figure 2:The overall structure of the method in this paper

    3.1 Backbone Network

    While CNNs have demonstrated proficiency across various image processing tasks,the inherent locality of convolutional operations and the inductive bias stemming from weight sharing impose constraints on long-distance dependencies.Such limitations curtail the ability of CNNs to glean comprehensive insights from an entire image and to facilitate information exchange beyond the confines of the convolutional kernel.In stark contrast,the self-attention mechanism inherent in the Transformer can adaptively modify its receptive domain,thereby furnishing long-range dependencies and circumventing the inductive bias endemic to CNNs.In this study,the Swin-T architecture is employed as the backbone network,facilitating the extraction of expansive global information and more nuanced features.The strategic confinement of attention computation to non-overlapping local windows ensures the computational complexity remains linear in relation to image size,thereby optimizing computational efficiency.A schematic representation of this network structure can be perused in Fig.3.

    Figure 3:Network structure of Swin-T

    The entire network adopts a stratified approach.Initially,an input image of dimensionsH×W× 3 is partitioned into (H/4) × (W/4) non-overlapping patches via the patch-splitting module,subsequently flattening each patch into a 48-dimensional token vector.This is followed by the linear embedding layer projecting the tensor of dimensions(H/4)×(W/4)×48 to an arbitrary dimensionC,culminating in a linear mapping of dimension(H/4)×(W/4)×C.These mapped tokens are then channelled into the dual Swin Transformer module,which ensures the token count remains consistent at(H/4)×(W/4)both at the input and output.The inaugural Swin Transformer module,in conjunction with the linear embedding layer,is designated as the primary layer.Thereafter,the patch merging layer collaborates with the Swin Transformer module across subsequent layers to elucidate feature descriptors across varied scales.Swin Transformer diverges from traditional architectures like ResNet,opting to segregate the input image via diminutive windows,consequently parsing it into equitably sized blocks termed local windows.Within each of these local windows,feature extraction ensues,thereby permitting both self-attention and localized convolutional operations to ascertain feature information across diverse scales and positions.numerous Swin Transformer Blocks(analogous to the basic blocks in ResNet)are harnessed for feature extraction across Stages 2–4.Eschewing the simplistic stacking of convolutional layers as observed in ResNet,Swin Transformer adopts a hierarchical group convolution to optimize feature amalgamation.Furthermore,Swin Transformer facilitates inter-window information exchange,enhancing the context-dependence of features.As the network becomes progressively intricate,every layer metamorphoses the tensor’s dimension,culminating in a stratified representation.Pertinently,each patch boasts a size of 4,each self-attention window spans 7,and the network depth across the four blocks is catalogued as[2,2,6,2].In summation,the multifaceted stages of Swin Transformer utilize local windows,hierarchical group convolution,and inter-window connections to glean feature insights across an array of positions and scales.This epitomizes Swin Transformer’s adeptness in tailoring its feature extraction capabilities to cater to diverse tasks and datasets.

    The Swin Transformer Block is composed of Layer Normalization (LN),a multi-head selfattention module,a shortcut connection,and anMLPfeaturing a Gaussian Error Linear Unit(GELU)activation function.A representation of two consecutively arranged Swin Transformer Blocks is depicted in Fig.4.Within the design of the multi-head self-attention module,the self-attention mechanism is intricately integrated into the window,encompassing two distinct types: the windowbased self-attention (W-MSA) and the shifted window-based self-attention (SW-MSA).TheWMSAmode partitions the image into multiple windows and executes self-attention computations individually for each window.Given an image fragmented intoh×wpatches,where each window comprisesM×Mpatches,the computational complexities ofMSAandW-MSAare elucidated in formulas(4) and (5),respectively.A noteworthy reduction in computational complexity is achieved throughW-MSA.Owing to the fact that the patch count within each window is significantly inferior to the total patch count of the image,the computational complexity associated withW-MSAexhibits linearity with respect to the image’s dimensions.

    SW-MSA,on the other hand,incorporates a shifted window approach,facilitating inter-window connections.This not only augments information transfer across different windows but also amplifies the receptive field,thereby bolstering the model’s representational prowess.The interrelationships embedded within the two serially arranged Swin Transformer Blocks are expounded in formula(6),where ?zlandzlrepresent the features output by the two types of window self-attention and the features output by theMLP,respectively.

    Figure 4:Two swin Transformer blocks in series

    3.2 Synchronized Batch Normalization

    In traditional applications,when batch normalization (BN) serves as the chosen normalization technique,the inputx={x1,x2,x3,x4} is segmented into multiple portions during each iteration of network training.These portions undergo forward and backward propagation processes,with the gradients being computed on disparate GPUs.Subsequently,a succeeding iteration is commenced post the amalgamation of gradients and parameters.This fundamental operation is visualized in Fig.5.

    Figure 5:Working principle of BN

    Due to each GPU processing data individually during both forward and backward propagations,the BN operation is independently executed on every GPU.Intrinsically,the samples normalized via BN are confined to their respective GPU,mirroring a reduction in the batch_size.A diminished batch_size can lead to the calculated mean and variance failing to capture the holistic data distribution,as a smaller batch_size introduces heightened variability.Notably,in the realm of deep learning,an expansive batch_size typically stabilizes the training trajectory.Cross-GPU synchronized batch normalization(SyncBN)retrieves the universal meanμand varianceσ2during the forward propagation phase,and computes the overarching gradient throughout the backward propagation.When deducing the global gradient in the backward propagation,the global BN statistics can be employed for input normalization,which invariably refines the accuracy of these statistics,as depicted in Fig.6.

    SyncBN ensures that the input data distribution across every network layer remains comparably consistent,permitting the adoption of larger learning rates.This facilitates the model’s diminished sensitivity towards network parameters,attenuates its reliance on initialization,simplifies the parameter tuning procedure,and promotes stable network learning.By leveraging normalized data subsequent to the activation function,SyncBN sustains gradients at relatively ample values,averting gradient vanishing or explosive behaviors.Such synchronized normalization has been effectively incorporated in applications like power grid device image anomaly detection[32]and apparel image analysis[33],facilitating seamless synchronization of network parameters across GPUs.

    Figure 6:Working principle of SyncBN

    Within each GPU,both the summations Σxiand Σxi2,are computed.Following this,a synchronized summation is executed to deduce the global variance and mean.Such a method allows the entire calculation to be finalized with a singular synchronization.This effectively reduces the number of cross-GPU synchronizations from two to just one,enhancing the speed of synchronized normalization,as articulated in formulas(7)and(8).Herein,xirepresents the sample point,andmis the cumulative number of sample points spanning multiple GPUs.

    3.3 Loss Function

    In scenarios where cross-entropy functions as the segmentation loss,certain complications arise.Particularly,when the quantity of foreground pixels is significantly lower than that of background pixels,the inherent competition mechanism of the cross-entropy loss function results in a predominance of the background pixels.Such a skew can severely bias the optimization trajectory of the model towards the background,proving detrimental,especially for segmenting minute rail surface defects.Additionally,cross-entropy loss,given its pixel-by-pixel computation nature,overlooks interpixel relationships.It is crucial to note that images inherently contain significant dependencies between pixels,which often encapsulate vital information about the structure of the subject.Relying solely on cross-entropy loss could lead to segmentation outputs that harbor imprecise,ambiguous predictions.

    Weeks later, when the real seeds began to break through the soil, the boy became entranced with the tiny seedlings6. He spent many afternoons helping7 Granddad water and hoe and watch them grow. And when the first baby vegetables were harvested, he liked them after all.

    To address these challenges,this study integrates the Lovász-hinge loss [34].This loss function inherently embodies the dependencies existing between pixels situated at varied positions within an image and further takes into account the disparity between background and target pixels.The computation associated with Lovász-hinge is inherently linked with the Jaccard Index (IoU).The pertinent loss function is illustrated in formula(9),wherein ?yrepresents the model’s prediction vector,y?denotes the ground truth,andcsymbolizes the object class.The set constituting the mispredicted pixels for classcis articulated in formula(10).Further,formula(9) can be reconceptualized as a function of the misprediction set,as delineated in formula(11).

    In the use of the Jaccard loss,it is mandated that the prediction result be discrete;a continuous prediction result cannot be derived post-discretization.Given that the Jaccard loss operates as a submodule function,its input space can be transformed from discrete {0,1}pto continuous Rpthrough the Lovász extension.Furthermore,this transformed value remains consistent with the output value of the original function on {0,1}p.In addressing binary segmentation challenges such as rail surface defect segmentation,this study employs the continuous and derivable Lovász-hinge,which is attained after the smooth extension of Jaccard loss,as the loss function for optimal model parameterization.The Lovász-hinge loss has also found applications in the first arrival picking problem in seismic signal processing.In this context,the first arrival picking problem is reframed as a contour detection challenge,with the aim of enhancing the accuracy of first arrival picking[35].Kerola et al.[36] introduced a groundbreaking representation learning methodology tailored for panoptic segmentation.They further presented a streamlined architecture and loss mechanism to master pixel-wise embeddings,effectively representing instances,object categories,and stuff classes.This paved the way for a unified embedding-based single-shot panoptic segmentation.The entity ΔJcis reconceptualized as a suite of functions that mispredictm,with the mathematical representation of the incorrect prediction definition illustrated in formula(12).

    The continuous linear distribution function after Lovász extension of ΔJcis shown in formula(13).

    The mispredictionmiis replaced by the hinge loss of the input imagex,as shown in formula(14),whereF(x)represents the network output.

    g(m)is obtained by taking the derivative ofm,as shown in formula(15).

    To sum up,the loss function used in this paper is shown in formula(16).

    4 Experiment

    Experimental verification was undertaken on two datasets in this study.The initial dataset was procured from the transportation railway line of Baiyin Nonferrous Group Co.,Ltd.,where the primary defect type identified was shelling defects.Given that the background image information of rail ties and ballast was redundant for detection,it was consistently transformed into 160 × 370 pixels images encompassing only the rail surface area.A total of 914 defect images were retained post-processing.This dataset was christened the Rail-BY dataset.The subsequent dataset was dubbed the Rail Surface Discrete Defects(RSDDs),curated by researchers from Beijing Jiaotong University[37].Owing to the proactive intervention of the railway maintenance department,a scarcity of rail surface defect samples was observed.An expansion of defect images was executed through strategies like rotation,mirroring,filtering,denoising,and leveraging deep convolutional generative adversarial networks(DCGAN).The augmented Rail-BY dataset encompassed 9570 defect images:8134 in the training set,957 for validation,and 479 for testing.The RSDDs dataset included 6950 defect images,segmented into 5907 for training,695 for validation,and 348 for testing.Representative samples are depicted in Figs.7 and 8,with additional expanded samples in Fig.9.

    Figure 7:A partial sample of the Rail-BY dataset

    Figure 8:A partial sample of the RSDDs dataset

    The apparatuses harnessed for experimentation comprised an Intel Core i9-10920X CPU@3.50 GHz,RAM of 128 GB,dual NVIDIA GeForce RTX 3090 GPU,underpinned by Python 3.8 and PyTorch 1.8.1.During the experiments,a benchmark of 160000 iterations was established,utilizing the AdamW optimizer.Parameters were set atβ1=0.9,β2=0.999,with a learning rate of 0.00006,batch_size of 2,and weight_decay set to 0.01.Loss trajectory for the validation set across the Rail-BY and RSDDs datasets is illustrated in Fig.10.A stability in training loss post the 160000 iterations indicates an optimal training outcome.

    To ascertain the efficacy of the enhanced methodology presented in this study,a comparative examination of the segmentation outcomes of various modules on the two datasets was performed.Table 1 delineates the segmentation outcomes for diverse modules.The foundational backbone network utilized was ResNet50,with normalization facilitated via BN,and the loss function being the cross-entropy loss function.An examination of Table 1 reveals that solely substituting ResNet50 with Swin-T ameliorated PA,IoU,Dice,and Precision by 8.06%,11.42%,7.3%,and 9.52%,respectively on the Rail-BY dataset.Conversely,on the RSDDs dataset,improvements were noted at 3.48%for PA,4.45% for IoU,2.59% for Dice,and 1.7% for Precision.A comprehensive enhancement in UPerNet’s segmentation accuracy was observed,with the most pronounced improvement reflected in the original UPerNet.Additionally,the incorporation of Swin-T for feature extraction culminated in a reduction of Params by 6.55 M.The backbone network grounded in the Transformer architecture decidedly surpassed CNN-based networks in terms of performance.The employment of the SyncBN normalization technique and the Lovász-hinge loss function predominantly augmented PA.Given that these don’t pertain to network structure modifications,their impact on network complexity was minimal.

    Table 1:Comparison of segmentation results of different modules in improved UPerNet

    Table 2 presents the training outcomes of the refined UPerNet network juxtaposed against other prevailing semantic segmentation networks across both datasets.The methodology advanced in this study manifested commendable outcomes on both datasets.When pitted against the high-performing DeepLabV3+,the Rail-BY dataset witnessed enhancements of 6.88%in PA,6.83%in IoU,and 4.2%in Dice.Precision escalated by 1.38%,whereas on the RSDDs dataset,PA surged by 1.13%,IoU by 0.64%,Dice by 0.37%,but Precision experienced a decrement of 0.41%.The original UPerNet network recorded an FPS of 20.1 Hz on the Rail-BY dataset and 20.2 Hz on the RSDDs dataset.However,the FPS of the improved UPerNet on the datasets diminished by 2.2 Hz and 2.3 Hz,respectively,rendering its detection speed performance relatively moderate.Notwithstanding,while the parameter count of the enhanced method presented herein was reduced relative to the original UPerNet model,it remains considerably expansive when compared to other mainstream semantic segmentation techniques.A visual comparison between the methodology advanced in this study and the outcomes from other segmentation networks is portrayed in Fig.11.

    Table 2:Comparison of defect segmentation results between proposed method and other networks

    Figure 11:(Continued)

    Figure 11:Comparison of visualization results between proposed method and other semantic segmentation methods

    5 Defects Parameter Calculation

    Deep learning detection algorithms prevalent today are predominantly employed to ascertain the presence of rail defects,yet they seldom incorporate specific parameters associated with these defects.Consequently,in this work,the actual length,width,and area of rail surface defects are computed based on semantic segmentation.Such a computation aids railway maintenance personnel by providing an intuitive comprehension of the exact dimensions of the defects,enabling them to undertake appropriate welding repairs.In the endeavor to calculate the length,width,and area of these defects,distinguishing different defects within the semantic segmentation image is imperative.For this purpose,a method founded on connected component analysis was utilized in this study.As the nomenclature implies,a connected component encompasses interconnected pixels within an image segment characterized by analogous pixel values.By deploying connected component analysis,individual components within an image can be discerned and marked.This technique has garnered considerable traction across various computer vision tasks.However,to obviate any potential interference caused by fluctuations in pixel values when isolating distinct connected components,images are typically subjected to binarization preceding connected component analysis.Integral to this analysis is the determination of neighborhood connectedness.Two predominant criteria for this assessment are the 4-connected neighborhood and the 8-connected neighborhood,the principles of which are delineated in Fig.12.

    Figure 12:4-connected neighborhood and 8-connected neighborhood of adjacent pixels

    In the realm of the 4-connected neighborhood definition,two pixels can be adjudged as neighboring solely if they exhibit contiguous connections either horizontally or vertically.When pointP0(x,y)is examined under the 4-connected neighborhood premise,the coordinates of its four adjacent pixels emerge asP1(x-1,y),P2(x+1,y),P3(x,y-1),andP4(x,y+1).Conversely,under the 8-connected neighborhood paradigm,besides the horizontal and vertical axes,pixels are also deemed interconnected if they adjoin across the diagonal axes.Accordingly,for the pointP0(x,y),the eight neighboring pixel coordinates,when evaluated under the 8-connected neighborhood definition,areP1(x-1,y),P2(x+1,y),P3(x,y-1),P4(x,y+1),P5(x-1,y-1),P6(x+1,y-1),P7(x-1,y+1),andP8(x+1,y+1).The methodology outlined in this study harnesses the 8-connected neighborhood definition.

    Upon subjecting the rail surface defects to semantic segmentation image analysis using connected components,the OpenCV-Python function cv2.ConnectedComponentsWithStats() is leveraged to discern the connected components.This function yields the count of connected components,the expanse of each individual component,and the dimensions of the encompassing rectangle.The procedural sequence through which defect parameters are derived,rooted in connected component analysis as delineated in this research,is exhibited in Fig.13.

    Figure 13:Connected component analysis process of this article

    Firstly,the semantic segmentation image,in the initial phase,is subjected to denoising via median filtering.Following this,it undergoes transformation into a binary grayscale image.Subsequent to this transformation,a binary image emerges through threshold segmentation.The concluding step embarks upon connected component analysis.From this analysis,several metrics can be derived:the quantity of connected components within each image,pixel markings within the image,and comprehensive statistics corresponding to each marking—encompassing the length,width,and area of the respective contours.To ensure distinctive identification of different connected components,a colormap facilitates the filling of each with a unique hue.Fig.14 exhibits the outcomes of this connected component analysis.It is imperative to note that the defect boundaries used within the connected component analysis are devoid of direction.As such,the included analysis does not delve into boundary directionality.Therefore,the graphics leveraged for connected component analysis in this research are undirected,rendering them neither as strongly nor weakly connected graphs.

    Figure 14:Example of connected component analysis results

    The rail head images featured in this study have an actual width of 73 mm,and all utilized rail surface defect images maintain a dimension of 160 × 370 pixels.Leveraging the proportionality principle,the actual length of the rail within the image is discerned to be approximately 168.81 mm.This translates to a total rail areaSentire_realof roughly 12323.31 mm2.formula(21)outlines the calculation methodology for the real rail defect length,drawing upon the ratio between the defect length in the semantic segmentation imageLdefect_segand the entirety of the image lengthLentire_seg.It’s specified thatLentire_real,representing the real rail length within the image,stands at 168.81 mm.

    According to the percentage of defect area in the semantic segmentation imageSdetect_segto the entire image areaSentire_seg,the actual rail defect areaSdefect_realcan be calculated,as shown in formula(22),whereSentire_realrepresents that the actual area of the rail area corresponding to the semantic segmentation image is 12323.31 mm2.

    Post connected component analysis,the actual dimensions of the defect within the image—its length,width,and area—can be deduced by incorporating the relevant coefficient.Given that the image’s background pixels too form part of a connected component,any component exceeding an area of 30000 pixels is classified as the backdrop.Fig.15 presents the algorithm’s operational outcomes.In this depiction,start_x and start_y signify the x and y coordinates of the starting point of the connected component,respectively.Meanwhile,“width”represents its image width,“height”corresponds to its image length,and “area” delineates its space within the image.The terms “actual width”,“actual height”,and“actual area”convey the real-world dimensions of the connected component.

    Figure 15:Operating results of damage area parameter calculation

    For example,in Fig.14a,an area,depicted in purple and characterized as connected,was found to be 64 pixels in length and 23 pixels in width,resulting in an area of 960 pixels.When translated to real-world dimensions,the defect exhibited an actual length of 29.18 mm,a width of 10.49 mm,and spanned an area of 199.68 mm2.Additionally,another connected region,colored in coffee,measured 131 pixels in length,45 pixels in width,and occupied an area of 3803 pixels.Correspondingly,this defect’s actual dimensions were determined to be 59.74 mm in length,20.52 mm in width,and an area of 791.02 mm2.For Fig.14b,the analysis revealed a connected area measuring 77 pixels in length,55 pixels in width,with a total area of 2899 pixels.The actual measurements of this defect were found to be 35.11 mm in length,25.08 mm in width,and 602.99 mm2in area.In Fig.14c,a blue connected area was identified with a length of 28 pixels,a width of 36 pixels,and an overall area of 650 pixels.In actual measurements,the defect was 12.77 mm long,16.42 mm wide,and spanned an area of 135.20 mm2.A second region,colored in coffee,was 26 pixels in length,41 pixels in width,and covered an area of 819 pixels.The actual measurements of this defect were 11.86 mm in length,18.70 mm in width,and an area of 170.35 mm2.A third region,depicted in purple,was found to be 40 pixels long,28 pixels wide,and had an area of 676 pixels.The real-world measurements of this defect were 18.24 mm in length,12.77 mm in width,and an area of 140.61 mm2.Fig.14d showcased a coffee-colored connected area that was 22 pixels long,21 pixels wide,with an area encompassing 353 pixels.This translated to an actual length of 10.03 mm,a width of 9.58mm,and an area of 73.42 mm2.Another area,colored in purple,had dimensions of 20 pixels in length,22 pixels in width,and covered 361 pixels.The real-world dimensions of this defect were determined to be 9.12 mm in length,10.03 mm in width,and 75.09 mm2in area.

    Through the use of connected component analysis,we are able to determine the actual dimensions of the defect,including its length,width,and area.This provides maintenance personnel with a clear and intuitive understanding of the defect’s parameters,enabling them to undertake targeted maintenance actions.

    6 Conclusion

    In addressing the inefficiencies of conventional machine vision methods and the irregular and intricate defect shapes encountered in rail surface defect detection,an enhanced UPerNet algorithm is presented in this study.Initially,Swin-T is employed as the backbone network to circumvent issues related to inductive bias and the locality of feature extraction inherent in CNN-based backbone networks.Given the influence of BN on the convergence of the semantic segmentation mode,SyncBN is adopted to refine the model,thereby enhancing the training stability during multi-GPU computations.Conclusively,the Lovász-hinge is selected as the loss function to address the challenge posed by cross-entropy loss,which tends to overlook the interdependence between pixels.Experimental findings indicate that the proposed methodology augments detection accuracy across both datasets.Specifically,for the Rail-BY dataset,the PA was recorded at 93.39%,the IoU at 83.69%,the Dice at 91.12%,and the Precision at 90.85%.Similarly,for the RSDDs dataset,the PA,IoU,Dice,and Precision values were determined to be 93.35%,87.58%,93.38%,and 93.41%,respectively,illustrating the method’s proficiency in accurately executing rail surface defect segmentation tasks.

    Subsequent to the semantic segmentation of rail surface defects,connected component analysis technology is utilized to discern distinct defects within a singular image.Thereafter,by enumerating the pixel count,the authentic length and area of each defect zone are deduced.Such insights prove invaluable for railway maintenance personnel in ascertaining the extent of the inflicted damage.

    Acknowledgement:Thanks to our tutors and researchers for their assistance and guidance.The authors also gratefully acknowledge the helpful comments and suggestions of the reviewers,which have improved the presentation.

    Funding Statement:This work is supported in part by the National Natural Science Foundation of China(Grant No.62066024).Gansu Province Higher Education Industry Support Plan(2021CYZC-34),Lanzhou Talent Innovation and Entrepreneurship Project(2021-RC-27,2021-RC-45).

    Author Contributions:Study conception and design: Yongzhi Min,Jiafeng Li,Yaxing Li;data collection:Jiafeng Li;analysis and interpretation of results:Yongzhi Min,Jiafeng Li,Yaxing Li;draft manuscript preparation:Jiafeng Li,Yaxing Li.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data used in this paper can be requested from the corresponding author upon request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    国产成人欧美在线观看| 99国产精品一区二区蜜桃av| 欧美丝袜亚洲另类 | 亚洲av免费在线观看| 国产黄色小视频在线观看| 国产淫片久久久久久久久 | 国产 一区 欧美 日韩| 国产熟女xx| 精品电影一区二区在线| 国产黄a三级三级三级人| 成人无遮挡网站| 99热这里只有精品一区| 国产精品99久久99久久久不卡| 1000部很黄的大片| 一进一出好大好爽视频| 国产精品久久视频播放| 亚洲国产中文字幕在线视频| 少妇高潮的动态图| 日韩欧美一区二区三区在线观看| 亚洲精华国产精华精| 国产v大片淫在线免费观看| 韩国av一区二区三区四区| 精品人妻偷拍中文字幕| 午夜福利免费观看在线| 国产美女午夜福利| 在线国产一区二区在线| 人人妻人人澡欧美一区二区| 亚洲av电影在线进入| 国产真实乱freesex| 亚洲在线自拍视频| 亚洲国产高清在线一区二区三| 亚洲美女视频黄频| 性色avwww在线观看| 色尼玛亚洲综合影院| 99久久九九国产精品国产免费| 亚洲国产精品999在线| 听说在线观看完整版免费高清| 国产精品美女特级片免费视频播放器| 久久久久久久久中文| 99久久久亚洲精品蜜臀av| 97超视频在线观看视频| 精品一区二区三区视频在线 | 国产一区二区三区在线臀色熟女| 最新在线观看一区二区三区| 香蕉久久夜色| 男人舔女人下体高潮全视频| 日韩欧美精品免费久久 | 国产精品亚洲一级av第二区| 欧美黄色片欧美黄色片| 18禁黄网站禁片午夜丰满| 精品国内亚洲2022精品成人| 国产免费一级a男人的天堂| 日韩欧美三级三区| 在线观看免费午夜福利视频| 中文字幕av在线有码专区| 日本免费一区二区三区高清不卡| 欧美一区二区国产精品久久精品| 久久精品国产99精品国产亚洲性色| 91九色精品人成在线观看| 亚洲精品一区av在线观看| 九九热线精品视视频播放| 欧美日韩一级在线毛片| 国产精品亚洲美女久久久| 男人舔奶头视频| 噜噜噜噜噜久久久久久91| 少妇人妻精品综合一区二区 | 欧美不卡视频在线免费观看| 一区福利在线观看| 男插女下体视频免费在线播放| 男女那种视频在线观看| 亚洲专区中文字幕在线| 欧美一区二区亚洲| 欧美中文日本在线观看视频| 亚洲av二区三区四区| 搡女人真爽免费视频火全软件 | 好看av亚洲va欧美ⅴa在| 午夜影院日韩av| 香蕉丝袜av| 性欧美人与动物交配| 99国产精品一区二区三区| 亚洲精品日韩av片在线观看 | 色综合婷婷激情| 此物有八面人人有两片| 国产高清有码在线观看视频| 丰满人妻熟妇乱又伦精品不卡| 在线a可以看的网站| 他把我摸到了高潮在线观看| av女优亚洲男人天堂| 丁香六月欧美| 午夜福利在线在线| 欧美激情在线99| 内地一区二区视频在线| 小说图片视频综合网站| 亚洲18禁久久av| 国产欧美日韩一区二区精品| 久久精品国产综合久久久| 亚洲自拍偷在线| 国产黄色小视频在线观看| 欧美极品一区二区三区四区| 婷婷精品国产亚洲av| 国产精品久久久久久精品电影| 日韩高清综合在线| 脱女人内裤的视频| 国产真实伦视频高清在线观看 | 亚洲av电影在线进入| 国产97色在线日韩免费| 欧美日韩瑟瑟在线播放| 久久中文看片网| 男人和女人高潮做爰伦理| 免费人成在线观看视频色| 成人国产一区最新在线观看| 日本a在线网址| 亚洲精品乱码久久久v下载方式 | 亚洲第一欧美日韩一区二区三区| 亚洲色图av天堂| 香蕉av资源在线| 青草久久国产| 午夜视频国产福利| 国产一区二区在线观看日韩 | 国产精品乱码一区二三区的特点| 国内揄拍国产精品人妻在线| 好男人在线观看高清免费视频| 国产av麻豆久久久久久久| 97碰自拍视频| 精品免费久久久久久久清纯| 亚洲精品乱码久久久v下载方式 | 欧美中文综合在线视频| 免费在线观看影片大全网站| a级毛片a级免费在线| 91av网一区二区| 免费人成视频x8x8入口观看| 成熟少妇高潮喷水视频| 一区二区三区激情视频| 亚洲国产欧美人成| 法律面前人人平等表现在哪些方面| 精品不卡国产一区二区三区| 亚洲av日韩精品久久久久久密| 最新在线观看一区二区三区| 欧美最新免费一区二区三区 | 叶爱在线成人免费视频播放| 亚洲不卡免费看| 国产亚洲精品久久久久久毛片| 波多野结衣高清作品| 51国产日韩欧美| 亚洲无线观看免费| 亚洲av免费在线观看| 久久久久国内视频| 色尼玛亚洲综合影院| 国产一级毛片七仙女欲春2| 91av网一区二区| 51午夜福利影视在线观看| 日本在线视频免费播放| 国产精品久久久久久亚洲av鲁大| 最后的刺客免费高清国语| 啦啦啦韩国在线观看视频| 亚洲精品在线观看二区| 午夜老司机福利剧场| av专区在线播放| 午夜福利18| 夜夜爽天天搞| 99精品在免费线老司机午夜| 精品国产三级普通话版| 女警被强在线播放| 亚洲精品一卡2卡三卡4卡5卡| 日本精品一区二区三区蜜桃| 国产v大片淫在线免费观看| 在线天堂最新版资源| 亚洲精品粉嫩美女一区| 国产成人啪精品午夜网站| 两个人的视频大全免费| 久久久久久久精品吃奶| 可以在线观看毛片的网站| 村上凉子中文字幕在线| 青草久久国产| 国产精品国产高清国产av| 美女高潮喷水抽搐中文字幕| 一级黄片播放器| 又黄又爽又免费观看的视频| 国产一区二区在线av高清观看| 老司机午夜福利在线观看视频| 在线a可以看的网站| 久久精品综合一区二区三区| 尤物成人国产欧美一区二区三区| 国产乱人视频| 久久久久国内视频| 日韩国内少妇激情av| 久久精品国产清高在天天线| 首页视频小说图片口味搜索| 日本黄大片高清| 超碰av人人做人人爽久久 | eeuss影院久久| 成年人黄色毛片网站| 91在线观看av| 99在线人妻在线中文字幕| 欧美3d第一页| 欧美一区二区精品小视频在线| 一夜夜www| 精品久久久久久久久久免费视频| 久久国产乱子伦精品免费另类| 午夜老司机福利剧场| 免费av观看视频| 久久这里只有精品中国| 51国产日韩欧美| 国语自产精品视频在线第100页| 一二三四社区在线视频社区8| 欧美丝袜亚洲另类 | 亚洲最大成人中文| 露出奶头的视频| 成人18禁在线播放| 亚洲国产精品sss在线观看| 久久久久九九精品影院| 内射极品少妇av片p| 日本精品一区二区三区蜜桃| 欧美国产日韩亚洲一区| 国产真实伦视频高清在线观看 | 亚洲精品影视一区二区三区av| 国产激情偷乱视频一区二区| 色噜噜av男人的天堂激情| 在线观看免费午夜福利视频| 老熟妇乱子伦视频在线观看| 18禁在线播放成人免费| 成人一区二区视频在线观看| 五月伊人婷婷丁香| 亚洲欧美激情综合另类| 在线天堂最新版资源| 叶爱在线成人免费视频播放| 99久久精品一区二区三区| 男女下面进入的视频免费午夜| 亚洲成人久久爱视频| 国产毛片a区久久久久| 国产av不卡久久| 内地一区二区视频在线| 91av网一区二区| 国产真实乱freesex| 白带黄色成豆腐渣| 性色av乱码一区二区三区2| 亚洲在线观看片| 91av网一区二区| 欧美日韩福利视频一区二区| 日韩中文字幕欧美一区二区| 亚洲黑人精品在线| 成人精品一区二区免费| 99精品欧美一区二区三区四区| 少妇的丰满在线观看| 国产一区二区三区视频了| 欧美zozozo另类| 久久久久久久久中文| 国产私拍福利视频在线观看| 亚洲人成网站在线播| 日韩人妻高清精品专区| 亚洲最大成人手机在线| 婷婷精品国产亚洲av| 99国产综合亚洲精品| 日本一本二区三区精品| av国产免费在线观看| 国产成人aa在线观看| 精品不卡国产一区二区三区| 久久亚洲真实| 天天一区二区日本电影三级| 超碰av人人做人人爽久久 | 宅男免费午夜| 成人18禁在线播放| av欧美777| 非洲黑人性xxxx精品又粗又长| 国产老妇女一区| 51午夜福利影视在线观看| 99久久精品一区二区三区| 高清日韩中文字幕在线| 蜜桃久久精品国产亚洲av| 波多野结衣高清无吗| 国产亚洲av嫩草精品影院| 两性午夜刺激爽爽歪歪视频在线观看| 老汉色av国产亚洲站长工具| 黑人欧美特级aaaaaa片| 99久久精品热视频| 内地一区二区视频在线| 欧美bdsm另类| 国产v大片淫在线免费观看| 日韩欧美国产一区二区入口| 久99久视频精品免费| 日本成人三级电影网站| 91久久精品国产一区二区成人 | 长腿黑丝高跟| 久久人妻av系列| 国产一区在线观看成人免费| 身体一侧抽搐| 波多野结衣高清作品| 可以在线观看的亚洲视频| 欧美+日韩+精品| 日本 av在线| 麻豆国产av国片精品| 色老头精品视频在线观看| 久久久久久大精品| 欧美最黄视频在线播放免费| 美女高潮喷水抽搐中文字幕| 久久国产精品人妻蜜桃| 宅男免费午夜| 在线播放国产精品三级| 国产高清激情床上av| 色视频www国产| 午夜福利高清视频| 精品欧美国产一区二区三| 色老头精品视频在线观看| 18禁在线播放成人免费| 在线观看舔阴道视频| 亚洲精品国产精品久久久不卡| 观看免费一级毛片| 国产精品免费一区二区三区在线| 亚洲天堂国产精品一区在线| 国产蜜桃级精品一区二区三区| 99久久综合精品五月天人人| 国产真实伦视频高清在线观看 | 午夜精品在线福利| 日韩中文字幕欧美一区二区| 亚洲五月天丁香| 久久久久国产精品人妻aⅴ院| 国内毛片毛片毛片毛片毛片| 欧美高清成人免费视频www| 五月玫瑰六月丁香| 午夜精品一区二区三区免费看| 久久精品国产清高在天天线| 制服人妻中文乱码| av欧美777| 国产v大片淫在线免费观看| 午夜福利在线在线| 两人在一起打扑克的视频| 午夜福利高清视频| 亚洲人与动物交配视频| 此物有八面人人有两片| 久久久久性生活片| 亚洲精品在线观看二区| 国产亚洲欧美98| 午夜福利在线在线| 亚洲成a人片在线一区二区| 精品电影一区二区在线| bbb黄色大片| 国产精品综合久久久久久久免费| 亚洲国产色片| 国内少妇人妻偷人精品xxx网站| 国产亚洲精品久久久com| av福利片在线观看| a在线观看视频网站| 亚洲色图av天堂| 女人十人毛片免费观看3o分钟| 日韩国内少妇激情av| 免费观看人在逋| 国产亚洲精品一区二区www| 床上黄色一级片| 脱女人内裤的视频| 欧美一区二区国产精品久久精品| 欧美高清成人免费视频www| 成人特级av手机在线观看| 国产激情偷乱视频一区二区| 一个人免费在线观看的高清视频| 亚洲中文日韩欧美视频| 一个人免费在线观看的高清视频| 好看av亚洲va欧美ⅴa在| 天天添夜夜摸| 内射极品少妇av片p| 制服丝袜大香蕉在线| 久久精品亚洲精品国产色婷小说| 91在线精品国自产拍蜜月 | 男人的好看免费观看在线视频| 婷婷精品国产亚洲av| 亚洲精品影视一区二区三区av| 国产黄片美女视频| 日韩欧美精品免费久久 | 亚洲av美国av| 国产探花在线观看一区二区| 久久精品影院6| 亚洲av成人精品一区久久| 亚洲精品色激情综合| 久久人妻av系列| 久久久久久久久久黄片| 国产精品三级大全| 国产亚洲精品综合一区在线观看| 亚洲av一区综合| 内射极品少妇av片p| 18禁美女被吸乳视频| 国产精品乱码一区二三区的特点| 国产精品久久久久久人妻精品电影| 日韩国内少妇激情av| 最近最新中文字幕大全免费视频| 老熟妇乱子伦视频在线观看| 最后的刺客免费高清国语| 亚洲中文字幕日韩| 国产精品嫩草影院av在线观看 | 欧美极品一区二区三区四区| 成人国产一区最新在线观看| 欧美av亚洲av综合av国产av| 久久精品国产亚洲av香蕉五月| 一进一出抽搐动态| 久久精品国产亚洲av涩爱 | 中国美女看黄片| 亚洲成av人片在线播放无| 九九久久精品国产亚洲av麻豆| tocl精华| 欧美乱色亚洲激情| 久久久久久久久久黄片| 老司机午夜福利在线观看视频| 国产精品国产高清国产av| 久久久久久国产a免费观看| 人妻久久中文字幕网| 亚洲精品亚洲一区二区| www.www免费av| 又黄又粗又硬又大视频| 国产高清视频在线观看网站| 蜜桃亚洲精品一区二区三区| av在线蜜桃| 丝袜美腿在线中文| 中文资源天堂在线| 欧美黑人欧美精品刺激| 人妻久久中文字幕网| 日韩欧美 国产精品| 啪啪无遮挡十八禁网站| 小蜜桃在线观看免费完整版高清| 97超视频在线观看视频| 啪啪无遮挡十八禁网站| 亚洲精品色激情综合| 欧美一级毛片孕妇| 日本五十路高清| 毛片女人毛片| 99久国产av精品| 国产三级在线视频| 色视频www国产| 男人舔奶头视频| 日本 欧美在线| 国产精品一区二区三区四区免费观看 | 亚洲精品一区av在线观看| 99在线视频只有这里精品首页| 又黄又爽又免费观看的视频| 精品免费久久久久久久清纯| 国产熟女xx| 欧美一区二区国产精品久久精品| 日日夜夜操网爽| 99精品欧美一区二区三区四区| 成人18禁在线播放| 久久精品国产自在天天线| 精品午夜福利视频在线观看一区| 久久久久久人人人人人| 国产淫片久久久久久久久 | av女优亚洲男人天堂| 一区二区三区激情视频| 热99re8久久精品国产| 丁香六月欧美| 亚洲国产欧洲综合997久久,| 特大巨黑吊av在线直播| 日本免费一区二区三区高清不卡| 久久久色成人| 身体一侧抽搐| 成年人黄色毛片网站| 国产老妇女一区| 69av精品久久久久久| 亚洲无线观看免费| 久久久久久久午夜电影| 夜夜躁狠狠躁天天躁| 一进一出好大好爽视频| 色av中文字幕| av天堂在线播放| 久久午夜亚洲精品久久| 麻豆成人av在线观看| 午夜亚洲福利在线播放| 真人做人爱边吃奶动态| АⅤ资源中文在线天堂| 真实男女啪啪啪动态图| 一级a爱片免费观看的视频| 成人av在线播放网站| 1000部很黄的大片| 啪啪无遮挡十八禁网站| 成人国产综合亚洲| 久久久久久久精品吃奶| 欧美日韩瑟瑟在线播放| 最近最新中文字幕大全免费视频| 欧美绝顶高潮抽搐喷水| 搡老妇女老女人老熟妇| 搡老熟女国产l中国老女人| 我要搜黄色片| 欧美日韩国产亚洲二区| 国产精品爽爽va在线观看网站| 看黄色毛片网站| 国产av不卡久久| 国产激情偷乱视频一区二区| 欧美日韩中文字幕国产精品一区二区三区| xxx96com| 一二三四社区在线视频社区8| 脱女人内裤的视频| 九色成人免费人妻av| 久久人妻av系列| 亚洲av成人av| 国内精品久久久久精免费| 久久精品影院6| 看免费av毛片| 国产av不卡久久| 看免费av毛片| 99在线人妻在线中文字幕| 国产探花在线观看一区二区| 长腿黑丝高跟| 午夜视频国产福利| 国产精品免费一区二区三区在线| 成人亚洲精品av一区二区| 99国产极品粉嫩在线观看| 精品熟女少妇八av免费久了| 亚洲精品在线美女| 18美女黄网站色大片免费观看| 亚洲国产精品合色在线| 成人午夜高清在线视频| 亚洲一区二区三区不卡视频| 久久精品影院6| 久久国产精品人妻蜜桃| 国产不卡一卡二| 国产精品久久久久久久久免 | 在线观看舔阴道视频| 18禁在线播放成人免费| 国产伦在线观看视频一区| 香蕉丝袜av| 欧美+日韩+精品| 18禁美女被吸乳视频| 特级一级黄色大片| 一夜夜www| 天天一区二区日本电影三级| 国产高清有码在线观看视频| 特级一级黄色大片| 在线观看美女被高潮喷水网站 | 97超级碰碰碰精品色视频在线观看| 亚洲成人久久性| 国产亚洲精品久久久com| 亚洲一区二区三区色噜噜| 久9热在线精品视频| 日韩欧美 国产精品| 国产真实乱freesex| 亚洲专区中文字幕在线| 欧美日韩福利视频一区二区| 麻豆久久精品国产亚洲av| 青草久久国产| 国产亚洲精品久久久久久毛片| 一本久久中文字幕| 九九在线视频观看精品| 一本久久中文字幕| 一区二区三区激情视频| a在线观看视频网站| 一级黄色大片毛片| 夜夜爽天天搞| 婷婷精品国产亚洲av| 日韩欧美精品免费久久 | 午夜福利18| 亚洲av电影不卡..在线观看| 色综合亚洲欧美另类图片| 成人18禁在线播放| 欧美午夜高清在线| 99久久成人亚洲精品观看| 婷婷六月久久综合丁香| 欧洲精品卡2卡3卡4卡5卡区| 在线观看一区二区三区| 免费看日本二区| 国产午夜精品久久久久久一区二区三区 | 又粗又爽又猛毛片免费看| 国产美女午夜福利| 日韩人妻高清精品专区| 国产精品美女特级片免费视频播放器| 1000部很黄的大片| 久久久国产成人免费| 亚洲成人久久性| 好男人在线观看高清免费视频| 午夜福利在线观看免费完整高清在 | 国产精品久久电影中文字幕| 美女免费视频网站| 日本三级黄在线观看| 国产伦精品一区二区三区视频9 | 免费无遮挡裸体视频| 亚洲欧美日韩卡通动漫| 国产精品一区二区三区四区免费观看 | 天堂动漫精品| 国内精品久久久久精免费| 亚洲国产欧美人成| 免费一级毛片在线播放高清视频| 一二三四社区在线视频社区8| av专区在线播放| 综合色av麻豆| 久久精品国产综合久久久| 99国产综合亚洲精品| 国产午夜精品论理片| 日本免费a在线| 国产免费男女视频| 超碰av人人做人人爽久久 | 国产午夜精品久久久久久一区二区三区 | 国产不卡一卡二| 亚洲avbb在线观看| 欧美在线黄色| 亚洲欧美精品综合久久99| 精华霜和精华液先用哪个| 亚洲av不卡在线观看| 午夜福利欧美成人| 亚洲国产精品久久男人天堂| 国内揄拍国产精品人妻在线| 国产成人啪精品午夜网站| www.色视频.com| 精品久久久久久久久久免费视频| 亚洲内射少妇av| 一级毛片高清免费大全| 天天添夜夜摸| 欧美精品啪啪一区二区三区| 啪啪无遮挡十八禁网站| 国产精品99久久99久久久不卡| 亚洲美女视频黄频| 色av中文字幕| 男女做爰动态图高潮gif福利片| 色尼玛亚洲综合影院| 99热只有精品国产| 18+在线观看网站| 18禁在线播放成人免费| 亚洲欧美日韩东京热| 日本 欧美在线| 久久精品91蜜桃| 日韩有码中文字幕| 日韩精品青青久久久久久|