• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Interactive medical image segmentation with self-adaptive confidence calibration?#

    2023-09-21 06:31:08ChuyunSHENWenhaoLIQisenXUBinHUBoJINHaibinCAIFengpingZHUYuxinLIXiangfengWANG

    Chuyun SHEN ,Wenhao LI ,Qisen XU ,Bin HU ,Bo JIN,Haibin CAI,Fengping ZHU,Yuxin LI,Xiangfeng WANG?

    1School of Computer Science and Technology, East China Normal University, Shanghai 200062, China

    2Huashan Hospital, Fudan University, Shanghai 200040, China

    3Software Engineering Institute, East China Normal University, Shanghai 200062, China

    Abstract: Interactive medical image segmentation based on human-in-the-loop machine learning is a novel paradigm that draws on human expert knowledge to assist medical image segmentation.However,existing methods often fall into what we call interactive misunderstanding,the essence of which is the dilemma in trading offshort-and long-term interaction information.To better use the interaction information at various timescales,we propose an interactive segmentation framework,called interactive MEdical image segmentation with self-adaptive Confidence CAlibration (MECCA),which combines action-based confidence learning and multi-agent reinforcement learning.A novel confidence network is learned by predicting the alignment level of the action with short-term interaction information.A confidence-based reward-shaping mechanism is then proposed to explicitly incorporate confidence in the policy gradient calculation,thus directly correcting the model’s interactive misunderstanding.MECCA also enables user-friendly interactions by reducing the interaction intensity and difficulty via label generation and interaction guidance,respectively.Numerical experiments on different segmentation tasks show that MECCA can significantly improve short-and long-term interaction information utilization efficiency with remarkably fewer labeled samples.The demo video is available at https://bit.ly/mecca-demo-video.

    Key words: Medical image segmentation;Interactive segmentation;Multi-agent reinforcement learning;Confidence learning;Semi-supervised learning

    1 Introduction

    Medical image segmentation is a crucial task in computer-assisted medical diagnosis.However,traditional convolutional neural networks (CNNs)(Wang et al.,2018,2019;Liao et al.,2020) used in segmentation algorithms often struggle to meet clinical standards due to the variability of pathological conditions,the presence of dark lesion areas,and the inconsistent quality of training data.Factors such as discrepancies in imaging scanners,operators,and annotators make the task of medical image segmentation more challenging.To improve accuracy,interactive image segmentation algorithms (Lin D et al.,2016;Xu et al.,2016;Rajchl et al.,2017;Bredell et al.,2018;Wang et al.,2018,2019;Liao et al.,2020;Ma et al.,2021) have been developed using interactive correction information,such as clicks,scribbles,or bounding boxes,to refine the results.

    Fig.1 illustrates the general interactive segmentation process,which contains two modules,i.e.,an interactive module and a utilization module.In the interactive module,users or experts provide interactive correction information,such as clicks and scribbles.The utilization module then uses this interactive correction information to efficiently refine the previous result.Interactive segmentation algorithms have the potential to outperform traditional algorithms by incorporating additional interactive correction information.The interaction process can be viewed as a continuous collaboration between the model and a human expert.Therefore,an effective interactive segmentation model should be able to incorporate and respond to expert feedback to improve its performance through collaboration.

    Fig.1 Interactive segmentation process.In the interactive module,the expert observes the current (or initial) segmentation and provides correction information (the red hint);in the utilization module,new segmentation is refined based on the correction information.References to color refer to the online version of this figure

    Methods such as BIFSeg (Wang et al.,2018) and DeepIGeoS (Wang et al.,2019) model user interactions as hard constraints through conditional random fields.However,these models focus more on one-step interaction,and refine segmentation after the first step cannot be efficiently used (Liao et al.,2020;Ma et al.,2021).Thus,these models cannot use long-term interaction information efficiently.The follow-up InterCNN (Bredell et al.,2018) and other methods (Liao et al.,2020;Ma et al.,2021) improve DeepIGeoS and BIFSeg,and model the problem as an iterative interaction problem,which is more focused on multi-step interaction.Therefore,these methods can cover the data distribution corresponding to different interaction levels in the training phase to effectively use long-term interactions,while ignoring the stochasticity or uncertainty of the model,which makes it difficult for them to use short-term interaction information effectively.

    Existing interactive methods cannot effectively use short-and long-term interaction information simultaneously,which leads to interactive misunderstanding.Fig.2 presents this phenomenon when implementing the popular interactive segmentation algorithm InterCNN (Bredell et al.,2018) on the BraTS2015 dataset.The algorithm ignores the expert’s correction information (Fig.2a),and is even adversely affected (Fig.2b).These inconsistencies between hint information and the refinement results indicate that existing algorithms still face the critical challenge of inefficient utilization of interactive correction information.The reason for the segmentation refining failure is that,at the end of the interaction process,the total loss of the model will guarantee the main area’s priority and ignore some small but challenging areas,such as edges.The regions that are insensitive to interactive correction information could be considered hard-to-segment regions (Shrivastava et al.,2016;Nie et al.,2019),leading to suboptimal refinement of the segmentation process.If these regions are not given additional attention during the training phase,the segmentation model may prioritize easy-to-segment areas at the expense of these more challenging regions.This problem is more severe for medical images,where the hard-tosegment regions usually are tumor boundaries and are very important for clinical diagnosis and surgery.Therefore,it has become urgent to improve the utilization of correction information,especially for the hard-to-segment regions.

    Fig.2 Segmentation refining failure after long-term interactions on BraTS2015 with InterCNN: (a) the model cannot fully understand or ignore the hint information;(b) the model could misunderstand the correction,which results in a worse result

    Another significant challenge in the development of interactive segmentation algorithms is the need for a large number of labeled images.Accurate annotation of medical images can be time-consuming and costly,making it difficult to obtain a sufficient quantity of high-quality data.On the other hand,it is much more cost-effective to obtain a larger number of unlabeled images,but existing interactive segmentation methods do not use these low-cost resources.As a result,it is important to explore how to use the benefits of unlabeled data to reduce the dependence on expert-annotated images.

    In this paper,we propose a novel interactive segmentation algorithm for three-dimensional (3D) medical images,called interactive MEdical image segmentation with self-adaptive Confidence CAlibration (MECCA).MECCA combines an action-based confidence learning network with the multi-agent reinforcement learning (MARL) framework (Zhang KQ et al.,2021).The action-based confidence network can evaluate the corrective action quality by directly calculating the confidence.Unlike learning the confidence of the overall segmentation result,the confidence of actions could be used to better evaluate whether the segmentation model has correctly used the experts’ interaction information.We formulate the iterative interaction process as a Markov decision process (MDP) to model the dynamic process and further introduce the MARL technique (Lee and Song,2018).Further,instead of setting each image or each patient (a series of images) as the agent,we consider each voxel (the smallest unit in the 3D space) as an agent,while each agent aims to learn the segmentation policy and makes its decision (Furuta et al.,2020).Specifically,after receiving the interactive correction information,each agent will modify its label by changing (increasing or decreasing) the category probability.The novel action-based confidence network will directly evaluate each agent’s corrective action obtained from the MARL corrective policy.Using this action-based confidence network,two additional techniques are further proposed to improve the utilization efficiency of the interactive correction information: (1) compared with former manual rewards,a self-adaptive reward function of each action is constructed which could provide more meticulous feedback under a flexible framework;(2) a simulated label generation mechanism is established using the interactive correction information as the weakly supervisory signal.Combining the confidence network,the simulated label generation mechanism can approximately generate the labels for unsupervised images and reduce over-reliance on labeled data.

    The main contributions of this work are summarized as follows:

    1.A novel framework is proposed for interactive medical image segmentation by combining the action-based confidence learning network with MARL.

    2.A self-adaptive feedback mechanism (selfadaptive reward) is constructed with the actionbased confidence network to alleviate interactive misunderstanding during the interaction process.

    3.Simulated supervisory signals can be generated based on the confidence learning network and actions;hence,much fewer labeled data (ground truth data) are needed to achieve the same performance.

    2 Related works

    2.1 Deep interactive image segmentation

    Xu et al.(2016) segmented images based on CNN interactively.DeepCut (Rajchl et al.,2017) and ScribbleSup (Lin D et al.,2016) both employ weak supervision to establish interactive image segmentation methods.DeepIGeoS (Wang et al.,2019) employs the geodesic distance metric to construct a hint map.The interactive segmentation process can be considered a sequential process which is natural to model by reinforcement learning (RL).Polygon-RNN (Castrejón et al.,2017) fundamentally segments each target as a polygon and iteratively chooses the polygon vertexes through a recurrent neural network (RNN).Polygon-RNN++(Acuna et al.,2018) employs almost the same idea as Polygon-RNN but learns to choose vertexes by RL.SeedNet (Lee and Song,2018) trains an expert interaction generation RL model that obtains new simulated interaction information at each interaction step.IteR-MRL (Liao et al.,2020) and BS-IRIS (Ma et al.,2021) both model the dynamic interaction process as an MDP and employ MARL models to segment images.Some researchers aimed to reduce the annotation cost of interactive image segmentation.IFSL (Feng et al.,2021) introduces interactive learning into the few-shot learning strategy and addresses the annotation burden of medical image segmentation models.IOG (Zhang SY et al.,2020) uses a practical inside-outside guidance approach to minimize the labeling cost.These interactive methods can hardly use experts’short-and long-term interaction information simultaneously,thus creating error correction operations.

    2.2 Uncertainty image segmentation

    Uncertainty estimation is helpful in the context of deployed machine learning systems because it can detect when a neural network is likely to make an incorrect prediction or when the input may be out of distribution.Recent methods generally adopt the neural network to learn what the uncertainty should be for any given input,i.e.,learningbased uncertainty estimation or confidence learning,as demonstrated in Kendall and Gal (2017),Moeskops et al.(2017),DeVries and Taylor (2018b),Hung et al.(2018),Robinson et al.(2018),Jungo and Reyes (2019),and Nie et al.(2019).These methods commonly consist of a segmentation network and a confidence network,and are more computationally efficient than other techniques.Thus,they are better suited when computational resources are limited or when real-time inference is required,such as the interactive segmentation scenario considered in this paper.Specifically,Kendall and Gal (2017) introduced a confidence network to predict the aleatoric and epistemic uncertainties by imitating classic Bayesian tools.The segmentation network of DeVries and Taylor (2018a,2018b) produces two separate outputs: prediction probabilities and a confidence estimate.Confidence estimate is motivated by interpolation between the predicted probability distribution and the target distribution during training,where the degree of interpolation is proportional to the confidence estimate.A series of works (Moeskops et al.,2017;Hung et al.,2018;Nie et al.,2019) focus on incorporating uncertainty estimation in the adversarial learning process,where the segmentation network corresponds to the generator,and the confidence network is the discriminator accordingly.Moeskops et al.(2017) first employed generative adversarial networks (GANs) to improve the CNN-based brain magnetic resonance imaging (MRI) segmentation method.The semi-supervised learning technique was used in Hung et al.(2018) to predict trustworthy regions in unlabeled images.Nie et al.(2019) proposed a difficulty-aware attention mechanism to handle those difficult samples or challenging regions.Different from the previous work on learning uncertainty through imitation,joint training,or adversarial learning,a simple but powerful alternative is to introduce an auxiliary task,such as predicting the overlap between a proposed segmentation and its ground truth (Robinson et al.,2018),or to predict the voxel-wise false positive and false negative (Jungo and Reyes,2019).

    2.3 Remark

    In our proposed algorithm,the confidence network should evaluate the confidence of calibrating actions instead of the segmentation results,which is the most significant difference between MECCA and previous uncertainty estimation methods.Therefore,we design a novel action-oriented auxiliary task to predict whether the direction of voxel-wise action is consistent with the ground truth.

    3 Proposed algorithm

    This section introduces the proposed interactive segmentation algorithm,MECCA,which can iteratively evaluate the refinement actions and feedback to the segmentation model.The algorithm framework follows the MARL structure (Sections 3.1 and 3.2),and an action-based confidence learning module (Section 3.3) is introduced to evaluate the confidence of corrective actions.This action-based confidence learning module can be used to establish the self-adaptive reward scheme and simulated label generation mechanism to use the interactive correction information efficiently.The architecture overview of MECCA is depicted in Fig.3a.The model’s state information includes the original 3D image,the previous segmentation probability,and the hint map generated from interaction and confidence maps.Based on the current state information,the segmentation module gives suggested actions to refine previous segmentation results by adjusting the segmentation probability of each voxel (agent).Further,the state information will be used to evaluate the confidence of the obtained actions through the confidence network,with a confidence map as the output.The self-adaptive reward is designed through a self-adaptive weighting scheme based on the action confidence evaluation (Section 3.4).The self-adaptive reward map can be considered a value map with the same size as the original input image (each agent has its self-adaptive reward).It can reflect the performance of the corresponding agent’s action.In addition,MECCA will suggest some lowconfidence regions for the user to segment during the interaction (Section 3.5).The confidence map can also be used to generate the simulated label by comparison with actions,which will be described in Section 3.6.The newly obtained hint map,the adjusted segmentation probability result,and the original 3D image form a new state.The process described above is repeated until the segmentation result meets the requirements.To emphasize,during the testing stage in Fig.3b,there is no need to calculate the selfadaptive rewards,and at the same time,we need only the obtained actions and suggested interaction areas.

    Fig.3 The architecture (a) and testing stage (b) of MECCA.In (a),the segmentation module outputs actions to change the segmentation probability of each voxel (agent) at each interaction step.Meanwhile,the confidence network will estimate the confidence of the actions,which will generate the self-adaptive reward and simulated label.The confidence map can provide the advice regions of the next interaction step to experts.In (b),all voxels in the segmentation probability map are initialized at a fixed value (set to 0.5 in the experiment).Users or experts randomly mark hint points according to the initialized segmentation probability map

    3.1 MARL-driven interactive segmentation

    We employ the MARL structure to formulate the interactive segmentation process and continuously give error-corrective actions at each interaction step.Letx=(x1,x2,···,xN) denote the input image,wherexi(i=1,2,···,N) denotes theithvoxel of the image.In MARL setting,every voxelxiis treated as an agent with its own refinement policyAt time stept,agentxiobtains actionfrom the segmentation network according to its current state.After taking the action,the agent will receive a rewardaccording to the segmentation result.

    3.2 Segmentation network

    The segmentation network adopts P-Net (Wang et al.,2019) as the backbone.It has two heads: a policy head and a value head.Two heads share the first three 3D convolutional blocks to extract low-level features.Each of the blocks has two convolution layers,and the size of the convolution kernel is fixed as 3×3 in all these convolution layers.Dilated convolution is employed in all the convolution kernels,which can reduce the loss of resolution.Both heads have another two 3D convolutional blocks to extract specific high-level features.Fig.4 shows the detailed architecture.Specifically,the policy head outputs policy,which is the distribution of action probabilities under the current state.By taking actions on different scales,the probabilities will be dynamically adjusted.The value head estimates the value of the current state,reflecting how good the current state is and estimating the expected return:

    Fig.4 Architecture of the segmentation network.The confidence network is the same as the value branch of the segmentation network,but the parameters are not shared

    whereTis the terminal time step in the interaction process,andγis the discount factor.θvdenotes the parameters of the value head,and the gradient with respect toθvis

    The policy head’s goal is to maximize the expected return by selecting proper actions in states(t).We useθpto denote the parameters of the policy head,and the gradient forθpis denoted as

    Usually,the policy head is updated more slowly.

    3.3 Action confidence learning

    As we mentioned in Section 1,there will be some situations where the segmentation model misunderstands or ignores the hint information.To some extent,these samples (or regions) with the phenomenon of interactive misunderstanding are hard samples (or regions).Although these samples may account for a small portion of the dataset,they are critical for improving generalization and robustness.The most important thing is finding a professional easy-or-hard representer (Nie et al.,2019) to identify them during the interaction.Focal loss (Lin TY et al.,2017) evaluates the easy-or-hard samples through a predicted probability.Nie et al.(2019) applied adversarial learning to train the easy-or-hard representer.Each representer has its advantages,but all evaluate easy-or-hard samples based on the final segmentation result.As such,these methods cannot be directly applied to interactive segmentation.For example,if the model predicts the category probability of a voxel to be 0.8,and then takes an actiona(t)i=-0.1,the next prediction will be 0.7 after interactionIfyi=1,the results are both correctly predicted becauseare both larger than 0.5,whereas for interactive segmentation,the probability is changing in the wrong direction.This change is what we call the interactive misunderstanding phenomenon,and the formal definition is shown in the following:

    Definition 1(Interactive misunderstanding) For a binary classification problem,the sign of the foreground labely=1 is denoted as positive,and the sign of the background labely=0 is denoted as negative.In an interactive medical image segmentation task (i.e.,a voxel-wise binary classification problem),for any voxeli,if the sign of the change of the segmentation probability outputted by the algorithm for two consecutive interaction steps,sign (Δ(p(i))),is not equal to sign (yi),then this phenomenon is defined as interactive misunderstanding.

    The confidence network learns the confidence of the given actions to avoid misunderstanding of hint information and to take accurate actions.We argue that confidential information can be used to regularize action choices and suggest more efficient interaction.The confidence network structure also uses P-Net as the backbone.The confidence network contains six 3D convolutional blocks.Each of the blocks has two convolution layers,and the size of the convolution kernel is fixed as 3×3 in all these convolution layers.The detailed architecture is also shown in Fig.4.

    The confidence network is trained using the previous state and action as input and a confidence map as output.The confidence network is optimized by minimizing the summation of binary cross-entropy loss over actions (shown in Eq.(6)) at each time stept.Here we useCto denote the confidence network,wCthe confidence network parameters,andLBCEthe binary cross-entropy loss:

    andg(t)indicates whether the direction of the action is consistent with the label.Fora ⊕b,the statement is true only if eithera>0 orb>0.

    One potential issue when training the confidence network is the imbalance of samples.Early in RL training,the error rate of actions outputted by the segmentation network is high,whereas most actions are correct when the network gradually converges.Inspired by discriminator training in GANs,this study introduces symmetric samples into Eq.(6) to speed up training.An obvious advantage is that it improves sample utilization efficiency because the confidence network can know what a“bad sample” is and obtain a corresponding“good sample.”

    3.4 Self-adaptive reward

    Essentially,no matter whether focal loss or adversarial learning is applied to train the segmentation model,the objective is to weigh these hard samples to prevent the model from being dominated by easy samples.However,the interactive segmentation task differs from those fully automatic segmentation tasks because the interactive segmentation model needs to cooperate with the user and understand the user’s hint information.The action to refine the segmentation result shows how the segmentation model understands hint information.Therefore,it is necessary to ensure that the hint information is correctly understood and that the correct action is taken.

    Specifically,the previously described actionconfidence learning can provide the segmentation model with a confidence map to alleviate interactive misunderstanding.Using this confidence map,easy-or-hard samples can be better recognized because the confidence values for these “hard regions”are lower than those for other regions.We formulate this action-aware learning as the self-adaptive reward function,r(t),which is shown in Eq.(8),to adapt this mechanism to MARL training:

    3.5 Interaction guide

    Another challenge for interactive image segmentation is that users usually need to decide where to interact in many slices,which is very timeconsuming,especially for 3D images.Our framework provides users with an interaction guide mechanism to save user interaction time.After refinement,our framework will suggest some possible areas for users to segment during the interaction.Specifically,our framework will filter out those areas with low action confidence and provide them to users (Fig.5).First,the original 3D image will be segmented with super voxels;each super voxel can be regarded as a group of voxels that share common characteristics.We use a simple linear iterative clustering (SLIC) (Achanta et al.,2012) technique with spacing=[2,2,2],compactness=0.1 to generate super voxels,and the number of super voxels equals 100 at the beginning and gradually declines during the refinement iterations for training and testing.Second,the proposed algorithm will compute the mean action confidence in each super voxel and rank them in descending order.Finally,the top five super voxels will be marked and recommended to users.Users need to select the best interaction positions from these super voxels.

    Fig.5 Interaction guide mechanism: (a) probability map;(b) action map;(c) confidence map;(d) advice region;(e) interaction point.The areas surrounded by the green lines are advice regions provided to users,and the red point in (e) is the real hint information selected from advice regions.The brighter the color (closer to yellow),the larger the positive value;conversely,the darker the color (closer to black),the smaller the negative value.References to color refer to the online version of this figure

    3.6 Simulated label generation

    In medical imaging,there are much more unlabeled data than labeled data due to difficulties in labeling medical images.To address this lack of annotations,the proposed algorithm leverages the action confidence not only to improve the utilization efficiency of hint information but also to generate a simulated label (Fig.6) for unlabeled data.We define the simulated label as follows:

    Fig.6 Illustration of simulated label generation.The simulated label generation mechanism uses the confidence map and the action map to generate the simulated label.The confidence map is used to calibrate the action,and the direction of the calibrated action is the simulated label of each voxel

    whereM=I(max (c(t),1 -c(t)) >δ),andis the advantage (defined in Mnih et al.(2016)) at time steptof takinga(t)in the condition of states(t),indicating the actual accumulated reward without being affected by the state and reducing the gradient variance.A mask,M,is used to constrain the training of unlabeled data.The backward gradients of unlabeled data occur only when the action confidence exceeds thresholdδ(which is gradually increased during the training process).Unlike traditional pseudo-label training,the supervised signal does not come from the segmentation network but from the confidence network.These filtered data with hint information are more valuable and provide more accurate supervised signals.Generally,the training processes with labeled data and simulated labeled data are carried out simultaneously.When using the simulated label generation mechanism,the pseudocode is as shown in Algorithm 1.

    4 Experiments and results

    4.1 Datasets and implementation details

    To comprehensively evaluate our proposed method,we apply our algorithm to four 3D medical image datasets.All datasets are divided into two parts:Dtrain/Dtest.The details of these datasets are as follows: (1) BraTS2015: Brain Tumor Segmentation Challenge 2015(Menze et al.,2015) contains 274(234/40) multi-parametric MRI images (Flair,T1,T1C,T2) from brain tumor patients.In our task,we use only the Flair image and segment the whole brain tumor.(2) BraTS2020: Brain Tumor Segmentation Challenge 2020 (Menze et al.,2015) contains 285(235/50) multi-parametric MRI images (Flair,T1,T1C,T2) from brain tumor patients.In our task,we use only the Flair image and segment the whole brain tumor.(3) MM-WHS: Multi-Modality Whole Heart Segmentation (Zhuang and Shen,2016) contains 24(20/4) multi-modality whole heart images covering the whole heart substructures.In our task,we choose to segment the left atrium blood cavity.(4) Medical Segmentation Decathlon: This is a generalizable 3D semantic segmentation dataset containing different organ segmentation tasks (Simpson et al.,2019).We use the spleen and liver datasets,which provide 61 (41/20) and 106 (96/10) CT images,respectively.

    We implement our method with PyTorch (Paszke et al.,2019).The demo video of MECCA is available at https://bit.ly/mecca-demo-video.The segmentation and confidence networks are both initialized using the Xavier (Glorot and Bengio,2010) method,and the learning rate is initialized to 1e-4.Other parameters are set as follows:T=5,γ=0.95,α=0.8,β=1.The maskMranges from 0.85 to 0.99 and increases by 0.000 25 at every epoch.Adam (Kingma and Ba,2015) is adopted as the optimizer.The original image is cropped by the bounding box based on the ground truth with a random extension in the range of 1-11 voxels.Each image is then resized and normalized to 55×55×30.The data are augmented by random flips and rotations.The proposed algorithm’s training time with one Nvidia 2080ti GPU varies from 5 to 13 h for different datasets.

    4.2 Experimental settings

    During the process of interaction,we adopt the edge points as the hint information for two reasons.On one hand,click operation saves more time than scribbles or other interactive methods.On the other hand,edge points can provide more information about object edges because the bottleneck of medical image segmentation is usually the inaccurate segmentation of object edges.Specifically,we provide each method with 45 points during the whole interaction process.At every step,users will click some edges that are not correctly predicted.We generate a 3D Gaussian (with a kernel size of eight voxels) centered on each of the edge points as the hint map to let networks receive hint information.Then the hint map will be inputted to the segmentation network as part of the state.We use the Dice score and the average symmetric surface distance (ASSD) to evaluate the segmentation results.According to these evaluation metrics,doctors can judge the patient’s condition:

    whereSpandSgdenote the prediction of an algorithm and the ground truth,respectively.

    whereSaandSbrepresent the sets of surface points of the segmentation results predicted by the algorithm and the ground truth,respectively.d(i,Sb) is the shortest Euclidean distance betweeniandSb.The Dice and ASSD in all tables are the average values of five algorithm test results.

    4.3 Comparisons with baselines

    We compare MECCA with four state-of-the-art interactive segmentation methods: DeepIGeoS (Wang et al.,2019),InterCNN (Bredell et al.,2018),IteR-MRL (Liao et al.,2020),and BS-IRIS (Ma et al.,2021).InterCNN is the multi-step version of DeepIGeoS.We also introduce a state-of-the-art medical image segmentation method,U-Net (Ronneberger et al.,2015),as a baseline.Table 1 shows the quantitative comparison of the seven segmentation methods on different datasets.For a fair comparison,except U-Net,all CNN-based methods adopt the same network structure (P-Net),which was proposed in Wang et al.(2019).In particular,U-Net and P-Net are methods without hint information.We can see that MECCA performs better than or is on par with the state-of-the-art methods on three datasets.We visualize the results in Fig.7,which shows that our method achieves more accurate results in edge segmentation.

    Table 1 Quantitative comparison of different methods

    Fig.7 Visualization of segmentation results by different methods.The green lines represent the boundaries of the ground truth,and the yellow lines represent the predicted boundaries.References to color refer to the online version of this figure

    To demonstrate that MECCA can take advantage of hint information more efficiently,we compare the relative improvement of different methods at each interaction step.Due to the different interaction processes of different methods,we set up a unified interaction process: all these methods have five interaction steps and receive 25 identical points at the first step to generate the initial segmentation result.After that,users will click five points at each step in the following four interaction steps (a total of 25+5×4=45 points),and these methods will iteratively refine previous segmentation results.The DeepIGeoS method does not model the interaction sequence,and simply combines current and previous hint information to refine the previous segmentation.The results are shown in Table 2 and Fig.8.

    Table 2 Dice scores at each interaction step by different methods

    Fig.8 The performance improvement of different interactive medical segmentation methods at different interaction steps.All these testing results were obtained on the BraTS2015 dataset

    The results show that MECCA performs better under the same amount of hint information and improves the Dice score more after the five steps.Compared with CNN-based methods,the main advantage of RL-based methods is that they can always maintain notable improvement.There are two reasons for this result.The first one is that RLbased methods model the whole interaction process to avoid interaction conflict.The second reason is the relative entropy-based reward which encourages the model to keep refining results.However,we should realize that the RL-based methods still cannot guarantee the high confidence of the corrective actions.As we can see in Fig.8,the performance of IteRMRL is not as good as those of other methods at the beginning,which is caused by numerous incorrect actions.However,after learning the confidence of actions and applying the self-adaptive reward to update the segmentation model,our proposed MECCA can perform well at each step and continue significantly refining the result.

    4.4 Comparison of different rewards

    One main contribution of our work is that the self-adaptive reward mechanism is further proposed based on the confidence map.This mechanism makes different actions have different feedback levels,so the segmentation network can identify wrong actions as much as possible.To measure the impact of different means of reward weighting on the segmentation result,we compare MECCA with the original IteR-MRL without a weighting reward,Focal-Reward IteR-MRL (Focal-IteR-MRL),and the one with a weighted reward in segmentation error regions (Err-IteR-MRL).Focal-IteR-MRL adopts the idea of focal loss (Lin TY et al.,2017):

    The reward function of Err-IteR-MRL is a linear scaling of the IteR-MRL reward:

    whereλiis a positive real number hyperparameter.The performance of different weighting rewards is shown in Fig.9.It can be seen that the actionbased reward weighting method is more suitable for interactive segmentation.Also,this study visualizes different masks that weigh the primary reward in Fig.10.As we can see,the focal mask pays more attention to regions with low prediction probabilities,and the error region mask contains all regions segmented incorrectly in the previous result.It is not enough for the focal mask to obtain more structured information by considering only the probability of prediction.As such,the performance of Focal-IteRMRL is erratic,even the poorest on the left atrium dataset.The performance of Err-IteR-MRL is more stable,but the improvement is not notable.This is because the weighting process of Err-IteR-MRL is based on the segmentation result while the refinement process of interactive image segmentation is based on actions.

    Fig.9 Average Dice scores of methods with different weighting rewards.The reward function of IteR-MRL is not weighted;the reward function of Focal-IteRMRL is weighted via Eq.(12),the reward function of Err-IteR-MRL is weighted in segmentation error regions,and the reward function of MECCA is weighted through action confidence

    Fig.10 Visualization of the original image (a),ground truth (b),action map (c),self-adaptive mask (d),error region mask (e),and focal mask (f).All masks are obtained at the first interaction step.The brighter the color (closer to yellow),the larger the positive value;conversely,the darker the color (closer to black),the smaller the negative value.References to color refer to the online version of this figure

    4.5 Weakly supervised segmentation

    As mentioned in Section 3.6,MECCA can reduce the dependence on voxel-level annotations of images using the simulated label generated from the action confidence map.This study validates MECCA on the BraTS2015 dataset by randomly selecting different proportions of samples as fully labeled data and using the rest of the training images as unlabeled data,which provide only hint information during the interaction.This study compares MECCA against P-Net (without hint),DeepIGeoS,and IteR-MRL (Liao et al.,2020).However,only MECCA uses unlabeled data,and the three baselines use only a fixed proportion of labeled data.This does not fully demonstrate the performance of MECCA on semi-supervised problems.For this reason,we introduce a state-of-the-art semi-supervised method UA-MT (Yu et al.,2019).Table 3 shows the results of different methods.MECCA achieves a Dice score of 87.14% with only 12.5% labeled data,and it performs better than the other methods with 25%labeled data.

    Table 3 Quantitative comparison between MECCA and other methods on the BraTS2015 dataset with different portions of labeled data

    Because UA-MT is not an interactive segmentation algorithm,its absolute performance is worse than those of interactive segmentation baselines but better than that of the non-interactive method PNet.In addition,it can be seen from the last column of Table 3 that UA-MT has the most negligible performance loss with different proportions of labeled data.After MECCA introduces the simulated label generation mechanism,the performance loss is similar to that of UA-MT.Interestingly,DeepIGeoS can maintain low performance loss without using a semi-supervised learning mechanism.Considering the performance loss of IteR-MRL,we can conclude that in the semi-supervised interactive segmentation task,the interactive misunderstanding phenomenon will exacerbate the performance loss caused by missing data.DeepIGeoS and MECCA alleviate the interactive misunderstanding phenomenon through hard constraints and self-adaptive confidence calibration,respectively,so both of them can achieve better results in semi-supervised settings.However,note that DeepIGeoS does not consider multi-step interactions or the relationship between consecutive interactions.It cannot fully use long-term interaction information,so there is a large gap with MECCA in absolute performance.

    We also evaluate MECCA with different numbers of interactions and percentages of labeled data.Based on the experimental results (Table 4),we conclude the following: (1) if there are enough labeled data,MECCA can achieve good results after a few interactions;(2) the number of interactions required by MECCA to achieve the same performance is roughly inversely proportional to the portion of labeled data;(3) although MECCA requires more interactions (about 2-3 times) when only part of the labeled data is available,it can approach the performance of the algorithm trained with all data being labeled in the end.

    Table 4 Dice scores with different numbers of interactions and percentages of labeled data

    4.6 Ablation studies

    Table 5 shows the impact of different mechanisms on MECCA.It is clear that the simulated label generation mechanism and self-adaptive reward mechanism significantly improve the algorithm’s performance.The simulated label generation mechanism can achieve a 1.70% and 2.26% gain over the initial algorithm without any mechanism for the BraTS2015 dataset and the Liver dataset in Medical Segmentation Decathlon,respectively.The main reason for this significant improvement is the utilization of these unlabeled data.

    Table 5 Ablation study of our proposed algorithm on BraTS2015 and Liver datasets with 25% labeled data

    Table 6 shows the comparison of interactive segmentation methods in computational and interaction time.The results show that MECCA takes only about half the time of other methods when performing an interaction.The main contribution to the reduction of interaction time is the interaction guide mechanism mentioned.Furthermore,to explore the efficiency of MECCA in authentic tasks,we design an interactive software platform.We ask oncologists to use it for natural interactive segmentation,and the results are shown in Table 7.It can be seen from the results that the MECCA algorithm also takes only about half the time of the baselines due to interaction guidance.In addition,as the segmentation results become more and more accurate,the interaction time required gradually decreases.

    Table 6 Comparison of computational and interaction time among different interactive segmentation methods

    Table 7 Quantitative comparison of interaction time among different interactive segmentation methods with different numbers of interactions on a realistic segmentation platform operated by experts

    5 Conclusions

    We present a novel action-based confidence learning method for interactive 3D image segmentation.Specifically,we propose a method for learning the confidence of actions that continuously refine the segmentation result during the interaction process,so that hint information can be used more effectively.Based on this process,a self-adaptive reward is proposed for the segmentation module,which can prevent the misunderstanding phenomenon during the interaction and help reduce the time cost of interaction by providing users with the advice regions with which they should interact next.In addition,the confidence map can replace the ground truth to generate feedback for the unlabeled samples for the segmentation module.These samples,without voxel-level annotations,can also be used to model training.By integrating these components,MECCA is demonstrated to improve medical segmentation efficiently using fewer annotated samples.

    Although our method can greatly enhance the utilization of short-and long-term interaction information with significantly fewer labeled samples,there are also certain limitations that we need to address.One of the primary challenges in medical image segmentation is the high dimension and resolution of medical images,leading to a requirement for numerous interactive clicks.Specifically,Wang et al.(2018) provided the actual use time collected from real physicians.Thus,there is a need for more efficient interactive methods,and Aljabri et al.(2022) introduced the currently available annotation tools for medical imaging.In future work,we may incorporate more efficient annotation methods to minimize the cost of physician interactions.

    Further,RL from human feedback (RLHF),which is used in ChatGPT (OpenAI,2022),is increasingly popular.This approach collects comparison data and trains a reward model,where the data are human rankings of the model’s outputs from best to worst.RLHF is another application that benefits from human feedback,and in future work,we may consider another interactive method in which radiographers rank the different outputs as reward signals.Another limitation in the field of medical image segmentation is the presence of multiple centers,where data are collected from different devices,resulting in varying distributions.With such non-independent and identically distributed (non-IID) data,the generalization ability of deep learning based models is limited (Li et al.,2021;Ye et al.,2022).To overcome this,our proposed MECCA method may require additional design;e.g.,Aljabri et al.(2022) introduced a novel robust weakly supervised learning paradigm.

    Finally,as technology continues to advance,changes in patient distribution and the introduction of new medical equipment may necessitate the development of online learning methods to ensure that the algorithm remains robust.Popular methods such as iCaRL (Rebuffiet al.,2017),GDumb (Prabhu et al.,2020),and iCaRL (Rebuffiet al.,2017) maintain a buffer to store the most important samples for the model.Specifically,in the RL area,we can borrow the ideas and theoretical justifications from lifelong RL,in which setting agents must solve a series of related tasks drawn from a task distribution rather than a single,isolated task (Abel et al.,2018).Abel et al.(2018) explored which knowledge should be transferred in lifelong RL through two simple families of results.Xie et al.(2020) leveraged latent variable models to learn a representation of the environment from experience and performed off-policy RL with this representation.

    Contributors

    Chuyun SHEN,Wenhao LI,and Qishen XU designed the research and conducted the experiments.Bin HU,Fengping ZHU,and Yuxin LI ensured the validity of the experiments.Bo JIN,Haibin CAI,and Xiangfeng WANG offered support across various experimental aspects.Chuyun SHEN drafted the paper.All the authors revised and finalized the paper.

    Compliance with ethics guidelines

    Chuyun SHEN,Wenhao LI,Qisen XU,Bin HU,Bo JIN,Haibin CAI,Fengping ZHU,Yuxin LI,and Xiangfeng WANG declare that they have no conflict of interest.

    Data availability

    The demo video is available at https://bit.ly/meccademo-video.The other data that support the findings of this study are available from the corresponding author upon reasonable request.

    List of supplementary materials

    1 More related works

    2 More visualizations

    3 Robustness of MECCA

    4 Comparison of baseline responses to the same user interaction

    Fig.S1 MECCA segmentation process

    Fig.S2 Qualitative segmentation results of MECCA for the BraTS2015 validation set

    Figs.S3-S5 Results of different methods’ responses to the same user interactions according to the same initial segmentation on different testing instances and different channels for the Liver dataset in Medical Segmentation Decathlon

    Table S1 Dice of our method which varies with the number of interactions under different cases

    Table S2 MECCA’s tolerance to inaccurate interaction points

    久久久精品欧美日韩精品| 国产精品亚洲av一区麻豆| 国产xxxxx性猛交| 亚洲色图av天堂| 亚洲av电影不卡..在线观看| а√天堂www在线а√下载| 国产欧美日韩精品亚洲av| 国产亚洲精品第一综合不卡| www日本在线高清视频| 中文字幕色久视频| av视频在线观看入口| 美女大奶头视频| 九色亚洲精品在线播放| 老司机在亚洲福利影院| 国产精品永久免费网站| 日韩大码丰满熟妇| 久久久久久大精品| 欧美日本亚洲视频在线播放| 欧美绝顶高潮抽搐喷水| 久久精品国产99精品国产亚洲性色 | 午夜免费激情av| 一区二区三区激情视频| 成人18禁在线播放| 亚洲精品久久国产高清桃花| 长腿黑丝高跟| 国产xxxxx性猛交| 国产精品免费视频内射| 91大片在线观看| 极品人妻少妇av视频| 久久久国产成人免费| 成在线人永久免费视频| 国产av在哪里看| 在线观看午夜福利视频| 日韩精品免费视频一区二区三区| 女同久久另类99精品国产91| 香蕉国产在线看| 视频在线观看一区二区三区| 两个人免费观看高清视频| 在线观看66精品国产| 国产亚洲精品一区二区www| 黄色丝袜av网址大全| 亚洲成a人片在线一区二区| 久久精品成人免费网站| 国产亚洲欧美在线一区二区| 精品国产美女av久久久久小说| 中出人妻视频一区二区| 这个男人来自地球电影免费观看| 午夜福利,免费看| 亚洲中文字幕日韩| 欧美日韩瑟瑟在线播放| 亚洲最大成人中文| 久久午夜综合久久蜜桃| 美女大奶头视频| 欧美成人免费av一区二区三区| 久久国产精品男人的天堂亚洲| 精品欧美一区二区三区在线| 国产免费男女视频| 后天国语完整版免费观看| 一a级毛片在线观看| 亚洲精品一卡2卡三卡4卡5卡| 99久久综合精品五月天人人| 操出白浆在线播放| 久久久久九九精品影院| 国产一区二区三区在线臀色熟女| 国产高清有码在线观看视频 | 99国产综合亚洲精品| 男女做爰动态图高潮gif福利片 | 国产蜜桃级精品一区二区三区| 精品人妻1区二区| 激情视频va一区二区三区| 波多野结衣巨乳人妻| 九色国产91popny在线| 女人精品久久久久毛片| 99精品久久久久人妻精品| 国产精品久久久久久人妻精品电影| 色播亚洲综合网| 亚洲色图 男人天堂 中文字幕| 欧美一级毛片孕妇| 免费观看精品视频网站| 婷婷丁香在线五月| 国产精华一区二区三区| 免费观看人在逋| 亚洲狠狠婷婷综合久久图片| 国产亚洲欧美在线一区二区| 午夜福利,免费看| 欧美黑人欧美精品刺激| 午夜福利免费观看在线| 长腿黑丝高跟| 免费在线观看完整版高清| 国产真人三级小视频在线观看| 日韩欧美国产一区二区入口| 国产亚洲精品久久久久久毛片| 亚洲自偷自拍图片 自拍| 国产av一区在线观看免费| 制服诱惑二区| 国产精品av久久久久免费| 国产蜜桃级精品一区二区三区| 波多野结衣av一区二区av| 欧美成人午夜精品| √禁漫天堂资源中文www| 禁无遮挡网站| 国内毛片毛片毛片毛片毛片| 此物有八面人人有两片| 久久午夜综合久久蜜桃| 国产亚洲av嫩草精品影院| 亚洲国产欧美一区二区综合| 亚洲国产精品成人综合色| 亚洲视频免费观看视频| 999久久久国产精品视频| 亚洲欧美日韩另类电影网站| 怎么达到女性高潮| 美女 人体艺术 gogo| 99久久国产精品久久久| 在线观看免费日韩欧美大片| 欧美日韩中文字幕国产精品一区二区三区 | 亚洲成a人片在线一区二区| 亚洲av电影在线进入| 亚洲国产看品久久| 午夜精品国产一区二区电影| 777久久人妻少妇嫩草av网站| 精品第一国产精品| 国产欧美日韩综合在线一区二区| 大型av网站在线播放| 嫩草影院精品99| 黄色视频不卡| 日韩三级视频一区二区三区| 精品乱码久久久久久99久播| 午夜福利一区二区在线看| 啦啦啦免费观看视频1| 国产精品,欧美在线| 黄色视频,在线免费观看| 国产蜜桃级精品一区二区三区| 日韩精品免费视频一区二区三区| 亚洲国产精品999在线| 午夜福利18| 久久久久久人人人人人| 国产欧美日韩一区二区三区在线| 丝袜在线中文字幕| 男人舔女人的私密视频| 12—13女人毛片做爰片一| 激情视频va一区二区三区| 成人特级黄色片久久久久久久| 亚洲电影在线观看av| 精品国产亚洲在线| 亚洲三区欧美一区| 麻豆成人av在线观看| 亚洲精品中文字幕在线视频| 亚洲国产欧美网| 高清在线国产一区| 三级毛片av免费| 日韩有码中文字幕| 色av中文字幕| 制服人妻中文乱码| 国产99久久九九免费精品| 国产精品亚洲美女久久久| 人人妻,人人澡人人爽秒播| 美女国产高潮福利片在线看| www.精华液| 可以免费在线观看a视频的电影网站| 午夜精品在线福利| 欧美日韩乱码在线| 91成年电影在线观看| 90打野战视频偷拍视频| 亚洲,欧美精品.| 日韩 欧美 亚洲 中文字幕| 乱人伦中国视频| 日本免费一区二区三区高清不卡 | 91麻豆精品激情在线观看国产| 在线视频色国产色| 久久人人精品亚洲av| 国产三级在线视频| 999久久久国产精品视频| 丝袜在线中文字幕| 国产欧美日韩一区二区三区在线| www.自偷自拍.com| www.自偷自拍.com| 又紧又爽又黄一区二区| 国产精品永久免费网站| 久久天躁狠狠躁夜夜2o2o| 精品乱码久久久久久99久播| 国产精品1区2区在线观看.| 99久久综合精品五月天人人| 精品久久久久久久毛片微露脸| 丝袜在线中文字幕| 精品国产乱子伦一区二区三区| 日日爽夜夜爽网站| 免费在线观看视频国产中文字幕亚洲| 91在线观看av| 午夜福利影视在线免费观看| 色老头精品视频在线观看| 九色亚洲精品在线播放| 国产真人三级小视频在线观看| 黄频高清免费视频| 一个人观看的视频www高清免费观看 | 午夜免费激情av| 一本久久中文字幕| 国产麻豆69| 啦啦啦观看免费观看视频高清 | av欧美777| 美女免费视频网站| 国产一区在线观看成人免费| 神马国产精品三级电影在线观看 | 真人一进一出gif抽搐免费| 国产精品爽爽va在线观看网站 | 50天的宝宝边吃奶边哭怎么回事| 女人被狂操c到高潮| 可以在线观看的亚洲视频| 久久久国产成人免费| 日韩有码中文字幕| 丁香六月欧美| 在线观看66精品国产| 日本免费一区二区三区高清不卡 | 麻豆国产av国片精品| 非洲黑人性xxxx精品又粗又长| 亚洲免费av在线视频| 亚洲avbb在线观看| 一级毛片精品| 又黄又爽又免费观看的视频| 国产一级毛片七仙女欲春2 | 国产一级毛片七仙女欲春2 | 亚洲成人久久性| 欧美+亚洲+日韩+国产| 亚洲精品国产精品久久久不卡| 国产精华一区二区三区| 成人三级做爰电影| 国内久久婷婷六月综合欲色啪| 最新美女视频免费是黄的| 9191精品国产免费久久| 黄色成人免费大全| 999久久久国产精品视频| 女人高潮潮喷娇喘18禁视频| 国产高清激情床上av| 国产精品日韩av在线免费观看 | 中文字幕久久专区| 好看av亚洲va欧美ⅴa在| 女同久久另类99精品国产91| 国产欧美日韩一区二区三| 男人的好看免费观看在线视频 | 激情视频va一区二区三区| 欧美+亚洲+日韩+国产| 巨乳人妻的诱惑在线观看| 天堂动漫精品| av视频在线观看入口| 国产熟女午夜一区二区三区| 中文字幕色久视频| 啦啦啦韩国在线观看视频| 成年版毛片免费区| 激情视频va一区二区三区| 女人高潮潮喷娇喘18禁视频| 久久婷婷人人爽人人干人人爱 | 级片在线观看| 咕卡用的链子| 老司机靠b影院| 99re在线观看精品视频| 久久香蕉国产精品| 亚洲精华国产精华精| 老司机午夜十八禁免费视频| √禁漫天堂资源中文www| 亚洲av成人一区二区三| 亚洲电影在线观看av| 欧美在线黄色| 精品久久久久久久毛片微露脸| www.www免费av| 久久久国产成人精品二区| 50天的宝宝边吃奶边哭怎么回事| 少妇熟女aⅴ在线视频| 悠悠久久av| 两个人免费观看高清视频| 丝袜人妻中文字幕| 人妻丰满熟妇av一区二区三区| 身体一侧抽搐| 精品一品国产午夜福利视频| 成人免费观看视频高清| 男人的好看免费观看在线视频 | 好男人电影高清在线观看| 欧美av亚洲av综合av国产av| 久久精品影院6| 国产精品 国内视频| 精品国产一区二区久久| 精品一区二区三区av网在线观看| 老熟妇乱子伦视频在线观看| 成人国产一区最新在线观看| 亚洲av美国av| 啪啪无遮挡十八禁网站| 无人区码免费观看不卡| 91大片在线观看| 亚洲精品中文字幕在线视频| 露出奶头的视频| 成人亚洲精品av一区二区| 精品国产国语对白av| 国产精品免费视频内射| 丁香欧美五月| 人人妻人人澡人人看| 色综合婷婷激情| 日韩 欧美 亚洲 中文字幕| 欧美乱妇无乱码| 日本三级黄在线观看| 久久精品亚洲精品国产色婷小说| 久久久久久免费高清国产稀缺| 美女大奶头视频| 亚洲男人天堂网一区| 亚洲精品中文字幕一二三四区| 美女 人体艺术 gogo| 不卡av一区二区三区| 男女做爰动态图高潮gif福利片 | 精品国产乱码久久久久久男人| 69精品国产乱码久久久| 窝窝影院91人妻| 欧美日韩黄片免| 日韩欧美三级三区| 日本a在线网址| 别揉我奶头~嗯~啊~动态视频| 国产精品永久免费网站| 女性被躁到高潮视频| 日韩欧美国产一区二区入口| 一个人免费在线观看的高清视频| 91麻豆av在线| 精品卡一卡二卡四卡免费| 99国产精品99久久久久| 日本三级黄在线观看| 男女之事视频高清在线观看| 纯流量卡能插随身wifi吗| 成熟少妇高潮喷水视频| 亚洲成av人片免费观看| 高清毛片免费观看视频网站| 成年女人毛片免费观看观看9| 超碰成人久久| 国产精品久久久人人做人人爽| 国产精品美女特级片免费视频播放器 | 亚洲,欧美精品.| 人成视频在线观看免费观看| 中文字幕色久视频| 成人永久免费在线观看视频| 日韩免费av在线播放| 国产精品,欧美在线| 中亚洲国语对白在线视频| 桃色一区二区三区在线观看| 热re99久久国产66热| 少妇粗大呻吟视频| 欧美激情极品国产一区二区三区| 成人手机av| 久久影院123| 一区二区三区高清视频在线| 亚洲九九香蕉| xxx96com| 两性午夜刺激爽爽歪歪视频在线观看 | 亚洲精品在线观看二区| 亚洲成av片中文字幕在线观看| 桃红色精品国产亚洲av| 级片在线观看| 亚洲熟女毛片儿| 亚洲精品久久成人aⅴ小说| 电影成人av| 丰满人妻熟妇乱又伦精品不卡| 国产精品乱码一区二三区的特点 | 精品熟女少妇八av免费久了| 看黄色毛片网站| 欧美成人午夜精品| 91麻豆精品激情在线观看国产| 男女下面插进去视频免费观看| 欧美在线一区亚洲| 亚洲av电影不卡..在线观看| 一级毛片高清免费大全| 在线观看免费视频日本深夜| 精品国产美女av久久久久小说| 一边摸一边抽搐一进一小说| 成人三级做爰电影| 女性被躁到高潮视频| 国产高清激情床上av| 男女下面插进去视频免费观看| 婷婷丁香在线五月| 女性被躁到高潮视频| 久久精品aⅴ一区二区三区四区| 欧美成人性av电影在线观看| 欧美黄色淫秽网站| 国产精品亚洲美女久久久| 丁香欧美五月| 国产成人精品无人区| 99久久综合精品五月天人人| 欧美绝顶高潮抽搐喷水| 波多野结衣av一区二区av| 搡老妇女老女人老熟妇| 中出人妻视频一区二区| 久久精品亚洲熟妇少妇任你| 午夜福利欧美成人| 国产成人影院久久av| 精品久久久久久久久久免费视频| 亚洲 欧美 日韩 在线 免费| 他把我摸到了高潮在线观看| 青草久久国产| 50天的宝宝边吃奶边哭怎么回事| 悠悠久久av| 91成年电影在线观看| 久久久久久久久免费视频了| 国产精品久久久久久精品电影 | 国产单亲对白刺激| 黑人巨大精品欧美一区二区mp4| 亚洲第一青青草原| 如日韩欧美国产精品一区二区三区| 精品国产乱子伦一区二区三区| 国产成人欧美| 成在线人永久免费视频| 欧美一级毛片孕妇| 精品不卡国产一区二区三区| 亚洲av第一区精品v没综合| 国产精品 欧美亚洲| 欧美亚洲日本最大视频资源| 91成年电影在线观看| 99久久国产精品久久久| 男男h啪啪无遮挡| 亚洲欧美日韩无卡精品| 久久久久久久精品吃奶| 成人18禁高潮啪啪吃奶动态图| 久久久久久久久中文| 91字幕亚洲| av有码第一页| 久久伊人香网站| 久久人人97超碰香蕉20202| 亚洲欧美一区二区三区黑人| 成人国语在线视频| 一边摸一边抽搐一进一出视频| 午夜福利欧美成人| 日韩av在线大香蕉| 亚洲在线自拍视频| 波多野结衣高清无吗| 狠狠狠狠99中文字幕| 级片在线观看| 国产精品久久久久久亚洲av鲁大| 夜夜爽天天搞| 国产激情欧美一区二区| 亚洲成人国产一区在线观看| 一进一出抽搐动态| 美女国产高潮福利片在线看| 日韩精品免费视频一区二区三区| 91国产中文字幕| 精品高清国产在线一区| 欧美av亚洲av综合av国产av| 色精品久久人妻99蜜桃| 免费观看精品视频网站| 天天躁狠狠躁夜夜躁狠狠躁| 久久性视频一级片| 757午夜福利合集在线观看| 国产xxxxx性猛交| 亚洲第一青青草原| 欧美色欧美亚洲另类二区 | 成人亚洲精品一区在线观看| 熟女少妇亚洲综合色aaa.| 国产精品综合久久久久久久免费 | 久久久国产成人精品二区| 夜夜看夜夜爽夜夜摸| 久久久久久久久久久久大奶| 欧美性长视频在线观看| 欧美日韩乱码在线| 一级毛片精品| 久久精品国产亚洲av香蕉五月| 三级毛片av免费| 国产亚洲av高清不卡| 中文字幕另类日韩欧美亚洲嫩草| 黄色a级毛片大全视频| 午夜福利欧美成人| 久99久视频精品免费| 国语自产精品视频在线第100页| 欧美日韩中文字幕国产精品一区二区三区 | 女警被强在线播放| 美女国产高潮福利片在线看| 黄色视频不卡| 国产人伦9x9x在线观看| 久久人人精品亚洲av| 好男人在线观看高清免费视频 | 两性午夜刺激爽爽歪歪视频在线观看 | 欧美丝袜亚洲另类 | 国产av一区在线观看免费| 嫩草影视91久久| 1024视频免费在线观看| 91在线观看av| 国产一卡二卡三卡精品| 中出人妻视频一区二区| 大型黄色视频在线免费观看| 1024视频免费在线观看| 美国免费a级毛片| 91精品三级在线观看| 精品久久蜜臀av无| 天天躁夜夜躁狠狠躁躁| av网站免费在线观看视频| 黄色毛片三级朝国网站| 一二三四社区在线视频社区8| 校园春色视频在线观看| 啦啦啦免费观看视频1| 成年人黄色毛片网站| 一边摸一边抽搐一进一出视频| 成人亚洲精品一区在线观看| 91大片在线观看| 国产又爽黄色视频| 国产成人一区二区三区免费视频网站| 麻豆成人av在线观看| 欧美日韩乱码在线| 国产1区2区3区精品| 久久中文看片网| 国产亚洲欧美98| 老汉色∧v一级毛片| 色播亚洲综合网| 91在线观看av| 亚洲精品国产精品久久久不卡| 成人18禁在线播放| 国产精品久久久av美女十八| 日本一区二区免费在线视频| 亚洲成人免费电影在线观看| 狠狠狠狠99中文字幕| 亚洲午夜理论影院| 欧美黄色片欧美黄色片| 91麻豆av在线| 欧美日本亚洲视频在线播放| 国产亚洲欧美98| 搡老岳熟女国产| 亚洲男人的天堂狠狠| 午夜免费观看网址| 看免费av毛片| 欧美日韩中文字幕国产精品一区二区三区 | 精品高清国产在线一区| 一个人免费在线观看的高清视频| 亚洲av成人一区二区三| 狠狠狠狠99中文字幕| 欧美日韩精品网址| 夜夜爽天天搞| 又黄又爽又免费观看的视频| 免费观看人在逋| 成人欧美大片| 一区二区日韩欧美中文字幕| 久久久国产成人精品二区| 精品国产亚洲在线| 久久精品91无色码中文字幕| 国语自产精品视频在线第100页| 亚洲国产欧美一区二区综合| 日韩国内少妇激情av| 国产成人精品在线电影| 欧美日本视频| 满18在线观看网站| 亚洲国产精品合色在线| 电影成人av| 女同久久另类99精品国产91| 亚洲五月婷婷丁香| 97人妻天天添夜夜摸| 午夜亚洲福利在线播放| 国产激情欧美一区二区| 亚洲,欧美精品.| 亚洲精品国产色婷婷电影| 久久久国产欧美日韩av| 亚洲人成伊人成综合网2020| 亚洲 国产 在线| 搡老岳熟女国产| 性欧美人与动物交配| 每晚都被弄得嗷嗷叫到高潮| 青草久久国产| 国产亚洲av高清不卡| 这个男人来自地球电影免费观看| 精品一区二区三区视频在线观看免费| 两性夫妻黄色片| 好男人电影高清在线观看| 中文字幕人妻熟女乱码| 亚洲欧美日韩另类电影网站| 午夜福利在线观看吧| 一级a爱视频在线免费观看| 免费观看人在逋| 免费一级毛片在线播放高清视频 | www.精华液| avwww免费| 亚洲avbb在线观看| 成人三级黄色视频| 久久精品国产99精品国产亚洲性色 | 在线观看免费视频日本深夜| 久久性视频一级片| 少妇粗大呻吟视频| 麻豆国产av国片精品| 久久婷婷成人综合色麻豆| 亚洲色图 男人天堂 中文字幕| 国产亚洲精品久久久久久毛片| 免费看十八禁软件| 黑人巨大精品欧美一区二区mp4| 精品欧美国产一区二区三| 亚洲国产欧美日韩在线播放| 97超级碰碰碰精品色视频在线观看| 国产亚洲精品第一综合不卡| 亚洲色图 男人天堂 中文字幕| 成人特级黄色片久久久久久久| 久久久久久久久久久久大奶| 欧美日韩瑟瑟在线播放| 免费搜索国产男女视频| 久久久久国产精品人妻aⅴ院| 久久人妻福利社区极品人妻图片| 亚洲国产日韩欧美精品在线观看 | 99精品在免费线老司机午夜| 五月玫瑰六月丁香| 日韩欧美国产在线观看| 久久久久久久久大av| 亚洲在线观看片| 亚洲内射少妇av| 91在线观看av| 狠狠狠狠99中文字幕| 亚洲成人免费电影在线观看| 亚洲人与动物交配视频| 日韩精品青青久久久久久| 在线看三级毛片| 欧美日本亚洲视频在线播放| 成年女人毛片免费观看观看9| 免费看a级黄色片| 久久精品国产99精品国产亚洲性色| 人人妻人人澡欧美一区二区| 无人区码免费观看不卡| 免费看a级黄色片| 国产精品久久久久久久久免| 午夜福利欧美成人| 内地一区二区视频在线| 香蕉av资源在线| 精品久久久噜噜| 欧美性猛交黑人性爽|