• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    View suggestion for interactive segmentation of indoor scenes

    2017-06-19 19:20:13ShengYangJieXuKangChenandHongboFu
    Computational Visual Media 2017年2期

    Sheng YangJie Xu,Kang Chen,and Hongbo Fu

    View suggestion for interactive segmentation of indoor scenes

    Sheng Yang1Jie Xu2,Kang Chen1,and Hongbo Fu3

    Point cloud segmentation is a fundamental problem.Due to the complexity ofreal-world scenes and the limitations of3D scanners,interactive segmentation is currently the only way to cope with all kinds of point clouds.However,interactively segmenting complex and large-scale scenes is very time-consuming. In this paper,we present a novel interactive system for segmenting point cloud scenes.Our system automatically suggests a series of camera views,in which users can conveniently specify segmentation guidance.In this way,users may focus on specifying segmentation hints instead of manually searching for desirable views of unsegmented objects,thus significantly reducing user effort.To achieve this,we introduce a novelview preference model,which is based on a set of dedicated view attributes,with weights learned from a user study.We also introduce support relations for both graph-cut-based segmentation and finding similar objects.Our experiments show that our segmentation technique helps users quickly segment various types of scenes,outperforming alternative methods.

    point cloud segmentation;view suggestion; interactive segmentation

    1 Introduction

    With the prevalence of consumer-grade depth sensors(e.g.,Microsoft Kinect),scanning our living environments is becoming easier.However,the resulting 3D point clouds are often noisy, incomplete,and distorted,posing various challenges to traditional point cloud processing algorithms. Thus,in recent years,growing attention has been paid to low-quality point cloud processing problems. Amongst them,semantic segmentation,which aims to provide a decomposition of a 3D point cloud into semantically meaningful objects,is one of the most fundamental problems,and is important for many subsequent tasks such as object detection[1],object recognition[2],scene understanding[3],etc.

    Semantic segmentation of 3D point clouds has been extensively studied,resulting in various techniques,based for instance on region growing[4, 5],graph-cut[6–8],learning[9–11],etc.Most of those approaches attempted to achieve semantic segmentation with little or even no user intervention. However,due to the complexity of real-world scenes and the limitations of 3D scanners,manual intervention is often inevitable[12].

    Previous interactive segmentation work(e.g.,Refs. [13,14])typically focuses on improving segmentation results given the same amount of user input (e.g.,provided by a commonly used stroke-based interface).We observed that when interactively segmenting scenes at a moderate or large scale, finding appropriate views to provide segmentation hints is very time-consuming.For example,for a scene with multiple rooms containing objects of various types,shapes,and sizes,objects can easily occlude each other,requiring careful selection of viewpoints for interactive segmentation.In addition,due to the discrete nature of point clouds, the distances between viewpoint and objects need to be carefully chosen to ensure the desired point density and that contextual information is in view.

    Fig.1 Given an input scene represented as a 3D point cloud(left),our system automatically suggests a series of reasonable views(middle) for easily inputting segmentation hints for semantic segmentation of the entire scene(right).

    Based on these observations,we present a new interactive system for segmenting cluttered point clouds of large real-world scenes.Our system is able to automatically suggest a series of camera views,in which users can conveniently specify segmentation guidance,i.e.,2D strokes in our case.To reduce user effort,we aim to optimize the suggestions, i.e.,to provide views that both contain plenty of undetermined objects and can clearly display them. To achieve this,we introduce a novelview preference model,which is based on a set of dedicated view attributes,whose weights are learned from a user study.Given a new scene,our system uses the learned view preference model to find the next best views one by one.In this way,users may focus on specifying segmentation hints,instead of manually searching for desired views for segmentation of unsegmented objects.

    To further reduce user effort in interactive segmentation,we incorporate support relations in a graph-cut-based segmentation framework,to find similar objects for segmentation propagation. We have compared the performance of interactive point cloud segmentation with and without view suggestion,and interactive segmentation of RGB-D images.The experiments show that our segmentation technique with view suggestion helps users quickly segment various types of scenes,and outperforms alternative methods.

    2 Related work

    2.1 Point cloud segmentation

    Semantic segmentation of 3D point clouds or RGBD images has long been an active research topic in the communities of computer graphics,vision,and robotics(see an insightful survey by Nguyen and Le[12]).Below we discuss the most relevant works, and categorize them into supervised,unsupervised, and interactive techniques.

    With the growing availability of free 3D datasets (e.g.,theNYU Depth Dataset[11,15],theSUN3Ddataset[16],and ShapeNet[17]),supervisedlearning-based segmentation algorithms typically exploit high-level semantic information from labeled datasets,and use the learned knowledge to help detect,recognize,and thus semantically separate object regions from backgrounds.Training data for supervised methods mainly come from two sources: CAD models and RGB-D images.High-quality CAD models are ideal training data as they provide full 3D a priori knowledge about geometric shapes[10, 18]and even contextual relationships[19],but the number and diversity of high-quality digital models are far from enough to cover everything;creating these models is expensive.Since RGB-D images are much easier to acquire,numerous methods(e.g., Refs.[1,3,20,21])learn various features from labeled RGB-D images.However,despite ease of acquisition,the labeling of RGB-D images is still labor intensive.Our system can be used as a convenient and robust segmentation tool to help produce high-quality training data.

    Unsupervised semantic segmentation methods often rely on regular patterns(i.e.,symmetry and repetition)observed from input data itself,and work extremely effectively on outdoor building facades[22,23].However,interior scenes present much more complex structures.To simplify the problem,previous works(e.g.,Refs.[9,24])focus on large-scale working environments containing limited types of objects(e.g.,office desk,office chair, monitor),each repeated many times.However,realworld scenes often contain unique objects.Thus, manual intervention is often still required to refine segmentation results produced by such automatic methods.

    Since automatic segmentation methods are far from perfect,in practice interactive methods are more frequently used for segmenting both indoorscenes[13,25]and outdoor scenes[14].The most commonly used interactive scheme is to let users specify representative foreground and background regions(typically via a stroke-based interface),which are then used to construct a probabilistic inference model(using,e.g.,conditional random fields)and optimized using graph-cut[7,8],or simply used as seeds for region growing[4,5,26].However,such an interactive scheme is designed for segmenting an individual image frame,and thus requires carefully selected views of 3D scene contents for projection to screen space.While the view selection process is time-consuming,especially for large-scale and complex scenes,it has gained little attention.Thus, our work is largely complementary to existing interactive and automatic segmentation techniques.

    2.2 Camera control

    Finding feasible views to display 3Dgraphics on a 2D screen is a fundamental problem,which in computer graphics is generally referred to as the camera control problem[27,28].According to the camera mode and the scale of the 3D contents,existing solutions to this problem can be divided into two categories:fl yaroundandwalk-through.We adapt these existing ideas to a new context for interactive point cloud segmentation,and tackle unique challenges such as the handling of cluttered point clouds,and dynamic selection of a series of good views for easy labeling of objects.

    Afl y-aroundcamera allows complete contents rendered to the screen,and is often used when displaying single objects or small-scale scenes.The core problem is best view selection,which aims to automatically select the most representative view of a 3D model.Various low-level view attributes(e.g., projected area,viewpoint entropy,silhouette length, and depth distribution)have been proposed[29,30] as a basis for solving this problem.The state of the art is probably the work by Secord et al.[29],which learns how to combine low-level view attributes based on an extensive user study ofhuman preferences.Our work differs from the existing best view selection techniques in terms of both inputs(cluttered point clouds versus clean surface or volume models)and outputs(a series of viewpoints looking at different parts of a scene versus a set of independent good views looking at the same target).

    Walk-throughcamera mode is thus more relevant to our problem;it is often used to navigate within large-scale virtual scenes[31,32].This problem essentially comprises two sub-problems:viewpoint selection[30]and path planning[33,34].Unlike the criteria used for fly-around cameras,measuring the quality of viewpoints within a scene is mainly based onviewpoint entropy[30].Path planning is needed to avoid penetrating objects or walls and to present smooth scene roaming.In contrast, our problem demands essentially discrete views for labeling,although smooth transition is weakly considered.Again,existing walk-through methods often take clean scenes as input.Additionally,their extracted walk-through paths can be pre-generated, while in our case the selection of the next best view depends on the segmentation progress and needs to be determined on the fly.We have also found that top views are very usefulfor our application,but they are seldom used for the walk-through applications.

    3 System overview

    The main contribution of our work is automatic suggestion of good views for easy labeling of objects in indoor scenes represented as 3D point clouds.We implement view suggestion in an interactive segmentation system,which takes an unsegmented scan of an indoor scene as input. However,our interactive system can be helpful for interactively refining automatically generated segmentation results.

    As shown in Fig.2,our system contains three components:preprocessing,view suggestion,and interactive segmentation.In the preprocessing step (Section 4)the system automatically aligns an input scene with Manhattan directions,extracts storeys(levels),and clusters the points into patches which should be treated as indivisible units in segmentation.

    Afterwards our system automatically finds candidate viewpoints and sample views for presenting as suggestions(Section 5.1).To evaluate the quality ofa view,we introduce a view preference model(Section 5.2),which involves severalattributes such as point density,projected area,and viewpoint entropy.The weights of these attributes are learned by conducting a user study of human preferences for views of scenes.At runtime,each view is evaluated using the learned view preference model,withsegmentation status and smooth transitions taken into account.The best view is then suggested to the user for interactive segmentation.The user may reject the current suggestion,and our system then updates the strategy for suggesting another view.

    Fig.2 System overview.

    Given a view,the user provides segmentation hints on unsegmented,over-segmented,or undersegmented objects that are in view,using a classic stroke-based interface.The patches corresponding to these strokes are used as seeds to trigger a graph-cut-based segmentation optimization.We extract support relations between patches in the preprocessing step,and incorporate them into the segmentation optimization(Section 6.1).To further reduce the amount of user intervention,we find similar objects to already-segmented objects for segmentation propagation(Section 6.2).The above steps of automatic view suggestion and interactive segmentation ofscenes in the selected suggested view are repeated until all objects have been labeled. For multi-storey cases,our system provides view suggestions storey by storey.

    4 Preprocessing

    In this section we briefly introduce our preprocessing step,which generates essential information for later use.Our system takes as input a 3D point cloud of an indoor scene,which can be acquired using different types ofscanning devices such as LiDAR or Microsoft Kinect.For example,for RGB-D streams, they can be registered to form a point cloud using KinectFusion or its variants[35,36].Like many other point cloud processing pipelines[37],we downsample the data,estimate point normals, and transform the input scene into Manhattan coordinates[38]with thez-axis pointing upwards. Then we extract horizontal planes(by RANSAC) as floors and an optional ceiling.As illustrated in Fig.3,we decompose a multiple-storey building into individual storeys,with the ceiling of each storey removed to achieve top views of rooms.

    4.1 Constructing representative patches

    Over-segmentation has often been used in existing works for preprocessing point clouds[7,14,24].We are interested in over-segmenting an input scene into semantic patches,instead of patches with similar sizes but with no semantics[7].Such patches not only help reduce computational costs but also serve as integralsemantic units to analyze the importance of views and support relations.

    Fig.3 Decomposition of one-storey(left)and two-storey(right) buildings into individual storeys,with the removal of ceilings(if any).

    Our solution is an extension of the region growing approach by Ref.[24].Since the original approach does not consider any color information,it might not split objects with similar shape but different colors(e.g.,causing objects on the table in Fig.4(c) to disappear).In addition,their approach has been shown to be effective only on good-quality scenes acquired by LiDAR devices,and can lead to numerous tiny patches for point clouds of low quality(e.g.,those acquired by Kinect).To address these problems,we improve their approach by appropriately adding a new color condition and performing two rounds of growing,the first on the input point cloud and the second on the patches resulting from the first round.

    The first round of region growing is applied to the input points.Specifically,letG0=〈V0,E0〉denote a graph,withV0representing the input points andE0the edges generated byk-nearest neighborhoods (k=15 in our implementation).As points in different colors more likely belong to different objects or parts,as well as the basic normal and position conditions(with the default parameterst0andt1)for patch growing in Ref.[24],we add a color condition as follows:

    Fig.4 Patch construction results.(a)Downsampled point clouds from diff erent datasets.(b)By Ref.[7].(c)By Ref.[24].(d)Our result.

    whereiandjare two colors.[0,2π)and[0,1]denote normalized HSV color values fori.=(cossin)denotes the 2D coordinate values ofiin the Hue–Saturation color disk.Intuitively,we allow two colors to be merged when their Euclidean distance in the Hue–Saturation color disk and intensity value difference are both within a small thresholdt2(=0.05 in our implementation).

    This region growing process results in a set of merged patches.However,in our experiments we found that by adding the color condition,individual points might be isolated as individual patches, especially when the quality ofthe input point cloud is poor.To address this problem,we perform a second round of region growing on the merged patches from the first pass,but with relatively looser rules. LikeG0is constructed over points,G1=〈V1,E1〉is constructed over patches.For merging,we sort the patches in descending order of the number of contained points.For each patch,we find the normalof the best-fit plane and its centroid to check the respective normal and position conditions(the corresponding parameters are denotedt3,t4).The color condition is based on the color distribution and is more specifically defined using theχ-squared distance[39]between the HSV color histograms of two patchesPiandPjinV1:

    whereIidenotes the normalized HSV color histogram of points inPi.Specifically,H,S, and V channels are discretized into 16,16,and 8 bins respectively.We discretize V values into fewer bins to reduce the influence of different lighting conditions.The thresholdt5is set to 1.6 in our system.The above patch-based region growing procedure is repeated multiple times(5 times in our implementation).After each iteration,the thresholds are relaxed by 20%to merge small isolated patches into larger patches.In our system,we sett0=0.8,t1=0.05,t3=0.75,andt4=0.2 for allinput scenes.

    5 View suggestion

    In this section we first discuss view sampling, then perceptual assessment of views,and finally how to automatically suggest views for interactive segmentation.

    5.1 View sampling

    View sampling simplifies our problem,since it results in a discrete set of views for assessment.A view is basically determined by three vectors:the camera position,view direction,and up direction.

    Inspired by previous works on urban city reconstruction using both airborne and street-borne data[40],we consider two types of views:interior perspectivesandtop views.Top views provide a good summary of a scene or its parts,since objects often lie on flat surfaces(e.g.,tables,floors)and are thus easily separable from a top view.In contrast,interior perspectives enable a more detailed examination of a scene from a closer distance,and are useful for dealing with objects of small size or objects that are blocked in top views(e.g.,objects in bookshelves).Thus these two types of views are largely complementary to each other.Next we explain how we sample these two types of views.

    For interior perspectives,we fix the orientation of a view so that its up vector in image space is aligned with the upright orientation of a scene[29]. An interior perspective can then be characterized by (x,y,z,θ,φ),wherex,y,andzdefine the camera position,andθandφcontrol the view direction. A possible viewpoint should meet the following requirements:it must stay inside a room(above the floor,away from the walls,below the ceiling if any) and avoid hitting objects.Objects are approximately detected as columns using the column representation proposed by Fisher et al.[41].We then perform uniform sampling in the space of a storey excluding the space occupied by the columns to get possible camera positions(x,y,z),each ofwhich is associated with a view sphere parameterized by(θ,φ)(see Fig.2).Bothθandφare sampled every 15°,resulting in 266 views at each viewpoint.

    For every top view,its view direction is fixed and always points downward along the negativez-axis. Lean view directions are not used since objects in such resulting views would often be severely blocked by vertical walls.Hence,θis fixed atπfor all top views in our implementation.We thus use(x,y,z,φ) for mapping the camera position of a view and its up direction.We uncover the ceiling of a room or storey and uniformly sample the viewpoints and poses inside its bounding box.It can be easily seen that higher viewpoints produce wider-range views but fewer details.In the case of a multiple-storey building,we decompose it into multiple storeys (Section 4),and sample interior perspectives and top views for individual storeys.

    5.2 View preference model

    For simplicity here we assume no object has been segmented in a view;we discuss how to incorporate the segmentation status in Section 5.3.The simplified problem is similar to the problem of best view selection for a single object;the latter is often represented as a polygonal mesh.In contrast we need to deal with scenes of objects,represented as cluttered point clouds.We observe that good views have at least the following properties.First,objects in such views should be easy to recognize.Second, such views should contain as manyunsegmentedobjects as possible so that only a small set of views are needed to cover every object in a scene. These two properties are somewhat conflicting.For example,a bird’s eye view of a scene might include many objects in the view,but individual objects might not be easily recognizable.Thus there is no single definition to describe whether a given view is good or not.

    We explored various attributes to describe the quality of a view from different perspectives;some are based on previous work while others were designed by us.A key challenge is to evaluate the impact of each attribute when combining them together into a unified view preference model.This is achieved by learning from view preferences ofhuman viewers in a user study.Below we first describe the important attributes indicated by the user study and then give the details of the study itself.

    In a preprocessing step,we pre-compute asupplementary meshfrom the point cloud using a greedy projection triangulation algorithm[42]. This supplementary mesh greatly facilitates visibility checking.In addition,observing that people tend to focus more on objects located at the center ofa view, we add weights to pixels in thesupplementary meshview:each pixelpicorresponds to a projected point; its weight is defined aswi=1,wherediis the normalized Euclidean distance betweenpiand the center of the view,whileλwis a small factor to help emphasize the central areas(λw=0.2 in our system).

    Our attributes are:

    Apa:projected area.This is a rather basic attribute and has been proved effective in previous works like Ref.[29].LetP={Pn}denote the set ofvisiblepatches in a view.Visibility is checked using the pre-computed supplementary mesh.Apa, the sum of the projected areas of visible patches in view,is then defined as

    whereSnis the projected area of a visible patch, where each pixel is weighted by the focus-attention weightwi.

    Ave:viewpoint entropy.To estimate the richness ofa view,we follow V′azquez et al.[30],and calculate view entropy on the supplementary mesh.

    Apd:point density.Due to the discrete nature of our point clouds,the distance between viewpoint and objects must be carefully set to provide the desired point density and contextual information in the view.We perform such a view-related density measurement by comparing the difference of views between the point cloud and the supplementary mesh.Letbe a pixel of the point cloud view, andbe the weighted areas that the pixels in the point cloud view and the mesh view belong to,for the same patchi.Apdrepresenting the average density of patches in view is then defined as

    where?(x)=1/(1+e?λ?·(x?t6))is a sigmoid function for evaluating the density of one patch. Intuitively,ifS′n/Snexceeds a thresholdt6,the density is acceptable and allows users to recognize objects.In our system,we setλ?=16 andt6=0.3.

    Besides the attributes listed above,we also explored additional attributes considering aesthetic perception,including:depth information(Add,Adc),object layout(Aop,Aod,Aov),andradial patches(Arp).Please refer to the Electronic Supplementary Material(ESM)for detailed formulations of these attributes.However,according to the user study, they are relatively dispensable.

    User study.In order to analyze the importance of these attributes to our view preference model, we conducted a user study.We chose three scenes captured by different types ofdevices(Kinect V2 and LiDAR)to alleviate the influence of diff erent input qualities.For each scene,we randomly sampled 60 views(using the sampling approach described in Section 5.1),including both top views(5%)and interior perspectives(95%).16 participants with basic knowledge ofcomputer graphics were recruited to manually evaluate the quality of each view.To ensure a proper comparative evaluation,each time, each participant was given a group ofviews from the same scene(6 views per group,randomly selected from 60 views)and was asked to rate each view on a scale from 1(worst)to 5(best).They were also told that views containing easily recognizable and multiple objects without severe occlusion should be given high scores.Representative views are shown in Fig.5.Please refer to the ESM for other scenes and results.

    Learning weights of attributes.After collecting the scores for each view,we studied the importance of attributes.We normalized the observed attributes and adopted lasso regression[43] to determine the weightsβiof the attributes:it isable to perform variable selection in order to enhance the accuracy of the predictions by the statistical model it produces.The learned model by lasso regression is simply formulated as

    Fig.5 Representative views from the user study.(a)High-score top view.(b,c)High-score interior views.(d)Low-score interior view(the tree seriously blocks the objects behind).(e,f)Low-score interior views.

    Lasso regression shows that among all the attributes included in the regression testing,view entropyandpoint densityplayed the most significant roles,while other attributes were dispensable.Thus, we discarded these other attributes and performed lasso regression again.The final weights provided forview entropyandpoint densitywere 2.11 and 3.05,respectively.Detailed information about the regression process can be found in the ESM.

    5.3 View suggestion

    Given an input scene,we first sample views(see Section 5.1)in the preprocessing step.At runtime, we use the learned view preference model(Eq.(5)) to automatically suggest views.Specifically,we generate a set of candidate views,denoted byV,by picking views withApa>0.6.The view inVwith the highest scoreis suggested as the first view.

    Given a suggested view,the user may perform interactive segmentation(see Section 6)and then request another view.To avoid repetitive suggestions,each time a view is suggested,we lower its score by a scale factor 0.7.We also allow users to reject a current view.In this case our system will suggest new views.We suggest new views according to the following principles:

    ?View suggestions should respect the current segmentation status.Views with many alreadysegmented objects should have lower priority.

    ?A smooth transition between the current and the new views is preferred.

    ?View suggestion should take into account user rejection of suggested views.

    To satisfy the first guideline,we update the score of each candidate view according to the current segmentation status.More specifically,we replaceAvewith the residual entropywhich is calculated by removing the terms corresponding to already-labeled patches inAve,since such labeled patches provide little information.In contrast,Apdremains unchanged because intuitively this term is used to clearly display point clouds to the screen, and thus is somewhat independent of labeling.

    To encourage a smooth transition fromVntoVn+1, we pickVn+1according to the following rule:

    whereis the online score of viewVx,andD(Ki,Kn)is the horizontal Euclidean distance between the current camera positionKnand the candidate positionKi.We use an exponentialdecay ofpreference to favor nearby rather than distant view candidates.

    If a user rejects a suggested view,it generally means that unlabeled patches in this view are less important,and thus,their priority should be lowered when assessing subsequent suggestions.To do so,we lower the residual entropyAveof unlabeled patches in rejected views by a scale factor 0.8.

    For multi-storey cases,our system begins with the highest storey and suggests views related to the current storey until the labeling progress of the current storey surpasses 90%.Once surpassed,our system will suggest the best view for the next lower storey and continue.

    6 Interactive segmentation

    We adopt a stroke-based interface similar to that in Ref.[14]for interactive segmentation.Given a view,users draw different strokes on unsegmented regions to indicate different objects(see Fig.6(left)). Graph-cut-based optimization[44]is then used to achieve the desired segmentation results.Our work contributes to this paradigm by introducing support relations into graph-cut,and finding similar objects to further reduce the amount of user intervention.

    6.1 Support relations for graph-cut

    Fig.6 Graph-cut without(middle)and with(right)support relations.Left:input strokes on unsegmented objects in a given view.

    Previous graph-cut-based point cloud segmentation techniques(e.g.,Refs.[7,14])mainly focus on theuse of low-level geometric features(e.g.,position, normal,color)to propagate the segmentation information from the user-specified seeds to the rest of the scene.Since man-made objects often exhibit box-like shapes,due to the sharp change in normal between different faces of such objects, multiple strokes(one for each face)may be needed to segment a single object(e.g.,a desk or monitor in Fig.6(middle)).

    To reduce the amount of user intervention, we introducesupport relationsas an additional measurement of distance between patches in the graph-cut formulation.This helps the reduction of both the number of user-specified strokes and the number of needed views.For example,with our support relations considered(see Fig.6(right)),the user only needs to input strokes on the top faces of a desk and monitor in a top view to segment the objects.Without,the user would need to switch to different views to input further strokes.

    Support relations are common in indoor scenes because of the influence of gravity.There are two levels of support relations:support between objects and support between primitive shapes which constitute an object.The former requires the availability of object-level segmentation(the output ofour system)[11].Our focus is on the second type of support relations to guide object-levelsegmentation.

    We take the patches constructed in Section 4.1 as input.First we classify all patches into two types:quasi-verticalQvand quasi-horizontalQhaccording to the angles between the planes fitted to them and the ground plane(relative to 45°). Then we concentrate on two common relationships: a quasi-verticalpatch supporting an quasi-horizontal patchQvhabove it,and a quasi-vertical patch being supported by a quasi-horizontal patchQhvbelow it, as shown in Fig.7.

    For each pair of an adjacent quasi-vertical patchPiand a quasi-horizontal patchPj(abovePi), the likelihood that patchesPiandPjhave aQvhrelationship is defined as follows:

    Fig.7 Automatically detected support relations.Left:the quasivertical patches co-support the top quasi-horizontal patches.

    whereU(Pi,Pj)is the number of quasi-vertical patches within the neighborhood ofPi(includingPi)which havePjas an upper patch andλQ=0.2 is a fixed parameter.Intuitively,more quasi-vertical patches supporting the same quasi-horizontal patch lead to higher likelihood of having theQvhrelationship.

    Similarly,for each pair of an adjacent quasivertical patchPiand a quasi-horizontal patchPj(belowPi),the likelihood that patchesPiandPjhave aQhvrelationship is defined as follows:

    whereWiis the area of the convex hull of points in patchPi.We consider aQhvrelationship for a quasi-vertical patchPionly when there is noQvhrelationship involvingPi.Consider Fig.7(middle): the laptop lid hasQhvrelationship with the table, the laptop base,and the keyboard,while the laptop base leads to the highest likelihood since they have about the same size.

    Then we defineTS(Pi,Pj),the likelihood for patchesPiandPjto belong to the same object using inference from the support map,as follows:

    whereQxcan be eitherQvhorQhv,decided by the spatial relationship between two concerning patches.The top two conditions are essentiallyQvhorQhv,respectively,and the third indicates cosupport relations between patches inQv,which exist ifand only iftwo verticalpatches have the same type of support relations(Qx)with the same horizontal patch(see the example in Fig.7(left)).Since support relations between patches are irrelevant to the segmentation status,in our system,they are computed during the preprocessing step.

    Graph-cut.Given a set of user-specified strokes, only pixels with the same patch indices on both the point cloud view and the supplementary mesh view are regarded as valid,to avoid penetrationlabeling.We then formulate support relations as well as distances,colors,and normals in the graph-cut formulation.Please refer to the ESM for details.

    6.2 Finding similar objects

    Duplicate objects often exist in a scene.It is redundant for a user to label them one by one.Our system employs a simple algorithm to automatically retrieve candidate objects similar to a given object in the scene.Ifthe found objects are in fact not similar, users can delete the automatically generated labels.

    Several methods for finding similar objects in point cloud scenes have already been proposed.For example,Kim et al.[9]introduced a method by first learning models of frequently occurring objects and then performing real-time matching.Mattausch et al.[24]took patch similarity and spatialconsistency into account and automatically found all similar objects in a scene by clustering,which,however, takes dozens of seconds for large-scale indoor scenes.

    We aim for a simple and efficient tool that can perform in real time,without pre-trained models or clustering.Our approach is based on the key assumption that similar objects should contain similar support structures,i.e.,both the support relations and the accompanying patches should be similar.This assumption allows us to identify similar objects by straightforwardly comparing allunlabeled candidate support structures with a given object, based on patch similarity and support relations.

    More specifically,given an already segmented object,denoted byC0,we first identify allcandidate patches which have not been segmented and have a high similarity(t7=0.3 in our implementation)to at least one patch inC0.For efficiency,we slightly change the similarity metric in Ref.[24]to define our similarityAi,jfor a given pair of patchesPiandPj(see the ESM).

    Then we group these candidate patches by connectedness to form patch constellations.For each pair of patches〈Pi,Pj〉in each constellationCn,we search for a corresponding pair〈Pk,Pl〉∈C0with the same type of support relation.Based on our key assumption,we compute the similarity likelihood Sim(Cn,C0)betweenCnandC0as follows:

    If the similarity likelihood Sim(Cn,C0)exceeds a thresholdt8(=0.25 in our implementation),we take it as a candidate similar object.We then use all patches in the found isomorphic support structures as seeds for a new object to perform another round of graph-cut.

    As shown in Fig.8 our algorithm manages to find several similar objects in two cases.In the left example,each chair is composed of different numbers of patches,but the algorithm is still able to find such similar isomorphic parts and perform graph-cut.The right example shows our ability to find diff erent types of objects:tables (green),monitors(cyan),telephones(yellow),and keyboards(purple).Note that objects adjacent to some keyboards are mistaken by graph-cut, demanding additionalstrokes for refinement.

    7 Results and evaluation

    We have tested our system on different qualities of datasets from different sources.We also performed a user study to evaluate the performance of view suggestion.Please see the accompanying video in the ESM for interactive demos.Some representative suggested views and segmentation results are shown in Fig.9.

    Datasets.In order to test our performance of view suggestion and segmentation,we chose middle-or large-scale scenes with rich sets ofobjects. Specifically,we used 2 out of 3 scenes(office2, office3)in the room detection datasets of UZH[24], 1 of 5 scenes(office1)in the Floored Panorama dataset[37],and 2 of 2 synthetic scenes(officeroom,living-room)with ground-truth trajectory in the ICL-NUIM dataset[45].The ICL-NIUM datawas generated by sampling point on surfaces of existing 3D mesh-based models,while others were acquired with LiDAR(with color information).Figure 9 illustrates all scenes tested.We additionally collected a scene captured by Microsoft Kinect2 to test for poor-quality scans.

    Fig.8 Results of finding similar objects(shown in the same color).

    Fig.9 Examples of results of our system.Left:view score map.Right:high score examples when no objects are labeled.

    Fig.10 Results of user study.Top,left:scenes with pre-defined target segmentations.(a)Acceptance rate of suggestions and hypothesis test results.Thep-values are from paired samplet-tests between our method and no suggestion(n)or RGB-D(d),giving full time(T),the time spent on manual navigation(N),the counts of minimap navigation(M)and used views(V).Right:average and standard deviation of the means of observed values.(b)Time spent on the whole process in second.(c)Time spent on manual navigation in second.(d)The count of used views.

    7.1 Validity of view suggestion

    To evaluate the effectiveness of view suggestion, we compared the performance of interactive segmentation with and without view suggestion.We also compared to stroke-based segmentation directly on RGB-D frames ifavailable.These three methods, denoted as M-S,M-N,and M-D,were compared in a user study.Please refer to the accompanying video in the ESM to see how each method works.

    Interfaces.For each of the methods compared, we provided a minimap for easy navigation.For M-D,while interactive segmentation was done on RGB-D images,we provided an additional point cloud viewer for examining the segmentation status.The only difference between M-S and MN was the availability of view suggestion.For MS,we encouraged the participants to use the view suggestion feature as much as possible,though minimap navigation and common 3D browsing were still allowed.All methods were tested in fullscreen mode(1080 p)to eliminate the influence of window size.For fair comparison,we used the same preprocessing,graph-cut,and similar-object-finding algorithms for all methods.

    Participants and apparatus.We recruited 12 university students,who had basic knowledge of graphics but different levels of experience in 3D browsing.In a training session,which lasted nearly 15 minutes on average,each participant was briefed and trained how to use the three methods.The entire user study was conducted on the same PC with i7 2.10 GHz CPU,a GTX 660Ti GPU,and 16 GB RAM.

    Tasks.We prepared 3 scenes with target segmentations.Scene A in Fig.10(a)came from the ICL-NUIM dataset with 42 objects and 35 RGB-D key-frames.Scene B was our own recorded Kinect2 data,including 22 objects and 33 keyframes.Scene C was a large-scale multiple-room scene with 21 objects but no associated RGBD images.These scenes are given in rows 1–3, respectively,in Fig.9.The participants were asked to interactively segment these scenes using the three methods,whose presentation was counterbalanced to alleviate bias due to familiarity with the scenes.For Scene C,we tested M-S and M-N only,due to its lack of RGB-D frames.We pre-defined a groundtruth segmentation for each scene with manually labeled objects and required the participants to reproduce the ground-truth segmentations(shown on another monitor simultaneously)as closely as possible.Since it was difficult to reproduce the segmentations exactly,each scene was considered as completed ifthe similarity(ratio ofpatches)between the current and target segmentations reached 90%.

    Measures.Our system recorded the following information for comparison:time spent on manual navigation,processing,and the whole progress (including planning,drawing,processing,and navigation);number of uses of view suggestion and minimap navigation operations;number of views used;number of executions ofgraph-cut and similarobject-finding algorithms;number and total length of input strokes.Additionally,we also recorded the acceptance ratio of view suggestion.We conducted a questionnaire survey at the end of the study.

    Results.The acceptance rates of our suggestedviews for Scenes A,B,and C were 83.9%,85.8%, and 93.1%,respectively,indicating the good quality of suggested views.Figures 10(b)–10(d)show the statistics of some observed values.The table in Fig.10(a)gives evaluation and hypothesis testing results,wherepis thep-value calculated through student’st-test[46],which is helpful for evaluating statistically significant differences among small-scale paired-samples.

    Using one-way analysis of variance(ANOVA),we found a significant difference in the average total time between the tested methods(p=0.0288 for all methods on Scenes A and B,p=0.0058 for M-S and M-N on all the scenes).In all tested scenes,our method significantly outperformed M-D (p<0.005 fromt-test)in both the number of views used and the total time.In Scene C,our method was significantly faster than M-N(p<0.001), although they achieved comparable performance on minimap navigation(p=0.0527).We observed that minimap navigation is crucial for large-scale scenes.In Scenes A and B,compared to M-N,our method significantly reduced the time for camera manipulation(p<0.005)and the usage of minimap navigation(p<0.03),but achieved comparable performance in total time(p>0.05).This might be because the expense ofmanualnavigation in such single-room scenes is relatively low.Our method also used significantly fewer views than M-N(p<0.02) when segmenting all the tested scenes.

    Our participants reported that with M-D,the RGB-D image views were more useful for object recognition due to their dense views,but(re-) checking the labeled objects,either through backand-forth traversal or with the help of the auxiliary point cloud view,was not convenient.Worse,the number of RGB-D key-frames can significantly increase as scenes get bigger.Some participants also reported that they subconsciously referred to the good RGB-D views or results of view suggestion to improve labeling efficiency when using M-N.In fact, the performance ofinteractive segmentation without view suggestion can be easily affected by familiarity with the scenes and the proficiency of 3D browsing.

    7.2 Robustness and efficiency

    Parameters.In the preprocessing step for extracting storeys and sampling viewpoints,we used different sets of parameters for diff erent datasets due to their different data quality and scene scale. The other parameters for patch construction,view suggestions,and interactive segmentation,remained unchanged for allthe tested scenes.

    Timing.On the same machine used in the user study,the preprocessing stage can cost dozens of minutes for the tested scenes.For a 140 m2onefloor room with 16 million points,it took about 1 minute per viewpoint to traverse through scenes to get the descriptors and 30 minutes in total;this could potentially be made faster by switching to quicker rendering tools.Since the score of each view can be calculated separately,we can also easily accelerate the method by parallelization.In the online stage, it only cost 50 ms to suggest an appropriate view, the graph-cut process only took 100 ms,and finding similar objects took 200 ms with respect to a labeled object.In summary,the online processing costs meet the requirement of real-time response.

    Limitations.In some scenes with complicated objects such as plants or trees,our system may“prefer”to suggest views involving such objects (Fig.11(left)),since they are taken as fragmentary patches(each including several leaves or stems)in patch construction,leading to high residual entropy on these views.However,once such objects are labeled,their influence willdisappear.Since patches are indivisible in our implementation,if two or more objects are gathered into one patch(e.g.,due to similar color and depth),it is impossible to segment the individualobjects.Also the boundaries ofobjects may not be accurate if the points on boundaries are ambiguous.This might be addressed by introducing an interactive repair tool for boundary refinement. In addition,our proposed support relations and the algorithm for finding similar objects may fail if objects are combined as multiple discrete components(see Fig.11(right)).Fortunately,users can modify the labeling by providing additionalinputs for segmentation.

    Fig.11 Less successful cases.Left:an example view with a tree with a high score.Right:failure to find similar objects in isolated patches(the chair in red).

    8 Conclusions and future work

    We have presented a novel system for interactive segmentation of large-scale indoor scenes, represented as cluttered point clouds.The key contributions of our work are the problem and solution of automatic view suggestion for interactive segmentation.Other contributions include support relations for graph-cut-based segmentation and finding similar objects.Our extensive evaluations show the advantages ofour approach over alternative methods.In future,besides addressing the discussed limitations,we are also interested in applying a similar idea ofview suggestion for other applications, e.g.,interactive editing of large-scale scenes.

    Acknowledgements

    This work was supported by the Joint NSFC–ISF Research Program(Project No.61561146393), the National Natural Science Foundation of China(Project No.61521002),the Research Grant of Beijing Higher Institution Engineering Research Center,and the Tsinghua–Tencent Joint Laboratory for Internet Innovation Technology. The work was partially supported by grants from the Research Grants Council of the Hong Kong Special Administrative Region,China(Project Nos. CityU113513 and CityU11300615).

    Electronic Supplementary Material Supplementary material is available in the online version of this article at http://dx.doi.org/10.1007/s41095-017-0078-4.

    [1]Lai,K.;Bo,L.;Ren,X.;Fox,D.Detection-based object labeling in 3D scenes.In:Proceedings of the IEEE International Conference on Robotics and Automation,1330–1337,2012.

    [2]Johnson,A.E.;Hebert,M.Using spin images for efficient object recognition in cluttered 3D scenes.IEEE Transactions on Pattern Analysis and Machine IntelligenceVol.21,No.5,433–449,1999.

    [3]Zheng,B.;Zhao,Y.;Yu,J.C.;Ikeuchi,K.;Zhu, S.C.Beyond point clouds:Scene understanding by reasoning geometry and physics.In:Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,3127–3134,2013.

    [4]Holz,D.;Behnke,S.Fast range image segmentation and smoothing using approximate surface reconstruction and region growing.In:Intelligent Autonomous Systems 12.Lee,S.;Cho,H.;Yoon, K.-J.;Lee,J.Eds.Springer Berlin Heidelberg,61–73, 2013.

    [5]Rabbani,T.;van den Heuvel,F.A.;Vosselmann, G.Segmentation of point clouds using smoothness constraint.International Archives of Photogrammetry, Remote Sensing and Spatial Information SciencesVol. 36,No.5,248–253,2006.

    [6]Boykov,Y.;Funka-Lea,G.Graph cuts and efficient N-D image segmentation.International Journal of Computer VisionVol.70,No.2,109–131,2006.

    [7]Golovinskiy,A.;Funkhouser,T.Min-cut based segmentation of point clouds.In:Proceedings of the IEEE 12th International Conference on Computer Vision Workshops,39–46,2009.

    [8]Sedlacek,D.;Zara,J.Graph cut based pointcloud segmentation for polygonal reconstruction.In:Advances in Visual Computing.Bebis,G.;Boyle,R.; Parvin,B.;Koracin,D.et al.Eds.Springer Berlin Heidelberg,218–227,2009.

    [9]Kim,Y.M.;Mitra,N.J.;Yan,D.-M.;Guibas,L. Acquiring 3D indoor environments with variability and repetition.ACM Transactions on GraphicsVol.31, No.6,Article No.138,2012.

    [10]Nan,L.;Xie,K.;Sharf,A.Asearch-classifyapproach for cluttered indoor scene understanding.ACM Transactions on GraphicsVol.31,No.6,Article No.137,2012.

    [11]Silberman,N.;Hoiem,D.;Kohli,P.;Fergus,R. Indoor segmentation and support inference from RGBD images.In:Computer Vision–ECCV 2012. Fitzgibbon,A.;Lazebnik,S.;Perona,P.;Sato,Y.; Schmid,C.Eds.Springer Berlin Heidelberg,746–760, 2012.

    [12]Nguyen,A.;Le,B.3D point cloud segmentation:A survey.In:Proceedings of the 6th IEEE Conference on Robotics,Automation and Mechatronics,225–230, 2013.

    [13]Shao,T.;Xu,W.;Zhou,K.;Wang,J.;Li,D.; Guo,B.An interactive approach to semantic modeling of indoor scenes with an RGBD camera.ACM Transactions on GraphicsVol.31,No.6,Article No. 136,2012.

    [14]Yuan,X.;Xu,H.;Nguyen,M.X.;Shesh,A.; Chen,B.Sketch-based segmentation of scanned outdoor environment models.In:Proceedings of the Eurographics Workshop on Sketch-Based Interfaces and Modeling,19–26,2005.

    [15]Silberman,N.;Fergus,R.Indoor scene segmentation using a structured light sensor.In:Proceedings of the IEEE International Conference on Computer Vision Workshops,601–608,2011.

    [16]Xiao,J.;Owens,A.;Torralba,A.SUN3D:A database of big spaces reconstructed using SfM and object labels.In:Proceedings of the IEEE International Conference on Computer Vision,1625–1632,2013.

    [17]Chang,A.X.;Funkhouser,T.;Guibas,L.;Hanrahan, P.;Huang,Q.;Li,Z.;Savarese,S.;Savva,M.;Song, S.;Su,H.;Xiao,J.;Yi,L.;Yu,F.ShapeNet:An information-rich 3D model repository.arXiv preprintarXiv:1512.03012,2015.

    [18]Hinterstoisser,S.;Lepetit,V.;Ilic,S.;Holzer,S.; Bradski,G.R.;Konolige,K.;Navab,N.Model based training,detection and pose estimation of texture-less 3D objects in heavily cluttered scenes.In:Computer Vision–ACCV 2012.Lee,K.M.;Matsushita,Y.;Rehg, J.M.;Hu,Z.Eds.Springer Berlin Heidelberg,548–562,2012.

    [19]Chen,K.;Lai,Y.-K.;Wu,Y.-X.;Martin,R.;Hu,S.-M.Automatic semantic modeling ofindoor scenes from low-quality RGB-D data using contextualinformation.ACM Transactions on GraphicsVol.33,No.6,Article No.208,2014.

    [20]Silberman,N.;Sontag,D.;Fergus,R.Instance segmentation ofindoor scenes using a coverage loss.In:Computer Vision–ECCV 2014.Fleet,D.;Pajdla,T.; Schiele,B.;Tuytelaars,T.Eds.Springer International Publishing,616–631,2014.

    [21]Chen,K.;Lai,Y.K.;Hu,S.-M.3D indoor scene modeling from RGB-D data:a survey.Computational Visual MediaVol.1,No.4,267–278,2015.

    [22]Shen,C.-H.;Huang,S.-S.;Fu,H.;Hu,S.-M.Adaptive partitioning of urban facades.ACM Transactions on GraphicsVol.30,No.6,Article No.184,2011.

    [23]Zhang,H.;Xu,K.;Jiang,W.;Lin,J.;Cohen-Or, D.;Chen,B.Layered analysis of irregular facades via symmetry maximization.ACM Transactions on GraphicsVol.32,No.4,Article No.121,2013.

    [24]Mattausch,O.;Panozzo,D.;Mura,C.;Sorkine-Hornung,O.;Pajarola,R.Object detection and classification from large-scale cluttered indoor scans.Computer Graphics ForumVol.33,No.2,11–21,2014.

    [25]Valentin,J.;Vineet,V.;Cheng,M.-M.;Kim,D.; Shotton,J.;Kohli,P.;Nie?ner,M.;Criminisi, A.;Izadi,S.;Torr,P.SemanticPaint:Interactive 3D labeling and learning at your fingertips.ACM Transactions on GraphicsVol.34,No.5,Article No. 154,2015.

    [26]Wong,Y.-S.;Chu,H.-K.;Mitra,N.J.SmartAnnotator an interactive tool for annotating indoor RGBD images.Computer Graphics ForumVol.34,No.2,447–457,2015.

    [27]Christie,M.;Olivier,P.Camera control in computer graphics:Models,techniques and applications.In: Proceedings of the ACM SIGGRAPH ASIA 2009 Courses,Article No.3,2009.

    [28]Scott,W.R.;Roth,G.;Rivest,J.-F.View planning for automated three-dimensionalobject reconstruction and inspection.ACM Computing SurveysVol.35,No. 1,64–96,2003.

    [29]Secord,A.;Lu,J.;Finkelstein,A.;Singh,M.;Nealen, A.Perceptual models of viewpoint preference.ACM Transactions on GraphicsVol.30,No.5,Article No. 109,2011.

    [30]V′azquez,P.-P.;Feixas,M.;Sbert,M.;Heidrich, W.Viewpoint selection using viewpoint entropy.In: Proceedings of the Vision Modeling and Visualization Conference,273–280,2001.

    [31]And′ujar,C.;V′azquez,P.;Fair′en,M.Way-Finder: Guided tours through complex walkthrough models.Computer Graphics ForumVol.23,No.3,499–508, 2004.

    [32]Li,T.-Y.;Lien,J.-M.;Chiu,S.-Y.;Yu,T.-H. Automatically generating virtual guided tours.In: Proceedings of the Computer Animation,99–106, 1999.

    [33]Christie,M.;Langu′enou,E.A constraint-based approach to camera path planning.In:Smart Graphics.Butz,A.;Kr¨uger,A.;Olivier,P.Eds. Springer Berlin Heidelberg,172–181,2003.

    [34]Salomon,B.;Garber,M.;Lin,M.C.;Manocha,D. Interactive navigation in complex environments using path planning.In:Proceedings of the Symposium on Interactive 3D Graphics,41–50,2003.

    [35]Choi,S.;Zhou,Q.-Y.;Koltun,V.Robust reconstruction of indoor scenes.In:Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,5556–5565,2015.

    [36]Newcombe,R.A.;Izadi,S.;Hilliges,O.;Molyneaux, D.;Kim,D.;Davison,A.J.;Kohli,P.;Shotton,J.; Hodges,S.;Fitzgibbon,A.KinectFusion:Real-time dense surface mapping and tracking.In:Proceedings of the 10th IEEE International Symposium on Mixed and Augmented Reality,127–136,2011.

    [37]Ikehata,S.;Yang,H.;Furukawa,Y.Structured indoor modeling.In:Proceedings of the IEEE International Conference on Computer Vision,1323–1331,2015.

    [38]Furukawa,Y.;Curless,B.;Seitz,S.M.;Szeliski, R.Manhattan-world stereo.In:Proceeding of the IEEE Conference on Computer Vision and Pattern Recognition,1422–1429,2009.

    [39]Asha,V.;Bhajantri,N.U.;Nagabhushan,P. GLCM-based chi-square histogram distance for automatic detection of defects on patterned textures.International Journal of Computational Vision and RoboticsVol.2,No.4,302–313,2011.

    [40]Fr¨uh,C.;Zakhor,A.Constructing 3D city models by merging aerial and ground views.IEEE Computer Graphics and ApplicationsVol.23,No.6,52–61,2003.

    [41]Fisher,M.;Savva,M.;Li,Y.;Hanrahan,P.;Nie?ner, M.Activity-centric scene synthesis for functional 3D scene modeling.ACM Transactions on GraphicsVol. 34,No.6,Article No.179,2015.

    [42]Marton,Z.C.;Rusu,R.B.;Beet,M.On fast surface reconstruction methods for large and noisy point clouds.In:Proceedings of the IEEE International Conference on Robotics and Automation,3218–3223, 2009.

    [43]Tibshirani,R.Regression shrinkage and selection via the lasso.Journal of the Royal Statistical SocietyVol. 58,No.1,267–288,1996.

    [44]Boykov,Y.;Veksler,O.;Zabih,R.Fast approximate energy minimization via graph cuts.IEEE Transactions on Pattern Analysis and Machine IntelligenceVol.23,No.11,1222–1239,2001.

    [45]Handa,A.;Whelan,T.;McDonald,J.;Davison, A.J.A benchmark for RGB-D visual odometry, 3D reconstruction and SLAM.In:Proceedings of the IEEE International Conference on Robotics and Automation,1524–1531,2014.

    [46]Gosset,W.S.The probable error of a mean.BiometrikaVol.6,No.1,1–25,1908.

    Jie Xu is a Ph.D.student at the Computer Science and Artificial Intelligence Laboratory in Massachusetts Institute of Technology. His research interests include computer graphics and geometric processing.

    Kang Chen received his B.S.degree in computer science from Nanjing University in 2012.He is currently a Ph.D.candidate in the Institute for Interdisciplinary Information Sciences, Tsinghua University.His research interests include computer graphics, geometric modeling and processing.

    Hongbo Fu is an associate professor in the School of Creative Media,City University of Hong Kong.He received his Ph.D.degree in computer science from the Hong Kong University of Science and Technology in 2007 and B.S.degree in information sciences from Peking University in 2002.His primary research interests fall in the fields of computer graphics and human computer interaction.He has served as an associate editor ofThe Visual Computer,Computers&Graphics,andComputer Graphics Forum.

    Open Access The articles published in this journal are distributed under the terms of the Creative Commons Attribution 4.0 International License(http:// creativecommons.org/licenses/by/4.0/),which permits unrestricted use,distribution,and reproduction in any medium,provided you give appropriate credit to the original author(s)and the source,provide a link to the Creative Commons license,and indicate if changes were made.

    Other papers from this open access journalare available free of charge from http://www.springer.com/journal/41095. To submit a manuscript,please go to https://www. editorialmanager.com/cvmj.

    g

    his B.S.degree in computer science from Wuhan University in 2014.He is currently a Ph.D.candidate in computer science in Tsinghua University.His research interests include computer graphics and point cloud processing.

    1 Tsinghua University,Beijing,China.E-mail:S. Yang,shengyang93fs@gmail.comK.Chen, chenkangnobel@hotmail.com.

    2 Massachusetts Institute of Technology,Cambridge,USA. E-mail:eternal answer@126.com.

    3 City University of Hong Kong,Hong Kong,China.E-mail:fuplus@gmail.com.

    Manuscript received:2016-12-02;accepted:2017-01-12

    国产又爽黄色视频| 老汉色av国产亚洲站长工具| 精品一区二区三区av网在线观看 | 妹子高潮喷水视频| 日韩大片免费观看网站| 亚洲精品中文字幕一二三四区 | 国产av又大| 欧美激情 高清一区二区三区| 老司机深夜福利视频在线观看| 午夜老司机福利片| 99国产精品免费福利视频| 久久人人97超碰香蕉20202| 桃花免费在线播放| 97在线人人人人妻| 精品久久久久久电影网| 老熟妇仑乱视频hdxx| 两性夫妻黄色片| 激情视频va一区二区三区| 香蕉国产在线看| 成人特级黄色片久久久久久久 | 19禁男女啪啪无遮挡网站| 午夜激情久久久久久久| 天天躁狠狠躁夜夜躁狠狠躁| 99热网站在线观看| 99re在线观看精品视频| 一级毛片电影观看| 99国产综合亚洲精品| av国产精品久久久久影院| 国产精品偷伦视频观看了| 久久久久精品人妻al黑| 色在线成人网| 岛国在线观看网站| 久久久国产一区二区| 麻豆av在线久日| 亚洲中文av在线| av欧美777| 天堂俺去俺来也www色官网| 亚洲色图综合在线观看| 亚洲国产欧美在线一区| 国产av又大| 91av网站免费观看| 中文字幕人妻丝袜一区二区| 日韩制服丝袜自拍偷拍| 老司机深夜福利视频在线观看| 最近最新中文字幕大全电影3 | 亚洲精品粉嫩美女一区| 老熟妇仑乱视频hdxx| 在线十欧美十亚洲十日本专区| 中文字幕最新亚洲高清| 国产av又大| 日韩欧美免费精品| 久久精品国产亚洲av香蕉五月 | 啦啦啦视频在线资源免费观看| 一本色道久久久久久精品综合| 精品久久蜜臀av无| 国产真人三级小视频在线观看| 亚洲精品在线美女| 亚洲自偷自拍图片 自拍| 久久人妻av系列| 久久久久久亚洲精品国产蜜桃av| 亚洲黑人精品在线| 亚洲中文av在线| 性高湖久久久久久久久免费观看| av欧美777| 久久久久久人人人人人| a级毛片黄视频| 精品一区二区三区av网在线观看 | 一本色道久久久久久精品综合| 一边摸一边抽搐一进一小说 | 三上悠亚av全集在线观看| 中文字幕高清在线视频| 熟女少妇亚洲综合色aaa.| 99精品久久久久人妻精品| 老鸭窝网址在线观看| 露出奶头的视频| 高清视频免费观看一区二区| av视频免费观看在线观看| 午夜两性在线视频| 午夜免费鲁丝| 夜夜夜夜夜久久久久| 99riav亚洲国产免费| 久久人妻熟女aⅴ| 欧美黄色片欧美黄色片| 免费少妇av软件| 午夜视频精品福利| 99九九在线精品视频| 国产成人av激情在线播放| 免费人妻精品一区二区三区视频| 久久久久久亚洲精品国产蜜桃av| 国产亚洲精品第一综合不卡| 亚洲欧美精品综合一区二区三区| 成人亚洲精品一区在线观看| 麻豆成人av在线观看| 亚洲中文日韩欧美视频| av一本久久久久| 成人国产av品久久久| 一本色道久久久久久精品综合| 满18在线观看网站| 亚洲精品一二三| 亚洲 国产 在线| 一二三四在线观看免费中文在| 欧美日韩国产mv在线观看视频| 久久久精品94久久精品| 国产精品秋霞免费鲁丝片| 一级,二级,三级黄色视频| 久久毛片免费看一区二区三区| 纯流量卡能插随身wifi吗| 色婷婷久久久亚洲欧美| 精品高清国产在线一区| 18禁裸乳无遮挡动漫免费视频| 欧美精品一区二区免费开放| 19禁男女啪啪无遮挡网站| 日韩欧美一区二区三区在线观看 | 免费女性裸体啪啪无遮挡网站| 久久久久久免费高清国产稀缺| 少妇的丰满在线观看| 人人妻人人澡人人爽人人夜夜| 咕卡用的链子| 免费看十八禁软件| 午夜免费成人在线视频| 成人18禁高潮啪啪吃奶动态图| 99精国产麻豆久久婷婷| 妹子高潮喷水视频| 777米奇影视久久| 男女床上黄色一级片免费看| 不卡av一区二区三区| 免费不卡黄色视频| 亚洲精品国产精品久久久不卡| 建设人人有责人人尽责人人享有的| 亚洲av片天天在线观看| 久久青草综合色| 老司机午夜十八禁免费视频| 一区二区三区精品91| 露出奶头的视频| 国产一区二区三区综合在线观看| 久热这里只有精品99| 男人舔女人的私密视频| 妹子高潮喷水视频| 变态另类成人亚洲欧美熟女 | 法律面前人人平等表现在哪些方面| 免费不卡黄色视频| 国产欧美日韩综合在线一区二区| 亚洲国产精品一区二区三区在线| 亚洲精品粉嫩美女一区| 亚洲久久久国产精品| kizo精华| 国产精品香港三级国产av潘金莲| 国产在线一区二区三区精| 老司机靠b影院| 国产精品亚洲一级av第二区| 我要看黄色一级片免费的| netflix在线观看网站| 一级毛片精品| 50天的宝宝边吃奶边哭怎么回事| 亚洲欧洲日产国产| 久久国产精品影院| 亚洲欧美日韩高清在线视频 | 欧美日韩黄片免| 精品福利观看| 99九九在线精品视频| 精品国产乱子伦一区二区三区| 多毛熟女@视频| 日韩大片免费观看网站| 午夜久久久在线观看| 亚洲综合色网址| 菩萨蛮人人尽说江南好唐韦庄| 久久久国产成人免费| 日本一区二区免费在线视频| 在线观看免费日韩欧美大片| 国产av又大| 18禁黄网站禁片午夜丰满| 黄色视频,在线免费观看| h视频一区二区三区| 成年动漫av网址| 成人特级黄色片久久久久久久 | 人妻久久中文字幕网| 一进一出抽搐动态| 制服诱惑二区| 夜夜夜夜夜久久久久| 国产亚洲欧美精品永久| 久久久精品国产亚洲av高清涩受| 人成视频在线观看免费观看| 亚洲一码二码三码区别大吗| 满18在线观看网站| 狠狠婷婷综合久久久久久88av| 国产亚洲精品久久久久5区| 亚洲欧美一区二区三区黑人| 99国产精品一区二区蜜桃av | 国产不卡av网站在线观看| 另类亚洲欧美激情| 国产精品香港三级国产av潘金莲| 色在线成人网| 久久久久久人人人人人| 蜜桃国产av成人99| 黄网站色视频无遮挡免费观看| 美女国产高潮福利片在线看| 黄色a级毛片大全视频| 最新美女视频免费是黄的| 欧美中文综合在线视频| 黄色丝袜av网址大全| 亚洲精品乱久久久久久| 亚洲熟女精品中文字幕| 一区二区三区精品91| 亚洲七黄色美女视频| 中文字幕另类日韩欧美亚洲嫩草| 五月开心婷婷网| kizo精华| 国产不卡av网站在线观看| 久热这里只有精品99| 他把我摸到了高潮在线观看 | 国产精品.久久久| 黄片小视频在线播放| 久久久水蜜桃国产精品网| 亚洲,欧美精品.| 日本黄色日本黄色录像| 美女高潮到喷水免费观看| 少妇裸体淫交视频免费看高清 | 亚洲一区中文字幕在线| 亚洲三区欧美一区| 三上悠亚av全集在线观看| 十八禁高潮呻吟视频| 美女午夜性视频免费| 18禁国产床啪视频网站| 女人久久www免费人成看片| 91av网站免费观看| 精品福利永久在线观看| 国产熟女午夜一区二区三区| 首页视频小说图片口味搜索| 久久久精品94久久精品| 国产精品一区二区免费欧美| 国产精品电影一区二区三区 | 老司机午夜福利在线观看视频| 国产成人欧美在线观看| 久久久国产成人精品二区| 这个男人来自地球电影免费观看| 久久久水蜜桃国产精品网| 99国产极品粉嫩在线观看| av福利片在线观看| 国产1区2区3区精品| 国产精品久久电影中文字幕| 别揉我奶头~嗯~啊~动态视频| 亚洲国产欧美一区二区综合| 十八禁网站免费在线| 国产三级中文精品| 99久久综合精品五月天人人| 一本久久中文字幕| 一个人免费在线观看电影 | 夜夜爽天天搞| 国产成人aa在线观看| 欧美日韩亚洲国产一区二区在线观看| xxx96com| 欧美乱色亚洲激情| 悠悠久久av| 国内揄拍国产精品人妻在线| 久久久久久久久久黄片| 舔av片在线| 欧美xxxx黑人xx丫x性爽| 国产精品美女特级片免费视频播放器 | 欧美又色又爽又黄视频| 国产精品av视频在线免费观看| 欧美黄色片欧美黄色片| 我要搜黄色片| 夜夜爽天天搞| 亚洲人与动物交配视频| 国产成人啪精品午夜网站| 99精品在免费线老司机午夜| 又黄又粗又硬又大视频| 夜夜看夜夜爽夜夜摸| 99久久精品热视频| 国产精品av久久久久免费| 中国美女看黄片| www.精华液| 黑人巨大精品欧美一区二区mp4| 亚洲av免费在线观看| 欧美成人一区二区免费高清观看 | 亚洲av片天天在线观看| 在线十欧美十亚洲十日本专区| 一二三四在线观看免费中文在| 国产av在哪里看| 一区福利在线观看| 国产黄a三级三级三级人| 观看美女的网站| 日本一二三区视频观看| 性欧美人与动物交配| 精品久久久久久久久久免费视频| 亚洲无线观看免费| 99国产精品一区二区三区| 免费在线观看影片大全网站| 男人舔女人的私密视频| 国产一区二区在线观看日韩 | 在线国产一区二区在线| 成人特级黄色片久久久久久久| av天堂在线播放| 欧美午夜高清在线| 亚洲av电影在线进入| 亚洲色图 男人天堂 中文字幕| 久久久国产成人精品二区| 天堂网av新在线| 久久久久免费精品人妻一区二区| 全区人妻精品视频| 亚洲乱码一区二区免费版| 国产野战对白在线观看| 国产精品电影一区二区三区| 少妇裸体淫交视频免费看高清| 亚洲国产欧美网| 亚洲中文av在线| 两性午夜刺激爽爽歪歪视频在线观看| 午夜日韩欧美国产| 色老头精品视频在线观看| 国产精品综合久久久久久久免费| 国产主播在线观看一区二区| 露出奶头的视频| 国产亚洲精品久久久com| 91麻豆av在线| 中文字幕高清在线视频| 久久香蕉精品热| 精华霜和精华液先用哪个| 岛国视频午夜一区免费看| 欧美成人性av电影在线观看| 熟女人妻精品中文字幕| 99久久成人亚洲精品观看| 国产精品99久久久久久久久| 91字幕亚洲| 美女黄网站色视频| 国内毛片毛片毛片毛片毛片| 十八禁人妻一区二区| 一卡2卡三卡四卡精品乱码亚洲| 99国产精品99久久久久| 人妻夜夜爽99麻豆av| 亚洲av美国av| 中国美女看黄片| 成人亚洲精品av一区二区| 看片在线看免费视频| 老熟妇乱子伦视频在线观看| 神马国产精品三级电影在线观看| 精品乱码久久久久久99久播| 日韩欧美免费精品| 国产成人欧美在线观看| 9191精品国产免费久久| 国产乱人视频| 精品乱码久久久久久99久播| 麻豆av在线久日| 99久久成人亚洲精品观看| 一个人看的www免费观看视频| 亚洲国产精品sss在线观看| 99国产综合亚洲精品| 99在线视频只有这里精品首页| 天堂影院成人在线观看| 一二三四在线观看免费中文在| 一级a爱片免费观看的视频| 国产精品av久久久久免费| 99久国产av精品| 日韩欧美在线乱码| 国产99白浆流出| 欧美午夜高清在线| 欧美激情在线99| 色综合欧美亚洲国产小说| 麻豆一二三区av精品| 亚洲专区中文字幕在线| 亚洲精品色激情综合| 欧美乱妇无乱码| 夜夜躁狠狠躁天天躁| 精品国产超薄肉色丝袜足j| 五月玫瑰六月丁香| 欧美乱妇无乱码| 国内精品久久久久精免费| 精品国产超薄肉色丝袜足j| 88av欧美| 男人的好看免费观看在线视频| 91老司机精品| 国产午夜福利久久久久久| 国产精品久久久久久精品电影| 99久久无色码亚洲精品果冻| 伦理电影免费视频| 欧美日韩精品网址| 黄色女人牲交| 亚洲五月天丁香| 国产v大片淫在线免费观看| 欧美乱色亚洲激情| 久久人人精品亚洲av| 天天添夜夜摸| 国产成人一区二区三区免费视频网站| 男女那种视频在线观看| 欧美日韩综合久久久久久 | 女人被狂操c到高潮| 精品久久久久久久久久免费视频| 曰老女人黄片| 久久精品aⅴ一区二区三区四区| 国产成人精品久久二区二区91| 欧美极品一区二区三区四区| 天天一区二区日本电影三级| 午夜福利18| 国产伦人伦偷精品视频| 在线看三级毛片| 一本综合久久免费| 亚洲成人精品中文字幕电影| 一个人免费在线观看的高清视频| 久久午夜亚洲精品久久| 高潮久久久久久久久久久不卡| 999精品在线视频| 亚洲精品美女久久av网站| tocl精华| 国产 一区 欧美 日韩| 国产乱人伦免费视频| 男人舔女人下体高潮全视频| 十八禁人妻一区二区| 啦啦啦观看免费观看视频高清| 我的老师免费观看完整版| 亚洲精品色激情综合| 亚洲欧美日韩高清专用| 好看av亚洲va欧美ⅴa在| 国产三级在线视频| 亚洲性夜色夜夜综合| 久久草成人影院| 欧美乱色亚洲激情| 久久精品夜夜夜夜夜久久蜜豆| 日韩av在线大香蕉| 国产高清videossex| 日韩中文字幕欧美一区二区| 日本精品一区二区三区蜜桃| 狂野欧美白嫩少妇大欣赏| 国产成人精品久久二区二区免费| 成人国产一区最新在线观看| 国产精品日韩av在线免费观看| 欧美黄色淫秽网站| 曰老女人黄片| 国产欧美日韩一区二区三| 男人的好看免费观看在线视频| 亚洲欧美日韩高清在线视频| 日韩人妻高清精品专区| 观看美女的网站| 高清在线国产一区| 老汉色∧v一级毛片| 欧美一区二区国产精品久久精品| 桃色一区二区三区在线观看| 成人高潮视频无遮挡免费网站| 国产蜜桃级精品一区二区三区| 欧美成狂野欧美在线观看| 日韩大尺度精品在线看网址| 男人舔奶头视频| 亚洲黑人精品在线| 99精品在免费线老司机午夜| 2021天堂中文幕一二区在线观| 51午夜福利影视在线观看| 国产av一区在线观看免费| 999精品在线视频| a级毛片a级免费在线| 亚洲av电影不卡..在线观看| 国产亚洲精品久久久久久毛片| 一级a爱片免费观看的视频| 夜夜看夜夜爽夜夜摸| www.999成人在线观看| 国产精品女同一区二区软件 | 99国产综合亚洲精品| 最新美女视频免费是黄的| 动漫黄色视频在线观看| or卡值多少钱| 国产久久久一区二区三区| 一进一出抽搐gif免费好疼| 久久久国产精品麻豆| 色综合欧美亚洲国产小说| 在线观看66精品国产| 19禁男女啪啪无遮挡网站| 最新在线观看一区二区三区| 又黄又爽又免费观看的视频| 欧美黄色片欧美黄色片| 操出白浆在线播放| 成人高潮视频无遮挡免费网站| 男女床上黄色一级片免费看| 欧美丝袜亚洲另类 | 99国产精品99久久久久| 在线视频色国产色| 亚洲精品久久国产高清桃花| 亚洲人成网站高清观看| 久久精品人妻少妇| 亚洲av熟女| 99热这里只有是精品50| 日韩欧美国产一区二区入口| 啪啪无遮挡十八禁网站| 成人欧美大片| 日韩 欧美 亚洲 中文字幕| 亚洲五月婷婷丁香| 在线观看舔阴道视频| 国产欧美日韩一区二区三| 亚洲人成电影免费在线| 欧美乱色亚洲激情| 亚洲av免费在线观看| 国内精品久久久久久久电影| av福利片在线观看| 淫秽高清视频在线观看| 熟女少妇亚洲综合色aaa.| 又紧又爽又黄一区二区| 亚洲熟妇中文字幕五十中出| 精品99又大又爽又粗少妇毛片 | 黑人巨大精品欧美一区二区mp4| 啦啦啦免费观看视频1| 一本久久中文字幕| 国产亚洲精品一区二区www| 国产精品,欧美在线| 很黄的视频免费| 日韩精品中文字幕看吧| 亚洲精品美女久久久久99蜜臀| 亚洲人成电影免费在线| 国产精品乱码一区二三区的特点| 色吧在线观看| 熟女人妻精品中文字幕| 国产黄片美女视频| 久久久久久国产a免费观看| 最新在线观看一区二区三区| 午夜两性在线视频| 一个人免费在线观看的高清视频| 午夜久久久久精精品| 久久久国产欧美日韩av| 麻豆国产av国片精品| 黄色日韩在线| 成人特级av手机在线观看| 国产人伦9x9x在线观看| 久久精品人妻少妇| 日韩高清综合在线| 国产激情欧美一区二区| 人人妻人人澡欧美一区二区| 久久久色成人| 久久久久国产精品人妻aⅴ院| 亚洲精品国产精品久久久不卡| 成年女人毛片免费观看观看9| 一区二区三区激情视频| 天天添夜夜摸| 亚洲欧美日韩高清在线视频| 国产一区二区在线观看日韩 | 男女做爰动态图高潮gif福利片| 18禁黄网站禁片午夜丰满| 夜夜看夜夜爽夜夜摸| 偷拍熟女少妇极品色| 一级毛片女人18水好多| 最新中文字幕久久久久 | 午夜成年电影在线免费观看| 日本与韩国留学比较| tocl精华| 亚洲专区国产一区二区| 制服丝袜大香蕉在线| 高潮久久久久久久久久久不卡| 曰老女人黄片| 色吧在线观看| 欧美日韩亚洲国产一区二区在线观看| 亚洲精品在线观看二区| 99精品久久久久人妻精品| 黑人巨大精品欧美一区二区mp4| 欧美成人性av电影在线观看| 搡老熟女国产l中国老女人| 一二三四社区在线视频社区8| 欧美国产日韩亚洲一区| 人人妻,人人澡人人爽秒播| 欧美色视频一区免费| 中文亚洲av片在线观看爽| 日韩av在线大香蕉| 一本综合久久免费| 国产精品九九99| 天堂网av新在线| av在线天堂中文字幕| 国产综合懂色| 国产伦精品一区二区三区视频9 | 亚洲精品美女久久久久99蜜臀| 在线播放国产精品三级| 女人高潮潮喷娇喘18禁视频| 国产成人av教育| 每晚都被弄得嗷嗷叫到高潮| 俺也久久电影网| 亚洲av五月六月丁香网| 少妇裸体淫交视频免费看高清| 欧美高清成人免费视频www| 国产精品九九99| 久久久精品大字幕| 丰满人妻一区二区三区视频av | 欧美日韩乱码在线| 一区二区三区国产精品乱码| 男女床上黄色一级片免费看| 免费看日本二区| 全区人妻精品视频| 欧美一区二区国产精品久久精品| 婷婷丁香在线五月| 999精品在线视频| 18禁美女被吸乳视频| 亚洲国产精品久久男人天堂| 成年女人看的毛片在线观看| 国产黄a三级三级三级人| 亚洲第一欧美日韩一区二区三区| 18禁黄网站禁片免费观看直播| 国产亚洲欧美98| 国产三级在线视频| 成年免费大片在线观看| 村上凉子中文字幕在线| 桃色一区二区三区在线观看| 男人舔女人的私密视频| 一进一出抽搐动态| 久久精品人妻少妇| 老司机午夜十八禁免费视频| 国产伦精品一区二区三区视频9 | 免费大片18禁| 欧美日韩国产亚洲二区| 搡老妇女老女人老熟妇| 成人永久免费在线观看视频| 叶爱在线成人免费视频播放| 99久久精品一区二区三区| 欧美日韩中文字幕国产精品一区二区三区| 中文字幕久久专区| 久久久久久久午夜电影| 一区二区三区国产精品乱码| 久久久国产精品麻豆| 午夜日韩欧美国产| 黄色视频,在线免费观看| 无限看片的www在线观看| 免费观看人在逋| 亚洲欧美日韩东京热| 国产伦人伦偷精品视频| 搡老熟女国产l中国老女人|