• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    CurveNet: Curvature-Based Multitask Learning Deep Networks for 3D Object Recognition

    2021-06-18 03:36:46MuzahidWanggenWanFerdousSohelLianyaoWuandLiHou
    IEEE/CAA Journal of Automatica Sinica 2021年6期

    A. A. M. Muzahid, Wanggen Wan,, Ferdous Sohel,,Lianyao Wu, and Li Hou

    Abstract—In computer vision fields, 3D object recognition is one of the most important tasks for many real-world applications.Three-dimensional convolutional neural networks (CNNs) have demonstrated their advantages in 3D object recognition. In this paper, we propose to use the principal curvature directions of 3D objects (using a CAD model) to represent the geometric features as inputs for the 3D CNN. Our framework, namely CurveNet,learns perceptually relevant salient features and predicts object class labels. Curvature directions incorporate complex surface information of a 3D object, which helps our framework to produce more precise and discriminative features for object recognition. Multitask learning is inspired by sharing features between two related tasks, where we consider pose classification as an auxiliary task to enable our CurveNet to better generalize object label classification. Experimental results show that our proposed framework using curvature vectors performs better than voxels as an input for 3D object classification. We further improved the performance of CurveNet by combining two networks with both curvature direction and voxels of a 3D object as the inputs. A Cross-Stitch module was adopted to learn effective shared features across multiple representations. We evaluated our methods using three publicly available datasets and achieved competitive performance in the 3D object recognition task.

    I. INTRODUCTION

    IN the field of computer vision, 2D image analysis using deep learning (DL) methods has already achieved remarkable progress and outperforms human vision in many cases (e.g., image classification and human face analysis) [1],[2]. However, understanding three-dimensional objects is still an open research problem of modern computer vision research. A real object in three-dimensional space provides more detailed information. With the availability of low-cost 3D acquisition devices, it is easier to capture 3D objects. The rise of public repositories of 3D models has drawn attention to computer vision research such as 3D object recognition,reconstruction, semantic segmentation, and retrieval [3].

    Convolutional neural network (CNN)-based 3D object recognition systems have advanced considerably [4], but 3D CNNs have been not as successful in identifying 3D objects as 2D CNNs, especially in object label prediction. There are several reasons behind this, e.g., selecting the input features of a 3D object [5] is critical due to its complex geometrical structures, comparatively smaller training databases, and the high computational cost required by 3D CNNs. The earliest volumetric deep learning approach is 3D ShapeNets [6],which deals with three tasks including 3D object classification. Recently, several approaches have been published that solve 3D object recognition tasks using deep CNNs; voxels [6]-[9], point clouds [10]-[12], and 2D multiview [13]-[15] are the most widely used representations of CNNs for 3D object recognition [4]. Two-dimensional multiview-based approaches that use 2D CNNs achieve high performance. This is because existing 2D-based CNNs can be directly used for 3D object recognition, and they require fewer computational resources. However, multiview representations have some technical issues; for example, choosing the number of views to capture the information of the entire 3D object is still an open issue. In addition, the projection of 3D data to the 2D domain discards intrinsic features (e.g., geometric,structural, and orientational information) of a 3D object.

    In 3D shape analysis, the three-dimensional representation is the only way to preserve the intrinsic features of a 3D object; therefore, new features of a 3D object and advanced neural networks need to be explored to improve 3D vision tasks [8]. AI-based computer vision systems are developed with advanced machine learning (ML) algorithms (e.g., deep learning). In addition, the object classification accuracy of a CNN is also highly influenced by the input features [4].

    In this study, we incorporated curvature directions as input features of a 3D object into our novel 3D CNN to identify object labels. Principal curvature directions of 3D objects are considered as perceptually relevant properties of the human visual system (HVS) that are widely used in 3D mesh visual quality evaluation [16], [17]. Considering perceptual features in the HVS, curvature maps represent the salient structural features of a 3D object [17]. The curvature map of a typical 3D CAD model is shown in Figs. 1(d) and 1(e). The plot of curvature vectors looks similar to a sketch of the original 3D model. The idea is that if a typical neural network model can recognize an object from the boundary information (from voxels) of a 3D object, then it can also recognize it from its sketch.

    Fig. 1. The comparison of different 3D representations (a) original 3D Bimba CAD model; (b) voxel representation; (c) point-cloud representation;(d) curvature direction map using maximum values [17]; and (e) curvature direction map using minimum values [17].

    Inspired by this, we propose CurveNet for 3D object classification, which is a novel volumetric CNN that inputs curvature points directly as low-level input features. In CNNs,multitask [18], [19] and multiscale [20], [21] learning-based approaches provide additional benefits that improve the performance of the network for object classification. In multitask learning, an auxiliary task is considered to optimize the loss and improve the performance of the main task. We added pose recognition as an auxiliary task to improve the performance of CurveNet for the object category-level classification. In general, the classification accuracy using curvature points should not be worse than that of voxel input as they provide more information (discriminative surface feature), especially for the curve surface or complex structured 3D model. Our experimental results demonstrate that curvature-based input features outperform the object boundary with voxel representations. A new inception module was introduced to generate a high-level global feature from the low-level surface input feature. Our proposed CurveNet benefits from multiscale learning from this new inception module. To improve the network performance further, we fused the two networks, which take curvature vector points and voxels as inputs. Data augmentation was applied to increase the number of training samples, which helped to reduce the risk of overfitting and improve the performance of the network.

    The key contributions in this paper are as follows: i) We propose to use curvature information of a 3D CAD model as an input for the 3D classifier, CurveNet. To the best of our knowledge, this is the first time that a 3D CNN has directly taken curvature vectors as inputs and outperformed boundary representation from voxels. ii) We introduced a new inception block consisting of three conv layers, and one max pooling layer, which is an efficient alternative to the conventional multiscale learning strategy. This improves the learning capability of CurveNet and generates very high-level global features with fewer parameters. iii) We extended our technique by combining multiple inputs (boundary information from voxels and curvature directions) and experimented with the influence of soft and hard parameter sharing to improve the performance of 3D object classification.

    The rest of the paper is organized as follows: The related work on 3D object recognition based on neural networks is reviewed in Section II. Section III describes the proposed CurveNet architecture including the feature extraction method, network architecture, and network fusion techniques.Network training and the experimental results and analysis are presented in Section IV. Finally, we draw a conclusion in Section V.

    II. RELATED WORKS

    The most common representations of 3D objects used for 3D object recognition are multiview, point clouds, volumetric,hypergraph, and 3D descriptors.

    1) Multiview:One of the earliest approaches to 3D object recognition was presented in the context of RGB-D images[22], [23]. This feature learning approach is similar to image classification, where an additional input is used for the depth information. The multiview convolutional neural network(MVCNN) [13] model is a pioneering multiview-based CNN model, in which a 3D object can be rendered to 2D images from 12 to 80 viewpoints. A pre-trained model from ImageNet[20] was used to train the network model on the ModelNet40 dataset [24] for the 3D classification task. RotationNet [25]was proposed to jointly predict the object label and pose using multiview images of a 3D object. To use the correlation information between the multiview image of a 3D object, Maet al.proposed [26] a novel multiview-based network for classification and retrieval tasks. However, multiview representation discards the geometric properties, and it is hard to determine the number of projections required to capture the entire surface of a 3D object.

    2) Point Cloud:The point cloud representations are popular for their ease of 3D shape representation and their direct usage in CNNs as inputs. PointNet [12] is one of the earliest approaches proposed to address the problems of 3D object classification and segmentation tasks. PointNet directly inputs the pointset and outputs the object class label. However,PointNet does not capture the local structure induced by the pointwise feature extraction method. This problem is addressed in PointNet++ [11] by processing a set of points sampled in a hierarchical order. Ben-Shabatet al.[27]proposed a hybrid point cloud representation, namely 3D modified fisher vectors (3DmFV), which efficiently combines discrete point structures with a continuous generalization of fisher vectors for 3D point cloud classification and part segmentation. Point2Sequence [10] was proposed by Liuet al.as a means of capturing fine-grained contextual information and learning point cloud features by aggregating multi-scale areas of each local region of 3D objects with attention. The relation-shape convolutional neural network (RS-CNN) [28]was introduced to learn contextual shape-aware information for point cloud classification and segmentation problems.However, point clouds are somewhat different from rasterized data (voxels or pixels), which are formed in a spatial irregular fashion, but the main benefit of their coordinates can be used straightforwardly in CNNs as inputs. Due to the advancement of 3D scanning tools, it is very easy and fast to develop a point cloud model, but the model fails to preserve the surface information of the object in the real environment because of the unordered point sampling.

    3) Volumetric:Volumetric representation is a very powerful representation that can describe the full geometry of a 3D object. It holds the intrinsic feature of viewpoint information.Voxel and octree are the most commonly used representation methods of volumetric data for 3D object recognition tasks[3]. Wuet al.proposed a deep learning model for volumetric shape analysis including recognition, namely 3DShapeNets[6], the pioneering and earliest approach to 3D object recognition with volumetric data. The following outlines several fruitful volumetric approaches proposed using 3D voxel grid representation [5], [7], [12], [25]. Recently, octreebased volumetric CNNs [30]-[32] significantly improved the performance of 3D object recognition, consuming less memory compared to voxel-based methods. In addition, there are several new approaches introduced to 3D volumetric CNNs such as multitask learning [33], including sub-volume supervision [9], [14], [32], and network fusion methods with multi-representation input, which have gained noticeable progress [8], [14], [34] in recent years. In general, volumetric CNNs require higher computational cost, which may be challenging for real-time applications. VoxNet and LightNet addressed this issue and introduced volumetric networks with the smallest parameters of about 0.92 M and 0.30 M,respectively. Brocket al.proposed the voxel-based Voxception-ResNet (VRN) [35] model that learned latent space using a variational autoencoder (VAE). The VRN ensemble has achieved the best classification accuracy until now on the ModelNet public dataset [24]. Ghadaiet al.[36]introduced a multi-level 3D CNN that learned multi-scale features by utilizing a coarse voxel grid representation.

    4) Others:Lucianoet al.introduced a deep similarity network fusion framework, namely DeepSNF [37], for 3D object classification, where features from the geometric shape descriptor are fed into a graph convolutional neural network[38]. Recently, several methods were introduced using graph CNNs [38], as these CNNs can process both sparse and unordered data. Among them, dynamic graph [39],hypergraph [40], and multi-hypergraph [41] stereographicbased methods achieve significant improvement in 3D object recognition tasks. These aforementioned methods follow supervised learning strategies; in addition, there are several unsupervised methods [42]-[44] introduced for 3D object recognition. However, the performance of unsupervised approaches is generally worse than that of supervised methods.

    III. PROPOSED METHOD

    We consider a volumetric CNN in this paper. In volumetric data, voxels are commonly used to represent a 3D object.However, to improve the performance and reduce the computational cost of volumetric methods, several approaches have been introduced, such as octree [26], [28], combined voxel with curvature map [45], and normal vector [8], used as an input for a 3D CNN. There is still a large performance gap between 3D and 2D CNNs as 2D-based methods benefit from techniques from numerous existing works in image processing and perform better in 3D object recognition tasks. To reduce this gap, we propose a novel 3D CNN network, called CurveNet, which uses curvature points as inputs. The main reason for using a 3D CNN is to extract rich information from spatial structures and learn the fine details from the non-plane surface at various orientations. This section presents input data preparation, CurveNet architecture, and its learning strategy in detail.

    A. Input Data Preparation

    Two popular ways of representing a 3D object with a CNN are multiview and volumetric. Multiview representation is a 2D image of a 3D object captured from multiple viewpoints that discard the structural information. On the other hand,volumetric representation preserves the intrinsic features with viewpoint information, which is widely used as an input of 3D CNNs.

    We extracted the curvature direction map of a 3D mesh (3D CAD model). The irregular curvature point vectors are formed into a 3D data tensor, which is encoded further to regular points that are input into the 3D CNN. To calculate curvature direction in 2D when applied to a circle of radiusr(Fig. 2), let us define two pointse1ande2on the circle, and their normals aren1andn2respectively from the circle’s center. The curvature of the circle can be calculated as

    The curvature calculation from (1) can be extended on arbitrary 3D meshes by projecting the vectorn1-n2onto the curvature edge as

    Fig. 2. Circular curvature computation in 2D.

    Fig. 3. Geometric elements used for curvature computation for a 3D mesh.

    To extract these curvature vectors (min, max), we used an open-source tool, mepp1https://github.com/MEPP-team/MEPP(3D mesh processing platform), with modified TPDM metrics [17] on Ubuntu. The curvature vectors (3D vectors for three dimensions) are calculated in an iterative manner and stored as unstructured points (3D tensors) on the local drive. To encode these points, we cut 3D tensors of each 3D model with 5k points, and distributed them uniformly by Poisson disk sampling using Meshlab [47] and transformed them into a 3D grid [7] as the input for the CNN.Fig. 1 shows a representation of a 3D CAD model from voxels, point clouds, and curvature points.

    B. CurveNet Network Architecture

    Our proposed CurveNet is a 3D deep convolutional neural network, consisting of conv and FC layers. Fig. 4 depicts the general block diagram of CurveNet. The network is designed in the form of a multitask learning strategy. In multitask learning, at least two or more tasks are assigned [18]. Between them, one is the main task, and the rest of the tasks are auxiliary tasks. Auxiliary tasks are used to improve the performance of the main task. In this work, object label prediction is the main task, and pose prediction is the auxiliary task of CurveNet. CurveNet directly inputs the binary of the 3D data tensor by its input layer, which has a size of 30×30×30. A new inception block was equipped to generate discriminative features using a multiscale training approach.We used three identical inception blocks. Each inception block consisted of three conv layers and one max-pooling layer. Convolution parameters are represented as (C,S) ×N,whereCis the convolution kernel,Sis stride, andNrefers to the number of output feature maps. The output of each conv layer is passed to a sequence of a batch normalization layer, a scale layer, and a leaky rectified linear unit (Leaky ReLU)layer with a negative slope of 0.1.

    We used a typical form of a max-pooling layer. We applied a non-overlapping pooling function using kernel size 2 with a stride of 2, so the resolution of the feature map decreased by a factor of 2. The output of conv-2 and conv-3 is the same with the number of output feature maps that are concatenated to be input into the conv layer of the next inception block.However, conv-2 and conv-3 input features from the output of conv-1 and max-pooling layers, respectively, where the dimensions of two input features are (d/2-1), and (d/2),respectively. These multiscale features are produced by the inception block and sent to the next block as an input, which generates more distinguishable features to recognize object class labels. The number of learned feature maps by Inception-1 is 112, and this is increased twice (224, and 448)by the following Inception blocks. CurveNet uses four FC layers at the top, where the last two FCs, FC-3 and FC-4, are used to get the prediction score of object label and pose,respectively. FC-3 and FC-4 both take the output of FC-2 as an input. We used a softmax classifier to compute the accuracy and loss [48] of CurveNet for object label classification tasks.

    C. Parameter Sharing by a Combination of Two Inputs

    In this work, we employed network fusion strategies on CurveNet using both curvature directions and surface boundary representation from voxels as inputs. To experience the influence of parameter sharing, two CurveNets were fused to improve the classification performance, which takes boundary information from voxels [7] and curvature information as inputs. We trained our CurveNet using three different fusion strategies, namely no sharing, hard parameter sharing, and soft parameter sharing. All fusion networks are illustrated in Fig. 5. Training samples were chosen from the same object with similar volume resolutions for two inputs.The aim was to train a fusion of CurveNets to improve the classification accuracy using multiple inputs and find an effective feature sharing strategy.

    Fig. 4. CurveNet framework. A new Inception module is proposed. Inceptions 1, 2, and 3 are identical but differ with the number of input and output features and parameters (C1 = 3, C2 = 1, C3 = 2, S1 = 2, S2 = 1, and m = 2). The total number of training parameters is 3.1 M.

    Fig. 5. Parameter sharing using network fusion methods with two CurveNets (a) no sharing; (b) hard parameter sharing; and (c) soft parameter sharing.

    The first fusion pipeline is shown in Fig. 5(a). This pipeline is trained with no sharing strategy, and all learned features are concatenated before classification. The classification decision comes from the linear combination of FC outputs. The combination of learned features by two networks is fused before classifiers (topmost FCs). Then, the high-label global features are shared to classify the object into two categories.Our goal was to identify the correct object label (main task).We designed the networks to predict the object pose as an auxiliary task to improve the network performance using orientation information for our desired task [18].

    In the second fusion pipeline, we follow the hard parameter sharing (see Fig. 5(b)) where one or more conv and FC layers can share their learned features. We shared parameters between Inception-2 and Inception-3 layers. We found that sharing all layers including FCs did not help at all to increase the performance but consumed 50 percent more training time.In multitask learning, some layers should be invariant for auxiliary tasks [18], and we found that two separate branches of FCs provided better results than shared FCs.

    We implemented soft-parameter sharing by adopting Cross-Stitch networks [49], as illustrated in Fig. 5(c), where it was determined whether distinct learned features were shared or learned. The weighted parameters of two networks are sharable and trainable. The decision of sharing to the following layers is mediated by α, a matrix of a linear combination of activation maps. The mediated parameter decides whether the features are shared or learned.Accordingly, the network distributes task-related features to improve the prediction scores of the desired task. Finally, all learned features inside the network are forwarded to classifier layers through the final mediator block, which has the ability to select suitable and important features for the specific task(label or pose recognition). In both cases, the final output layer (softmax function) of two networks is averaged, and the largest prediction indicates the target class. All our fusion networks are followed by a layer-based parameter sharing strategy.

    IV. EXPERIMENTS AND ANALYSIS

    A. Training and Evaluation Setup

    We evaluated the classification performance of our proposed CurveNet on two different types of 3D objects,specifically 3D CAD models and LiDAR point clouds using Princeton ModelNet [24] and Sydney Urban objects datasets,respectively. Our CurveNet was implemented on the TensorFlow deep learning framework, which was installed on a Xeon-x5650 CPU with a Teslak20c GPU enabled by cuDNN8.0.

    We trained our CurveNet end-to-end in a supervised manner by stochastic gradient descent (SGB) [50] with the momentum parameter set to 0.9. The batch size was set to 32. Initial learning was set to 0.01 and decreased by a factor of 10. We added dropout regularization after each output layer with a ratio of 0.6. We computed the multiclass loss function using multinomial cross-entropy losses [51]. Our CurveNet took approximately 26 hours to be trained for 90 epochs. We applied voting strategy during testing where multiple rotations of a sample 3D object were fed to the network, and outputs were averaged to get the final prediction of the object class label.

    B. Data Alignment and Augmentation

    Modelnet40 and ModelNet10 are the two main datasets widely used to evaluate classification performance for 3D objects. However, the 3D models in ModelNet10 are aligned perfectly; in contrast, more than 1000 models in ModelNet40 are not aligned. We first sorted the classes of non-aligned objects and prepared a reference set for each class to be aligned. We adopted an unsupervised generation of viewpoint[52] methods to align objects according to class, and labeling of object pose was assigned iteratively by running three different subsets using 90-degree rotations. Some rotationally invariant objects (e.g., samples in flower pots and bottles)were forcibly assigned to class 1 as they did not provide any significant information in different poses.

    The visual attribute of a 3D object is influenced by the object pose and the number of samples per class. We extended the dataset samples by rotating each 3D model around the horizontal direction along thez-axis. We prepared two augmented subsets of the original datasets. We generated 12 and 24 copies of each 3D model by rotating by intervals of 30 and 15 degrees for two subsets, respectively. This augmentation procedure has been applied for both ModelNet40 and ModelNet10 datasets. This augmentation method helps to increase the number of training samples of each class (both pose and category) and helps to improve the performance of the network.

    C. Classification Performance on ModelNet Datasets

    The ModelNet datasets were built especially to evaluate 3D object classifiers by Changet al.[24] in 2015. ModelNet40 is a large-scale dataset containing 12311 3D CAD models that are classified into 40 categories. ModelNet10 is a small dataset that is a subset of ModelNet40, containing 4899 3D CAD models in 10 categories. The datasets come with their own training and testing splits. For consistency, we used the original splits from the datasets in our experiments. During the test, we sent all augmented models of a test sample to the network in one batch, and the object category was decided by averaging the activations of the output layer.

    1) Classification Results:We evaluated our model using both surface boundary information from voxels and curvature information. The classification accuracy (average class) is presented in Table I. We found that the normalized maximum curvature direction vector showed superiority over voxels.Our proposed method improves the classification accuracy by 1.2% compared to surface boundary information using the same network. It also outperforms 3DShapeNet [6] and VoxNet [7], which are state-of-the-art volumetric methods for 3D object classification. The improvements are mainly the result of two factors: curvature information as an input feature and auxiliary learning by adding pose classification. Curvature provides considerable depth and perceptual features to the CNN, and auxiliary pose learning helps to improve object classification accuracy, which is the main task of our network.In addition, the voting method is another common approach to improving the performance of the classifier. Table II shows the influence of pose learning and voting methods on improving the classification accuracy using our CurveNet. We used two augmented datasets from ModelNet40 using 12 and 24 copies of each sample. Experimental results show that classification accuracy is improved by 3.8% using 12-view and 3.2% using 24-view while using both the pose learning and voting approach.

    The best performance was achieved using 24-view, with an average classification accuracy of 90.2% on ModelNet40.

    From these experiments, we found that the number of training samples has an effect on the classification performance, i.e., the class with a large number of training samples has high accuracy. However, high interclass similarity between two classes can lead to a slight deviation in this trend, e.g., Chair vs. Stool and Dresser vs. Nightstand (seeFig. 6). Similar observations have also been reported in earlier works [30], [32]. When we visualize the confusion matrix of ModelNet40 in Fig. 7, it illustrates that this ambiguity only occurs if the target object belongs to some similar categories of objects, e.g., Dresser, Flower Pot, Sink, and Vase, which are examples of most confused objects in the database as they have close visual properties in instances where our network misclassified some of them.

    TABLE I COMPARISON OF CLASSIFICATION ACCURACY OF OUR CURVENET AND STATE-OF-THE-ART-METRICS ON MODELNET40

    TABLE II THE INFLUENCE OF POSE CLASSIFICATION AND VOTING METHODS ON CLASSIFYING OBJECT LABELS ON MODELNET40

    Fig. 6. Classification accuracy per class on ModelNet40 dataset.

    2) Parameter Sharing:To improve the classification performance further, both voxel and curvature features were fused using our proposed volumetric model. We applied three methods of parameter sharing between two networks, which were no sharing, hard sharing, and soft sharing, as shown in Fig. 5. Both networks (for voxel and curvature) were pretrained independently to reduce the training time.

    In the case of no sharing, generated features from different layers were concatenated before the last FC (classifier) layer.Moreover, for hard sharing, Inception-1 and Inception-2 layers were shared, then shared features were sent to FC layers consecutively. For channel-wise attention, the generated features across all layers were shared using a crossstitch algorithm (see [55] for details).

    Overall, the classification results of fusion networks were improved compared to a single network. The results are presented in Table III. We find that soft sharing has a great impact on achieving good results. It achieved average class and sample accuracy of 90.7% and 93.5%, respectively.Moreover, hard sharing performs slightly better than no sharing, as the average class and sample accuracy increased by 0.2% and 0.6%, respectively.

    Fig. 7. The confusion matrix of the ModelNet40 dataset shows the similarity between a pair of 3D objects; lower diagonal values indicate a higher degree of similarity between the objects.

    The experimental result indicates that the network with soft sharing benefited from mediated features (see Section III-C),by which it was decided whether the parameters needed to be shared or learned. On the other hand, sharing all parameters by hard sharing does not help to improve the classification result significantly; rather, using too many combinations of parameter coefficients may produce a more confusing scorethat may lead the network to a locally optimal solution.

    TABLE III PERFORMANCE EVALUATION OF CURVENET WITH SHARING FEATURES BY COMBINING CURVATURE AND BOUNDARY INFORMATION FROM VOXELS ON MODELNET40 DATABASE

    3) Comparison With Other Methods:The classification accuracy of our model is compared with state-of-the-art methods in Table IV. We found that Multiview and pointcloud methods still dominated in the 3D object classification task. In addition, our method achieved the highest accuracy among volumetric methods but was inferior to the VRN ensemble [35]. Our CurveNet has 3.1 M parameters, whereas fusion contains about 6.2 M. In comparison to the 90 M parameters of the VRN ensemble, our network is lighter, and the performance is comparable to other methods. In addition,our network also performed well on a low scale ModelNet10 database and achieved an average classification accuracy of 94.2%, which is also comparable to other methods.Theoretically, deep networks should perform better on large datasets as they need more training samples. However,samples in ModelNet10 are built with some selected classes with clean objects from ModelNet40. The confusion matrix for ModelNet10 is shown in Fig. 8, where higher diagonal values exhibit fewer similarities with other objects. In comparison to ModelNet40, all classes in ModelNet10 obtained a higher confusion score, which implies a lower similarity between the objects, resulting in better classification accuracy achieved over the ModelNet40 dataset.

    D. Sydney Urban Objects Dataset

    The Sydney Urban Objects dataset2http://www.acfr.usyd.edu.au/papers/SydneyUrbanObjectsDataset.shtml.consists of labeled Velodyne LiDAR scans of 631 urban road objects (point cloud samples) in 26 distinct categories including vehicles,pedestrians, signs, and trees. As it provides single 3D point cloud samples scanned on a common scenario of urban road,we considered this dataset to be of special interest in terms of object classification and observing the performance of our CurveNet. Because of unordered point-samples, we converted the points to volumetric representation [7] by putting all points around the objects in a bounding box and encoding them to a 303occupancy grid voxel. We empirically set the voxel size to 0.2 m. Data augmentation and test voting wereapplied accordingly with 24 rotations around thez-axis.

    TABLE IV THE COMPARISON OF CLASSIFICATION ACCURACY ACHIEVED BY OUR CURVENET AND OTHER METHODS ON MODELNET40 AND MODELNET10 DATASETS

    Fig. 8. Confusion matrix of ModelNet40 dataset.

    Classification Results:To compare the classification result with state-of-the-art methods, we followed the original dataset’s protocol [56]. The dataset was folded to four standard training and testing splits for a subset of 588 samples in 14 classes. This dataset was smaller than ModelNet;therefore, we still kept a drop-out connection of 0.6 to mitigate the overfitting during training. We report the classification performance by the averageF1score. We present the classification performance of different methods in Table V. Our model obtained an averageF1of 79.3%. The classification performance of CurveNet is significantly better than those of volumetric methods such as VoxNet [7], ORION[33], and NormalNet [8] but is slightly worse than that of LightNet [9]. This improvement implies that the classification performance on a low scale dataset is also influenced by combined multiscale features using our novel inception module and auxiliary learning on orientation information.

    TABLE V CLASSIFICATION ACCURACY OF DIFFERENT VOLUMETRIC METHODS ON SYDNEY URBAN OBJECTS DATASET

    V. CONCLUSION

    In this paper, we introduce CurveNet, which uses a curvature direction vector with a 3D CNN for 3D object recognition. Our framework takes into account multitask(label and pose) and multiscale learning (in an inception module) strategies. We applied network fusion and data augmentation methods to improve the recognition rate. Using the relevant perceptual features, our proposed CurveNet performed better than existing voxel representation models.The results are competitive with state-of-the-art point-cloudbased and multiview-based methods. In contrast, our model only works for 3D triangle mesh (3D CAD models). A weakness of our model is that it is not suitable for other 3D data representations (e.g., point clouds).

    国产精华一区二区三区| 亚洲欧美成人精品一区二区| 最近视频中文字幕2019在线8| 你懂的网址亚洲精品在线观看 | 99热这里只有是精品在线观看| 久久久久久国产a免费观看| 久久久久久九九精品二区国产| 日韩高清综合在线| 亚洲精品久久久久久婷婷小说 | 国产综合懂色| 春色校园在线视频观看| 少妇丰满av| 最近中文字幕2019免费版| 亚洲av不卡在线观看| 嫩草影院新地址| 亚洲激情五月婷婷啪啪| 国产精品一二三区在线看| 深爱激情五月婷婷| av免费观看日本| 伊人久久精品亚洲午夜| 又爽又黄a免费视频| 免费观看的影片在线观看| 男女国产视频网站| 2021天堂中文幕一二区在线观| 啦啦啦观看免费观看视频高清| 久久国内精品自在自线图片| 久久久久精品久久久久真实原创| 少妇高潮的动态图| 男人的好看免费观看在线视频| 亚洲精品乱久久久久久| 国产女主播在线喷水免费视频网站 | 亚洲精品自拍成人| 国产高清三级在线| 国产精品久久久久久精品电影小说 | 精品久久久久久久久久久久久| 亚洲欧美中文字幕日韩二区| 免费看美女性在线毛片视频| 国产亚洲5aaaaa淫片| 日韩在线高清观看一区二区三区| 如何舔出高潮| 舔av片在线| 男女下面进入的视频免费午夜| www.av在线官网国产| 国产在线一区二区三区精 | 国产乱来视频区| 亚洲国产色片| 女人十人毛片免费观看3o分钟| 成人特级av手机在线观看| 午夜a级毛片| 免费观看a级毛片全部| 国产一级毛片七仙女欲春2| 免费观看精品视频网站| 亚洲不卡免费看| 国产精品精品国产色婷婷| 国产精品一及| 日韩大片免费观看网站 | 校园人妻丝袜中文字幕| 精品久久久久久久久av| 婷婷色综合大香蕉| 国产av码专区亚洲av| 乱码一卡2卡4卡精品| 久99久视频精品免费| 18禁在线播放成人免费| 亚洲精品日韩在线中文字幕| 中文字幕久久专区| 毛片女人毛片| 亚洲自偷自拍三级| 欧美又色又爽又黄视频| 少妇熟女aⅴ在线视频| 如何舔出高潮| 国产精品美女特级片免费视频播放器| 欧美一区二区亚洲| 亚洲精品aⅴ在线观看| 可以在线观看毛片的网站| 免费在线观看成人毛片| 岛国在线免费视频观看| 亚洲一级一片aⅴ在线观看| 欧美最新免费一区二区三区| 国产高清三级在线| 99久久精品国产国产毛片| 国产伦精品一区二区三区视频9| 免费无遮挡裸体视频| 中文字幕制服av| 午夜老司机福利剧场| 长腿黑丝高跟| 欧美日本亚洲视频在线播放| 黄色一级大片看看| 真实男女啪啪啪动态图| 国产麻豆成人av免费视频| av视频在线观看入口| 国产亚洲91精品色在线| 亚洲精品成人久久久久久| 午夜福利在线观看免费完整高清在| 精品国内亚洲2022精品成人| 韩国高清视频一区二区三区| 91精品伊人久久大香线蕉| 高清日韩中文字幕在线| 亚洲精品久久久久久婷婷小说 | 亚洲性久久影院| 能在线免费观看的黄片| 国产黄片视频在线免费观看| 久久亚洲精品不卡| 国内少妇人妻偷人精品xxx网站| www.av在线官网国产| 国产午夜精品论理片| 国产又黄又爽又无遮挡在线| 在线免费观看不下载黄p国产| 一级毛片久久久久久久久女| 日韩大片免费观看网站 | 国产成年人精品一区二区| 国产极品天堂在线| 人体艺术视频欧美日本| 国产亚洲精品av在线| 国产在视频线在精品| 亚洲一区高清亚洲精品| 国产片特级美女逼逼视频| 好男人视频免费观看在线| 99久久无色码亚洲精品果冻| 国产亚洲最大av| 亚洲av.av天堂| 黄片wwwwww| 国产亚洲5aaaaa淫片| 亚洲国产高清在线一区二区三| 午夜福利网站1000一区二区三区| 2022亚洲国产成人精品| 身体一侧抽搐| 成人三级黄色视频| 亚洲精品日韩av片在线观看| av在线亚洲专区| 十八禁国产超污无遮挡网站| 久久精品国产亚洲网站| 久久国内精品自在自线图片| 一级二级三级毛片免费看| 少妇的逼好多水| 五月伊人婷婷丁香| 神马国产精品三级电影在线观看| 日韩,欧美,国产一区二区三区 | 日产精品乱码卡一卡2卡三| 欧美三级亚洲精品| 熟妇人妻久久中文字幕3abv| 国产日韩欧美在线精品| 少妇丰满av| 亚洲精品亚洲一区二区| 美女脱内裤让男人舔精品视频| 国产免费男女视频| 大又大粗又爽又黄少妇毛片口| 亚洲天堂国产精品一区在线| 欧美一区二区国产精品久久精品| 亚洲国产欧美在线一区| 国产精品人妻久久久影院| 国产精品,欧美在线| 久久婷婷人人爽人人干人人爱| 免费播放大片免费观看视频在线观看 | 狂野欧美白嫩少妇大欣赏| 日本欧美国产在线视频| 国产成人a∨麻豆精品| 久久6这里有精品| 成人无遮挡网站| 男插女下体视频免费在线播放| 欧美色视频一区免费| 午夜精品在线福利| 国产麻豆成人av免费视频| 国产精华一区二区三区| 一本久久精品| 六月丁香七月| 校园人妻丝袜中文字幕| 亚洲精品国产成人久久av| 国产精品,欧美在线| 国产精品人妻久久久影院| 国产精品一及| 久久久久久伊人网av| 国产一区二区在线观看日韩| 91aial.com中文字幕在线观看| 久久欧美精品欧美久久欧美| 欧美又色又爽又黄视频| 我的老师免费观看完整版| 国产毛片a区久久久久| 两性午夜刺激爽爽歪歪视频在线观看| 国产精品野战在线观看| 啦啦啦观看免费观看视频高清| 国产成人a区在线观看| 日本三级黄在线观看| 亚洲av不卡在线观看| 久久久国产成人精品二区| 麻豆成人午夜福利视频| 久久久久网色| 国产精品久久久久久精品电影小说 | av福利片在线观看| 看十八女毛片水多多多| 亚洲性久久影院| 久久鲁丝午夜福利片| 免费av不卡在线播放| 亚洲中文字幕日韩| 日本免费在线观看一区| 插逼视频在线观看| 插逼视频在线观看| 国语自产精品视频在线第100页| 国产午夜精品论理片| 国产精品不卡视频一区二区| 国产精品人妻久久久久久| 午夜视频国产福利| 精品久久久久久久人妻蜜臀av| 成人一区二区视频在线观看| 高清在线视频一区二区三区 | 美女高潮的动态| 亚洲一区高清亚洲精品| 非洲黑人性xxxx精品又粗又长| 一本一本综合久久| 一区二区三区乱码不卡18| 69人妻影院| 中文精品一卡2卡3卡4更新| 七月丁香在线播放| 爱豆传媒免费全集在线观看| av免费在线看不卡| 久久精品久久久久久噜噜老黄 | 视频中文字幕在线观看| 日本猛色少妇xxxxx猛交久久| 亚洲怡红院男人天堂| 日本午夜av视频| av免费观看日本| 少妇人妻一区二区三区视频| 只有这里有精品99| 国产高潮美女av| av.在线天堂| 黄色日韩在线| 精品酒店卫生间| 久久亚洲国产成人精品v| 麻豆一二三区av精品| 亚洲不卡免费看| 久久久久久久久久久免费av| 国内揄拍国产精品人妻在线| 日本色播在线视频| 亚洲乱码一区二区免费版| 国产精品综合久久久久久久免费| 亚洲精品乱码久久久久久按摩| 只有这里有精品99| 亚洲精品亚洲一区二区| 2022亚洲国产成人精品| 国产免费一级a男人的天堂| 精品国内亚洲2022精品成人| 欧美三级亚洲精品| 久久久久精品久久久久真实原创| 国产一区二区在线av高清观看| 最近手机中文字幕大全| 亚洲天堂国产精品一区在线| 色网站视频免费| 在线观看一区二区三区| 男女边吃奶边做爰视频| 亚洲成人久久爱视频| 免费黄色在线免费观看| 中文字幕亚洲精品专区| 美女高潮的动态| a级毛片免费高清观看在线播放| 最近中文字幕高清免费大全6| 亚洲欧美清纯卡通| 国产69精品久久久久777片| 国产又黄又爽又无遮挡在线| 国内精品一区二区在线观看| 纵有疾风起免费观看全集完整版 | 久久6这里有精品| 国产真实乱freesex| 级片在线观看| 国产毛片a区久久久久| 人体艺术视频欧美日本| 日本色播在线视频| 99热网站在线观看| 最近中文字幕高清免费大全6| 久99久视频精品免费| 如何舔出高潮| 成人欧美大片| 人人妻人人澡人人爽人人夜夜 | 精品久久久久久久久久久久久| 日本与韩国留学比较| 美女高潮的动态| 午夜精品国产一区二区电影 | 亚洲色图av天堂| 欧美精品一区二区大全| 永久网站在线| 久久这里只有精品中国| 国产高清国产精品国产三级 | 我的女老师完整版在线观看| 噜噜噜噜噜久久久久久91| 最近手机中文字幕大全| 久热久热在线精品观看| 99热精品在线国产| 中文亚洲av片在线观看爽| 国产精品福利在线免费观看| 久久99精品国语久久久| 美女国产视频在线观看| 日韩成人av中文字幕在线观看| 插阴视频在线观看视频| 午夜福利成人在线免费观看| 免费一级毛片在线播放高清视频| 少妇熟女aⅴ在线视频| 永久网站在线| 久久这里只有精品中国| 美女xxoo啪啪120秒动态图| av卡一久久| 美女高潮的动态| 亚洲av成人精品一区久久| 三级男女做爰猛烈吃奶摸视频| 久久精品夜夜夜夜夜久久蜜豆| 深爱激情五月婷婷| 欧美xxxx性猛交bbbb| 午夜久久久久精精品| av在线亚洲专区| 永久免费av网站大全| 精品人妻视频免费看| АⅤ资源中文在线天堂| 91久久精品国产一区二区成人| 国产在线一区二区三区精 | 亚洲乱码一区二区免费版| av免费在线看不卡| 一二三四中文在线观看免费高清| 乱系列少妇在线播放| 晚上一个人看的免费电影| h日本视频在线播放| 男女下面进入的视频免费午夜| 国产一区二区在线av高清观看| 国产免费视频播放在线视频 | 免费看美女性在线毛片视频| 一区二区三区乱码不卡18| 亚洲人成网站在线播| 伦精品一区二区三区| 2021天堂中文幕一二区在线观| 五月玫瑰六月丁香| 一边摸一边抽搐一进一小说| 欧美成人免费av一区二区三区| 成人综合一区亚洲| av.在线天堂| 国产亚洲午夜精品一区二区久久 | 可以在线观看毛片的网站| 全区人妻精品视频| 搡女人真爽免费视频火全软件| 视频中文字幕在线观看| 韩国高清视频一区二区三区| kizo精华| 日韩av在线免费看完整版不卡| 日韩三级伦理在线观看| 久久精品影院6| 亚洲精品国产av成人精品| 高清在线视频一区二区三区 | 国产免费一级a男人的天堂| 美女国产视频在线观看| 狠狠狠狠99中文字幕| 成人av在线播放网站| 欧美成人午夜免费资源| 国产免费又黄又爽又色| 伊人久久精品亚洲午夜| 日产精品乱码卡一卡2卡三| 亚洲av免费在线观看| 亚洲五月天丁香| 干丝袜人妻中文字幕| 免费一级毛片在线播放高清视频| 午夜福利在线观看免费完整高清在| 日本wwww免费看| 97热精品久久久久久| 国产在线一区二区三区精 | 我的老师免费观看完整版| 国产伦精品一区二区三区视频9| 国产视频首页在线观看| 国产成人aa在线观看| 精品久久久噜噜| 我的老师免费观看完整版| 亚洲精品乱久久久久久| 国产成人a∨麻豆精品| 久久久国产成人精品二区| 91狼人影院| 日韩精品青青久久久久久| 国产视频首页在线观看| 亚洲四区av| 在线播放国产精品三级| 精品酒店卫生间| 免费电影在线观看免费观看| 亚洲va在线va天堂va国产| 最近最新中文字幕大全电影3| 熟女电影av网| 七月丁香在线播放| 精品欧美国产一区二区三| 日本一本二区三区精品| 最近中文字幕2019免费版| 26uuu在线亚洲综合色| 久久综合国产亚洲精品| 能在线免费观看的黄片| 国产成人精品一,二区| 老司机影院毛片| 热99在线观看视频| 少妇裸体淫交视频免费看高清| 久久草成人影院| 久久精品国产亚洲av涩爱| 2021天堂中文幕一二区在线观| 18禁在线播放成人免费| 成人二区视频| 欧美激情国产日韩精品一区| 91午夜精品亚洲一区二区三区| 国产黄a三级三级三级人| 午夜激情欧美在线| 看黄色毛片网站| 日韩精品青青久久久久久| 国产黄色小视频在线观看| 一级av片app| 女人久久www免费人成看片 | av国产久精品久网站免费入址| 亚洲国产精品久久男人天堂| 国内精品一区二区在线观看| 精品一区二区免费观看| 男插女下体视频免费在线播放| 午夜精品一区二区三区免费看| 久久久久久九九精品二区国产| 成人特级av手机在线观看| 久久99热6这里只有精品| 国产午夜精品论理片| 97人妻精品一区二区三区麻豆| av在线播放精品| 99久久人妻综合| 成人毛片a级毛片在线播放| 成年女人永久免费观看视频| 欧美xxxx性猛交bbbb| 欧美+日韩+精品| 26uuu在线亚洲综合色| 日本欧美国产在线视频| 午夜视频国产福利| 自拍偷自拍亚洲精品老妇| 一个人免费在线观看电影| 99久久人妻综合| 日韩人妻高清精品专区| 国产精品日韩av在线免费观看| 99热这里只有是精品在线观看| 日韩成人伦理影院| 亚洲国产精品sss在线观看| 久久午夜福利片| 99久久精品一区二区三区| 亚洲成人中文字幕在线播放| 18禁在线无遮挡免费观看视频| 亚洲国产最新在线播放| 国产精品野战在线观看| 国产免费一级a男人的天堂| АⅤ资源中文在线天堂| 久99久视频精品免费| 免费av观看视频| 日韩精品青青久久久久久| 午夜亚洲福利在线播放| 亚洲精品一区蜜桃| 亚洲中文字幕一区二区三区有码在线看| 麻豆国产97在线/欧美| 秋霞在线观看毛片| 网址你懂的国产日韩在线| av又黄又爽大尺度在线免费看 | 亚洲精品国产av成人精品| 精品久久久久久久人妻蜜臀av| 日本五十路高清| 丝袜美腿在线中文| 久久久久久久亚洲中文字幕| 边亲边吃奶的免费视频| 噜噜噜噜噜久久久久久91| 嘟嘟电影网在线观看| 人体艺术视频欧美日本| 国产一区二区亚洲精品在线观看| 中文字幕亚洲精品专区| 寂寞人妻少妇视频99o| av在线天堂中文字幕| 天堂影院成人在线观看| 亚洲精品久久久久久婷婷小说 | 亚洲欧美清纯卡通| 女人久久www免费人成看片 | 亚洲人成网站高清观看| 91aial.com中文字幕在线观看| 美女cb高潮喷水在线观看| 欧美+日韩+精品| 国产欧美另类精品又又久久亚洲欧美| 人妻少妇偷人精品九色| 中国美白少妇内射xxxbb| 午夜a级毛片| 国产黄色视频一区二区在线观看 | 狂野欧美激情性xxxx在线观看| 别揉我奶头 嗯啊视频| 精品一区二区三区人妻视频| 色哟哟·www| 特大巨黑吊av在线直播| 免费av不卡在线播放| 91久久精品国产一区二区三区| 精品99又大又爽又粗少妇毛片| 欧美日韩综合久久久久久| 亚洲国产欧美人成| 久热久热在线精品观看| 人体艺术视频欧美日本| 91久久精品国产一区二区三区| 内地一区二区视频在线| 国产免费福利视频在线观看| 欧美性猛交╳xxx乱大交人| 午夜视频国产福利| 国产精品一及| 国产麻豆成人av免费视频| 黄色一级大片看看| 久久人人爽人人爽人人片va| 成人二区视频| 久久国产乱子免费精品| 啦啦啦啦在线视频资源| 边亲边吃奶的免费视频| 舔av片在线| 亚洲在线自拍视频| 亚洲国产欧洲综合997久久,| 91精品国产九色| 亚洲真实伦在线观看| 最近的中文字幕免费完整| 精品人妻偷拍中文字幕| 好男人在线观看高清免费视频| 欧美日韩综合久久久久久| 亚洲va在线va天堂va国产| 亚洲中文字幕日韩| 18禁裸乳无遮挡免费网站照片| 精品国产露脸久久av麻豆 | 国产精品无大码| 国产精华一区二区三区| 亚洲性久久影院| 亚洲欧洲日产国产| 久久久a久久爽久久v久久| 亚洲欧美成人综合另类久久久 | 永久网站在线| 18禁在线播放成人免费| 国产精品电影一区二区三区| 欧美bdsm另类| 午夜激情欧美在线| 你懂的网址亚洲精品在线观看 | 亚洲最大成人中文| 中文字幕熟女人妻在线| 国产在线一区二区三区精 | 亚洲婷婷狠狠爱综合网| 国产成人freesex在线| 中文字幕熟女人妻在线| 国产在线男女| 中文精品一卡2卡3卡4更新| 国产高清三级在线| 日韩大片免费观看网站 | 国产午夜精品一二区理论片| 五月玫瑰六月丁香| 国产亚洲av片在线观看秒播厂 | 欧美变态另类bdsm刘玥| 我的女老师完整版在线观看| 久久草成人影院| 天天躁夜夜躁狠狠久久av| 久久久久九九精品影院| 日韩国内少妇激情av| 欧美一区二区国产精品久久精品| 亚洲精品乱码久久久v下载方式| 欧美又色又爽又黄视频| 国内精品宾馆在线| 国语自产精品视频在线第100页| 激情 狠狠 欧美| 99热这里只有精品一区| 男人舔奶头视频| 男人和女人高潮做爰伦理| 国内精品一区二区在线观看| 淫秽高清视频在线观看| 久久久精品大字幕| 亚洲怡红院男人天堂| 女人被狂操c到高潮| 高清午夜精品一区二区三区| 一夜夜www| 啦啦啦韩国在线观看视频| 22中文网久久字幕| 国产黄色小视频在线观看| 欧美潮喷喷水| 免费不卡的大黄色大毛片视频在线观看 | 国产v大片淫在线免费观看| 身体一侧抽搐| videossex国产| 久久久久免费精品人妻一区二区| 一边亲一边摸免费视频| 99久久人妻综合| 91aial.com中文字幕在线观看| 国产美女午夜福利| 国产高清不卡午夜福利| 99国产精品一区二区蜜桃av| 亚洲国产日韩欧美精品在线观看| 日韩大片免费观看网站 | 国产欧美日韩精品一区二区| 婷婷六月久久综合丁香| 久久精品久久久久久久性| 久久精品人妻少妇| 麻豆成人午夜福利视频| 亚洲精品国产成人久久av| 日韩亚洲欧美综合| 1000部很黄的大片| 人妻少妇偷人精品九色| 在线观看av片永久免费下载| 麻豆乱淫一区二区| 97人妻精品一区二区三区麻豆| 久热久热在线精品观看| 3wmmmm亚洲av在线观看| 亚洲av不卡在线观看| 日本色播在线视频| 久久精品国产99精品国产亚洲性色| 久久欧美精品欧美久久欧美| 99热全是精品| 汤姆久久久久久久影院中文字幕 | 日本与韩国留学比较| 麻豆av噜噜一区二区三区| 日本与韩国留学比较| 99热6这里只有精品| 国产极品天堂在线| 国产单亲对白刺激| 美女大奶头视频| 女的被弄到高潮叫床怎么办| 国产三级在线视频| 日本黄色片子视频| 国产精品福利在线免费观看| 久久久久性生活片| 国产真实伦视频高清在线观看| 久久精品国产99精品国产亚洲性色| 成人三级黄色视频| 久久久国产成人精品二区| 日本午夜av视频| 18禁动态无遮挡网站| 五月伊人婷婷丁香|