1. Introduction
With improvements in the performance of point cloud data acquisition equipment such as LiDAR, point clouds have become increasingly widely used in the fields of robotic automated driving and virtual reality, among others. It has become one of the most important data formats in 3D representation, and has been widely used in tasks such as object classification [1,2,3,4], segmentation [2,4,5], pose estimation [6,7], object recognition [8], and object detection [9,10].
Point cloud processing technology is also widely used in extended reality fields such as virtual, augmented, and mixed reality. Extended reality technologies represent a paradigm that enhances and supports Industry 4.0 in diverse settings [11,12]. Digital twins are one of the disruptive technologies associated with the Industry 4.0 concept. Combining the advanced point cloud processing algorithms with cameras and sensors [13] will facilitate the development of Industry 4.0 and related applications [14,15].
There are three typical representations of 3D data: voxels [16], meshes [17,18], and point clouds [19]. A voxel-based representation can apply a traditional convolutional neural network (CNN) to 3D data. However, as the resolution increases, the storage and computing resource consumption of the voxel method significantly increases. Therefore, it is not suitable for high-resolution point cloud reconstruction. Compared with a voxel, a point cloud is a simpler and more unified structure; it can represent 3D shapes more efficiently and is easier to manipulate when geometric transformations are performed.
Real-world point cloud data are usually incomplete. For example, owing to occlusion or interference, the point cloud data scanned by LiDAR are partially incomplete, resulting in the loss of geometric information of the objects. The incompleteness of point cloud data affects further processing. Therefore, converting a partial point cloud into a complete point cloud is of great value for downstream applications such as classification, segmentation, and object detection.
The difficulty in processing point clouds is that the point cloud is disordered and rotationally invariant; therefore, it is difficult to apply traditional convolution operations to point clouds. PointNet [2] and PointNet++ [4], proposed by Qi et al., provide solutions to the point cloud disorder problem. They directly operate on the point cloud for classification and segmentation, avoiding the loss of information caused by the point cloud during data format conversion. FoldingNet [20] contains an auto-encoder where the encoder part can extract the global features of the point cloud and the decoder part can recover data of the original point cloud as accurately as possible from the global features. These two studies laid the foundation for a point cloud completion network. Yuan et al. [21] also adopted an encoder–decoder architecture. The difference is that their decoder adopts a two-stage generation framework to generate a detailed point cloud. The aforementioned networks directly output the complete point cloud; however, the unoccluded parts need not be generated by the network. In addition, the decoder of FoldingNet folds 2D lattices into 3D shapes, which are more difficult to learn and train. In recent years, transformers [22] have achieved excellent results in the fields of natural language processing and computer vision [23]. Inspired by this, Zhao et al. [24] applied transformers to point cloud scenes and proposed point transformers. Point transformers have demonstrated excellent performance in tasks such as classification and segmentation. However, the encoder of most completion networks adopts a multilayer perceptron (MLP) or a similar architecture, and the feature extraction ability is limited.
In response to the above problems, we designed several improvements to the existing networks. The main contributions of this study are as follows:
(1). We think that the unoccluded part of the point cloud does not need to be generated by the network; hence, our network only predicts the occluded part and then stitches the output of the network with the unoccluded part into a complete point cloud of the shape.
(2). We replaced the 2D lattice in the FoldingNet decoder with a 3D lattice and directly deformed the three-dimensional point cloud into a point cloud of the occluded part. This can simplify network training and improve network performance.
(3). The feature extraction capability of the MLP encoder is limited, and to improve it, we used a transformer module as the encoder of our completion network.
The Section 2 of this paper introduces the related work of point cloud completion, the Section 3 introduces our network model and loss function in detail, the Section 4 introduces the specific implementation and results of the experiment, the Section 5 discusses possible further improvements for the network, and the Section 6 summarizes the study.
2. Related Work
Point cloud completion methods can be divided into two categories: traditional and learning-based point cloud completion methods. Traditional methods include geometry- and template-based methods. Learning-based methods mainly use encoder–decoder architecture networks or multisegment generation networks.
2.1. Traditional Completion Methods
Geometry-based methods use information from incomplete input shapes to obtain complete shapes. It needs the geometric properties of the shape, such as the continuity of the surface and the symmetry of the shape. Surface-oriented methods [25,26] employ smooth interpolation to fill incomplete holes on the surfaces of the shape. Symmetry-based methods [7,27] first identify the symmetry axis and recurrent structures of the shape and then copy the shape of the unoccluded part to the missing part. These methods require that the missing parts can be inferred from the unoccluded parts; therefore, they are only suitable for data that are not severely occluded. However, real-world data are often severely occluded, which sometimes makes these methods ineffective. Model-based methods complete shapes by matching incomplete input shapes to models in large databases. The direct retrieval method [28,29] directly matches the input with the model in the database as the final result of the completion. Partial retrieval methods [17,30] divide the input into several parts to match the models in the database and then combine the matching results to generate the final completion result. Deform-based methods [31,32] deform retrieved shapes to obtain shapes that better match the input. The geometric primitive method [33,34] uses geometric primitives instead of large databases and matches the input with geometric primitives to synthesize the final shape.
The advantage of the traditional method is that it is easy to implement with a simple algorithm. The disadvantage is that when the incomplete area of the input point cloud is too large, the geometry of the missing area cannot be estimated.
2.2. Learning-Based Methods
Learning-based methods use neural networks and large amounts of data for shape completion. Some studies [35,36] represented shapes as voxels, and generalized traditional 2D convolution to 3D convolution. The PointNet [2] and PointNet++ [4] networks solve the problems caused by the disorder and rotation invariance of point clouds and obtain high-dimensional features of point clouds. The decoder of FoldingNet [20] demonstrated the feasibility of restoring point clouds from high-dimensional features. PCN [21] uses an encoder similar to that of FoldingNet to extract features and employs two stages in the decoder to generate high-density point clouds. TopNet [37] models the point cloud generation process as the growth of a rooted tree, and uses a hierarchical point cloud generation decoder. SA-Net [38] applies a self-attention mechanism to the network, which effectively preserves local information. SoftPoolNet [3] replaces max pooling with SoftPool and retains more information. PF-Net [39] uses an idea similar to fractal geometry, taking an incomplete point cloud as the input, but only outputting the missing part of the point cloud. SnowflakeNet [40] models the generation of a complete point cloud as a snowflake-like growth of points in a 3D space, revealing local geometric details.
The main advantage of the learning-based methods is that they have strong applicability, and there is no restriction on incomplete shapes or incomplete areas in the input point cloud. Even if the incompleteness is serious, it can be completed. The disadvantage is that they require a large amount of data for training. If the training data are too small, the learning-based methods cannot fit the shape well.
3. Methods
This section introduces the network architecture design. Our network predicts the point cloud of the occluded part from that of the unoccluded input part. Figure 1 illustrates the architecture of the network. The encoder uses the point cloud X of the unoccluded part as input and outputs a one-dimensional global feature vector. According to the global vector, the decoder deforms the 3D lattice into the missing part of the point cloud, Yocc. We optimized the network by calculating the loss between the Yocc and ground truth (GT), which retains only the occluded part, GTocc. Finally, the Yocc was stitched with the unoccluded part to obtain the complete point cloud, Ycomp. We evaluated the completion performance of the network by computing the loss between the Ycomp and GT. Next, we detail the architecture of the encoder, decoder, and use of the loss functions.
3.1. Encoder
To ensure that the encoder has excellent feature extraction, we used the point transformer [24] proposed by Zhao et al. as the encoder. As shown in Figure 2, the unoccluded part point cloud X passes through an MLP-point transformer module and is transformed into an matrix. It is then processed using the transition down and point transformer modules n times, and an matrix is obtained, where V is 32 × 4n. Finally, average pooling is performed on this matrix, and a global feature with the shape (1, V) is obtained.
The specific architecture of the point transformer layer is shown in Figure 2a. The self-attention feature, denoted as , of features corresponding to each point in the point cloud, denoted as , is calculated with the feature set of k-nearest neighbors:
(1)
(2)
where φ, ψ, and α are linear layers; γ and θ are nonlinear MLPs with two linear layers and one ReLU layer, respectively; ⊙ is the vector dot product; ẟ is the relative position encoding of two points, denoted as , , where , are the three-dimensional coordinates of points i and j, respectively.The transition down module reduces the number of points by farthest point sampling [38]. After each transition down module, the number of points becomes one-quarter of the original. We used the transition down module with the same architecture and parameters as the original point transformer. The specific architectures of the transition down module and point transformer module are shown in Figure 2.
3.2. Decoder
To generate the point cloud of the occluded part and complete the point cloud, we adopted an improved fold-based decoder architecture as the decoder of the network. The decoder transforms the 3D lattice into a point cloud of the occluded part of the shape. As shown in Figure 1, the global feature of the (1, V) output from the encoder is first repeated M times to form an M × V matrix, which is spliced with the coordinates of the three-dimensional lattice into an M × (V + 3) matrix. The 3D lattice is a cube with coordinates ranging from −1 to 1. There are M points in the cube. Second, the spliced matrix is input into the three-layer perceptron to complete the first deformation. The output 3D coordinates are spliced with the global features and copied M times to obtain an M × (V + 3) matrix, which is input into the three-layer perceptron to achieve the second deformation. Finally, the network outputs the reconstructed point cloud of the occluded part and splices it with the point cloud of the unoccluded part to obtain the completion result.
The decoder part implements a mapping from the 3D lattice to the missing part of the point cloud shape. The global feature output by the encoder serves as a parameter to guide the deformation operation of the decoder, essentially storing the force required to perform deformation. Because the multilayer perceptron is effective at approximating nonlinear functions, it can precisely apply the required force to deform the 3D point cloud and deform the 3D lattice into any desired shape.
3.3. Loss Function
The loss function measures the difference between two point clouds. Owing to the disordered nature of point clouds, the loss function should be insensitive to the order of points. We used the Chamfer distance (CD) proposed by Fan et al. [34] as our loss function.
(3)
Equation (3) is a symmetrical version of the formula used to calculate the CD between two point clouds. It measures the average closest point distance between the output point cloud (S1) and the GT point cloud (S2). The first term forces the output points to lie close to the GT points and the second term ensures that the GT point cloud is covered by the output point cloud.
In our experiment, we first calculated the CD distance between the point cloud Yocc of the occluded part that is output by the network and the point cloud of the occluded part in the GT (denoted as ). This distance is denoted as . We optimized the network according to . Then, the output was spliced with the input point cloud X of the unoccluded part to obtain the complete point cloud . The effect of point cloud completion was evaluated by calculating the CD between the and GT.
4. Experience and Results
In this section, we first describe how to create a dataset for training our network. We then compare the experimental results of our network with those of FoldingNet. Finally, we describe the ablation experiments used to verify the effectiveness of various changes in our network.
4.1. Environment and Network Parameters
4.1.1. System Environment
We implemented our analysis on a PC with Ubuntu18.04 as the operating system, an Intel Core i7-6800K CPU (Intel Technology (China) Co., Ltd.), and an NVIDIA GTX 1080Ti GPU (NVIDIA Semi-Conductor Technology (Shanghai) Co., Ltd.), and the experimental frameworks were python 3.8 (google. Inc.) and pytorch 1.8.2 (Open source software).
4.1.2. Network Specific Parameters
The number of points in the input of the unoccluded part (N) was 1536. The number of the transition down and point transformer modules (n) was set to four, and the encoder output a global feature vector with the shape (1512). The cube lattice input to the decoder was set to 512 points with sides of 8 units. Thus, the occluded part of the point cloud output by the decoder contained 512 points. Finally, it was spliced with the unoccluded part to obtain a point cloud of 2048 points.
4.1.3. Model Training Parameters
Adam was used as the network optimizer. The batch size was set to 10; the initial learning rate was set to10−4; and after every three rounds of training, the learning rate was reduced to 0.9. The network tended to converge after approximately 200 rounds. To ensure that the network was optimal, we conducted 300 training rounds.
4.2. Data Generation and Implementation Detail
To train our network, we used the car categories from the standard ShapeNet dataset. This category has 3162 shapes, and we used 2458 as the training set and 704 as the test set. We uniformly sampled 2048 points of the CAD model of the shape to obtain point cloud data.
All point cloud data were centered at the origin, and the coordinates were normalized to the range [−1, 1]. As shown in Figure 3, we used the sampled 2048 points as the GT of the complete point cloud and deleted 512 points near a random point using the k-nearest neighbor method to simulate occlusion. We used the deleted 512 points as the GT for training the network (denoted as GTocc), and the remaining 1536 points were used to simulate the unoccluded part as the input of the network.
4.3. Results
In this subsection, we qualitatively and quantitatively compare the experimental results of our network with those of the FoldingNet. FoldingNet was trained in two ways: the first was inputting the unoccluded part point cloud and directly predicting the complete point cloud. The 2D lattice of the decoder was initialized with a size of 32 × 64 to output 2048 points. In this experiment, we compared the original completion method of FoldingNet; see the FoldingNet (1) column in Figure 4 and the first row in Table 1. The second method was inputting the unoccluded part of the point cloud and predicting only the occluded part. The two-dimensional lattice in the decoder was initialized with a size of 16 × 32 to output 512 points. This experiment showed that under the same input and output, our completion results were better than those of FoldingNet; see the FoldingNet (2) column in Figure 4 and the second row in Table 1, where is the CD between the predicted occluded point cloud and GTocc, and is the CD between the complete point cloud and GT.
In Figure 4, we present a visualization of the results of our method and of the two FoldingNet training methods. From this, we can observe that the point cloud density output obtained using our method was more reasonable. The texture and distribution of the point cloud output by our method were closer to the GT and could be combined with the unoccluded parts without being obtrusive. In Table 1, we quantitatively compare the proposed method with FoldingNet and show that the CD of our method was much smaller than that of FoldingNet, which directly outputs the complete point cloud. Compared with FoldingNet, which outputs the occluded parts, the CD of our method was 8% smaller for the predicted occluded parts, and the CD of the complete point cloud output by our method was significantly smaller. These findings indicate that our method outperforms FoldingNet, both visually and quantitatively, suggesting that the improvements we proposed for the original network are effective.
4.4. Ablation Study
In this study, we verified the effectiveness of each of our changes through ablation experiments and qualitatively and quantitatively analyzed the experimental results. The datasets quantitatively analyzed the experimental results. The datasets used in the experiments were the point cloud shapes obtained from car models in ShapeNet. We chose the CD loss as the evaluation metric.
4.4.1. Transformer Encoder
In this subsection, we evaluate the effectiveness of the transformer encoder in extracting point cloud features. We replaced the transformer encoder with the encoder originally used by FoldingNet and did not change the other parts of the network structure. Comparing the data in the first and second rows in Table 2, after replacing the transformer module, the CD loss of the occluded part and complete point cloud increased. This showed that the transformer module has better information extraction ability and improves the performance of the entire network.
4.4.2. 3D Lattice
In this subsection, we evaluate the effectiveness of the 3D lattices in the decoder. We replaced the 3D lattice with the original 2D lattice in the FoldingNet decoder, while the other parts of the network structure remained unchanged. Comparing the data in the first and fourth rows in Table 2, after replacing the 3D lattice with a 2D lattice, the CD loss of the occluded part increased by 15.5%. In addition, as shown in the circled part in Figure 5, compared with the 2D lattice, the output point cloud density of the 3D lattice was more uniform, and the connection with the unoccluded part was more natural and unified in vision.
5. Discussion
5.1. Some Poor Completion
For most point cloud shapes, the point cloud density of the occluded part output by our network was relatively uniform and could be smoothly spliced with the unoccluded part; however, the distribution of the output point cloud was still slightly different from that of the occluded part. In some shapes, there was a gap between the point cloud of the occluded part and that of the unoccluded part (as shown in Figure 6). We speculate that this is because the network does not fully learn the distribution rules of point clouds in space, resulting in a certain degree of difference in the distribution between the output and original point clouds. If the distribution of the output point cloud of the generative adversarial network is close as possible to the original point cloud, the results may be improved.
5.2. The Effect of Density
In theory, the higher the density, the more conducive the network is to extracting the features of the point cloud and the better the completion effect of the point cloud. To verify the effect of density on the performance of the network, we conducted experiments with point clouds with 1024 points, and compared it with the experiments in Section 4 (using point clouds of 2048 points). The data showed that using the point clouds with 1024 points for training, the CD of the occluded part was 5.272 × 10−2, which is an increase compared with the 4.104 × 10−2 in the previous experiment.
In short, high-density point clouds are more conducive to completion. The shape of the completion is shown in Figure 7.
5.3. The Effect of the Scale of Occlusion
In theory, the more occluded parts, the more difficult it is to extract the features of the point cloud, and the completion effect may be affected to a certain extent. To verify the effect of the volume of the occluded part on network performance, we designed an experiment with 50% occlusion. The experimental results showed that under 50% occlusion, the CD of the missing part was 4.112 × 10−2, while the CD of the missing part in Section 4 (25% occlusion) was 4.104 × 10−2. The results showed that as the degree of occlusion increased, the CD only slightly increased, indicating that the network we designed could still effectively extract the global features of objects from severely occluded data.
In short, the larger the volume of the occluded part, the worse the completion performance. The shape of completion is shown in Figure 8.
5.4. The Behavior of the Network on the Other Categories
To verify that our network works on other categories of shapes, we conducted experiments on seven other categories, and the results are shown in Table 3.
Among the seven categories, performance in the airplane category was the best, and performance in the cabinet category was the worst. The network produced different performances for the different categories of shapes. We think that for objects with more details, it is more difficult for the network to extract the detailed features, and the shape is more difficult to fit. Overall, our network is usable in other categories as well.
6. Conclusions
This study proposed an end-to-end deep neural network for point cloud completion. Our network improved upon FoldingNet. We used a transformer as the encoder of the point cloud completion network to extract the global features of the point cloud; we also replaced the 2D lattice in the decoder with a 3D lattice for the output point cloud density to be more uniform and detailed. We conducted experiments on a point cloud dataset sampled from the ShapeNet car category model. The experimental results showed that the changes we made to the cloud completion network improved its performance. The point cloud completion network proposed in this study can enable machines to acquire and analyze information about surrounding objects with increased accuracy and improve their perception of their surroundings.
Conceptualization, X.M.; methodology, X.M., J.S., and X.L.; software, X.M.; validation, X.M. and X.L.; formal analysis, J.S.; investigation, X.M.; resources, J.S. and X.L.; data curation, J.S.; writing—original draft preparation, X.M.; writing—review and editing, X.L.; visualization, X.M.; supervision, X.L.; project administration, J.S. and X.L.; funding acquisition, X.L. and J.S. All authors have read and agreed to the published version of the manuscript.
Not applicable.
Not applicable.
Not applicable.
The authors declare no conflict of interest.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Figure 1. Network architecture mainly consists of a transformer-based encoder and an improved fold-based decoder.
Figure 2. Network architecture of the transformer encoder comprises (a) point transformer layer, (b) point transformer block, and (c) transition down block.
Figure 4. Output results of different methods. Occluded part represents the predicted point cloud of the occluded part, complete represents the complete point cloud after splicing with the unoccluded part. Since FoldingNet (1) was trained using the first method, which directly outputs the completion point cloud, there is no occluded part column.
Quantitative comparison of different methods.
Methods |
|
|
---|---|---|
FoldingNet (1) | — | 4.403 |
FoldingNet (2) | 4.461 | 1.034 |
Ours | 4.104 | 0.965 |
Ablation study.
Methods |
|
|
---|---|---|
Ours | 4.104 | 0.964 |
Without Transformer | 4.228 | 0.984 |
Without 3D | 4.747 | 1.088 |
The behavior of the network on different categories.
Category |
|
---|---|
Airplane | 2.720 |
Cabinet | 6.099 |
Chair | 5.965 |
Lamp | 5.925 |
Sofa | 5.374 |
Table | 5.563 |
Watercraft | 4.411 |
Car | 4.104 |
References
1. Hegde, V.; Zadeh, R. FusionNet: 3D Object Classification Using Multiple Data Representations. arXiv; 2016; arXiv: 1607.05695
2. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. Proceedings of the 30th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); Honolulu, HI, USA, 21–26 July 2017.
3. Wang, Y.; Tan, D.J.; Navab, N.; Tombari, F. SoftPoolNet: Shape Descriptor for Point Cloud Completion and Classification. Proceedings of the 16th European Conference; Glasgow, UK, 23–28 August 2020.
4. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet plus plus: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. Proceedings of the 31st Annual Conference on Neural Information Processing Systems (NIPS); Long Beach, CA, USA, 4–9 December 2017.
5. Atik, M.E.; Duran, Z. An Efficient Ensemble Deep Learning Approach for Semantic Point Cloud Segmentation Based on 3D Geometric Features and Range Images. Sensors; 2022; 22, 6210. [DOI: https://dx.doi.org/10.3390/s22166210] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/36015964]
6. Mousavian, A.; Anguelov, D.; Flynn, J.; Kosecka, J. 3D bounding box estimation using deep learning and geometry. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Honolulu, HI, USA, 21–26 July 2017; pp. 7074-7082.
7. Li, Y.; Snavely, N.; Huttenlocher, D.; Fua, P. Worldwide pose estimation using 3d point clouds. Computer Vision—ECCV 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 15-29.
8. Alhamzi, K.; Elmogy, M.; Barakat, S. 3d object recognition based on local and global features using point cloud library. Int. J. Adv. Comput. Technol.; 2015; 7, 43.
9. Wang, D.Z.; Posner, I. Voting for voting in online point cloud object detection. Robot. Sci. Syst.; 2015; 1, pp. 10-15.
10. Zhou, Y.; Tuzel, O. Voxelnet: End-to-end learning for point cloud based 3d object detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Salt Lake City, UT, USA, 18–22 June 2018; pp. 4490-4499.
11. Cárdenas-Robledo, L.A.; Hernández-Uribe, O.; Reta, C.; Cantoral-Ceballos, J.A. Extended reality applications in industry 4.0.—A systematic literature review. Telemat. Inform.; 2022; 73, 101863. [DOI: https://dx.doi.org/10.1016/j.tele.2022.101863]
12. Tsaramirsis, G.; Kantaros, A.; Al-Darraji, I.; Piromalis, D.; Apostolopoulos, C.; Pavlopoulou, A.; Alrammal, M.; Ismail, Z.; Buhari, S.M.; Stojmenovic, M. et al. A modern approach towards an industry 4.0 model: From driving technologies to management. J. Sens.; 2022; 2022, 5023011. [DOI: https://dx.doi.org/10.1155/2022/5023011]
13. Kum, S.; Oh, S.; Yeom, J.; Moon, J. Optimization of Edge Resources for Deep Learning Application with Batch and Model Management. Sensors; 2022; 22, 6717. [DOI: https://dx.doi.org/10.3390/s22176717] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/36081177]
14. Piromalis, D.; Kantaros, A. Digital Twins in the Automotive Industry: The Road toward Physical-Digital Convergence. Appl. Syst. Innov.; 2022; 5, 65. [DOI: https://dx.doi.org/10.3390/asi5040065]
15. Martínez-Olvera, C. Towards the Development of a Digital Twin for a Sustainable Mass Customization 4.0 Environment: A Literature Review of Relevant Concepts. Automation; 2022; 3, pp. 197-222. [DOI: https://dx.doi.org/10.3390/automation3010010]
16. Eigen, D.; Puhrsch, C.; Fergus, R. Depth Map Prediction from a Single Image using a Multi-Scale Deep Network. Proceedings of the 28th Conference on Neural Information Processing Systems (NIPS); Montreal, QC, Canada, 8–13 December 2014.
17. Gregor, R.; Schreck, T.; Sipiran, I. Approximate Symmetry Detection in Partial 3D Meshes. Comput. Graph. Forum J. Eur. Assoc. Comput. Graph.; 2014; 33, pp. 131-140.
18. Wang, N.; Zhang, Y.; Li, Z.; Fu, Y.; Liu, W.; Jiang, Y.-G. Pixel2Mesh: Generating 3D Mesh Models from Single RGB Images. Proceedings of the 15th European Conference on Computer Vision (ECCV); Munich, Germany, 8–14 December 2018.
19. You, C.C.; Lim, S.P.; Lim, S.C.; Tan, J.S.; Lee, C.K.; Min, Y.; Khaw, Y.M.J. A Survey on Surface Reconstruction Techniques for Structured and Unstructured Data. Proceedings of the 2020 IEEE Conference on Open Systems (ICOS); Kota Kinabalu, Malaysia, 17–19 November 2020.
20. Yang, Y.; Feng, C.; Shen, Y.; Tian, D. FoldingNet: Point Cloud Auto-Encoder via Deep Grid Deformation. Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition; Salt Lake City, UT, USA, 18–23 June 2018; pp. 206-215.
21. Yuan, W.; Khot, T.; Held, D.; Mertz, C.; Hebert, M. PCN: Point Completion Network. Proceedings of the 6th International Conference on 3D Vision (3DV); Verona, Italy, 5–8 September 2018.
22. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. Proceedings of the Advances in Neural Information Processing Systems; Long Beach, CA, USA, 4–9 December 2017.
23. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv; 2020; arXiv: 2010.11929
24. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.H.S.; Koltun, V. Point Transformer. Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV); Virtual, 11–17 October 2021.
25. Sarkar, K.; Varanasi, K.; Stricker, D. Learning quadrangulated patches for 3D shape parameterization and completion. Proceedings of the International Conference on 3D Vision (3DV); Qingdao, China, 7 June 2018.
26. Berger, M.; Tagliassacchi, A.; Seversky, L.; Alliez, P.; Levine, J.; Sharf, A.; Silva, C. State of the Art in Surface Reconstruction from Point Clouds. Proceedings of the Eurographics 2014—State of the Art Reports; Strasbourg, France, 7–11 April 2014.
27. Sung, M.; Kim, V.G.; Angst, R.; Guibas, L. Data-driven structural priors for shape completion. ACM Trans. Graph.; 2015; 34, pp. 1-11. [DOI: https://dx.doi.org/10.1145/2816795.2818094]
28. Li, Y.; Dai, A.; Guibas, L.; Niebner, M. Database-Assisted Object Retrieval for Real-Time 3D Reconstruction. Comput. Graph. Forum; 2015; 34, pp. 435-446. [DOI: https://dx.doi.org/10.1111/cgf.12573]
29. Nan, L.; Xie, K.; Sharf, A. A search-classify approach for cluttered indoor scene understanding. ACM Trans. Graph.; 2012; 31, 137. [DOI: https://dx.doi.org/10.1145/2366145.2366156]
30. Martinovic, A.; Gool, L.V. Bayesian Grammar Learning for Inverse Procedural Modeling. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Portland, OR, USA, 23 June 2013.
31. Gupta, S.; Arbeláez, P.; Girshick, R.; Malik, J. Aligning 3D models to RGB-D images of cluttered scenes. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Boston, MA, USA, 7–12 June 2015.
32. Rock, J.; Gupta, T.; Thorsen, J.; Gwak, J.; Shin, D.; Hoiem, D. Completing 3D object shape from one depth image. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Boston, MA, USA, 7–12 June 2015.
33. Yin, K.; Huang, H.; Zhang, H.; Gong, M.; Cohen-Or, D.; Chen, B. Morfit: Interactive surface reconstruction from incomplete point clouds with curve-driven topology and geometry control. Acm Trans. Graph.; 2014; 33, 202. [DOI: https://dx.doi.org/10.1145/2661229.2661241]
34. Mitra, N.J.; Pauly, M.; Wand, M.; Ceylan, D. Symmetry in 3D Geometry: Extraction and Applications. Comput. Graph. Forum; 2013; 32, pp. 1-23. [DOI: https://dx.doi.org/10.1111/cgf.12010]
35. Sharma, A.; Grau, O.; Fritz, M. VConv-DAE: Deep Volumetric Shape Learning Without Object Labels. Computer Vision—ECCV 2016 Workshops; Springer: Cham, Switzerland, 2016.
36. Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3D ShapeNets: A Deep Representation for Volumetric Shapes. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Boston, MA, USA, 7–12 June 2015.
37. Tchapmi, L.P.; Kosaraju, V.; Rezatofighi, H.; Reid, I.; Savarese, S. TopNet: Structural Point Cloud Decoder. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Long Beach, CA, USA, 15–20 June 2019.
38. Yang, Y.B.; Zhang, Q.L. SA-Net: Shuffle Attention for Deep Convolutional Neural Networks. Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); Toronto, ON, Canada, 13 May 2021.
39. Huang, Z.; Yu, Y.; Xu, J.; Ni, F.; Le, X. PF-Net: Point Fractal Network for 3D Point Cloud Completion. arXiv; 2020; arXiv: 2003.00410
40. Xiang, P.; Wen, X.; Liu, Y.S.; Cao, Y.P.; Wan, P.; Zheng, W.; Han, Z. SnowflakeNet: Point Cloud Completion by Snowflake Point Deconvolution with Skip-Transformer. arXiv; 2021; arXiv: 2108.04444
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
With the development of autonomous driving, augmented reality, and other fields, it is becoming increasingly important for machines to more accurately and comprehensively perceive their surrounding environment. LiDAR is one of the most important tools used by machines to obtain information about the surrounding environment. However, because of occlusion, the point cloud data obtained by LiDAR are not the complete shape of the object, and completing the incomplete point cloud shape is of great significance for further data analysis, such as classification and segmentation. In this study, we examined the completion of a 3D point cloud and improved upon the FoldingNet auto-encoder. Specifically, we used the encoder–decoder architecture to design our point cloud completion network. The encoder part uses the transformer module to enhance point cloud feature extraction, and the decoder part changes the 2D lattice used by the A network into a 3D lattice so that the network can better fit the shape of the 3D point cloud. We conducted experiments on point cloud datasets sampled from the ShapeNet car-category CAD models to verify the effectiveness of the various improvements made to the network.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details
1 State Key Laboratory of Integrated Optoelectronics, College of Electronic Science and Engineering, Jilin University, Changchun 130012, China
2 State Key Laboratory of Integrated Optoelectronics, College of Electronic Science and Engineering, Jilin University, Changchun 130012, China; Peng Cheng Laboratory, Shenzhen 518000, China