Semantic scene reconstruction from sparse and incomplete point clouds is a critical task for point scene understanding. It aims to recognize semantic labels for objects and recover their complete shapes as meshes. Existing methods often fail to realize high-quality instance reconstruction due to inadequate shape representation and underutilization of proposal point clouds. To address these issues, we optimize the previous BSP/occupancy-to-mesh reconstruction framework to points-to-mesh and accomplish multi-level utilization of proposals. We chose point cloud as the representation of completion to reduce the difficulty of restoring curved shallow parts. Benefiting from the optimization, we can match and merge proposal point clouds with the restored ones, avoiding missing parts existing in inputs. We design an effective pose normalization module to extract point-based features from normalized proposals, which are fused with features extracted from voxelized proposals, avoiding the detailed geometry lost in voxelization and enhancing the reconstruction's robustness to different input postures. The suitable points-to-mesh reconstruction framework and full utilization of proposals make our method improve reconstruction results efficiently. Detailed experiments on the challenging ScanNet dataset of the semantic scene reconstruction benchmark show that our network outperforms state-of-the-art methods in both completion and mapping metrics.
Incomplete Point clouds obtained from one-side scanning always result in structural loss in 3D shape representations, thus many learning-based methods are proposed to restore complete point clouds from partial ones. However, most of them only utilize global features of inputs to generate outputs, which might lose details. In this paper, a new method that utilizes both global and local features is proposed. First, Local features are extracted from inputs and analyzed under the conditions interpreted by global features. Second, conditional local feature vectors are deeply fused with each other via graph convolution and self-attention. Third, deeply-fused features are decoded for generating coarse point clouds. Last, global features extracted from inputs and coarse outputs are combined to generate fine outputs with high-density. Our network is trained and tested on eight categories of objects in ModelNet. The results show that our network is able to overcome instability in local feature awareness, restore complete point clouds with more details and smoother shapes, and outperform most of those existing methods both intuitively and quantitatively. Our source codes will be available at: https://github.com/wuhang100/LRA-Net.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.