论文标题
PBP-NET:3D点云分段的点投影和反向预测网络
PBP-Net: Point Projection and Back-Projection Network for 3D Point Cloud Segmentation
论文作者
论文摘要
经过3D扫描技术的大量发展,最近提出了许多针对3D视觉任务的方法,包括一些利用2D卷积神经网络(CNN)的方法。但是,即使2D CNN在许多2D视觉任务中都达到了高性能,但现有作品并未有效地将其应用于3D视觉任务。特别是,由于每个点的密集预测难度,这需要丰富的特征表示,因此对细分的研究尚未得到很好的研究。在本文中,我们提出了一个名为Point投影和反向投影网络(PBP-NET)的简单有效的体系结构,该架构利用2D CNN进行3D点云分割。引入了3个模块,每个模块都将3D点云投射到2D平面上,使用2D CNN骨架提取功能,然后在原始的3D点云上进行反向项目。为了证明使用2D CNN进行有效的3D特征提取,我们进行了各种实验,包括与最近方法的比较。我们通过消融研究分析了提出的模块,并对对象零件分割(Shapenet-Part数据集)和室内场景语义分割(S3DIS数据集)进行实验。实验结果表明,提出的PBP-NET与现有最新方法的性能相当。
Following considerable development in 3D scanning technologies, many studies have recently been proposed with various approaches for 3D vision tasks, including some methods that utilize 2D convolutional neural networks (CNNs). However, even though 2D CNNs have achieved high performance in many 2D vision tasks, existing works have not effectively applied them onto 3D vision tasks. In particular, segmentation has not been well studied because of the difficulty of dense prediction for each point, which requires rich feature representation. In this paper, we propose a simple and efficient architecture named point projection and back-projection network (PBP-Net), which leverages 2D CNNs for the 3D point cloud segmentation. 3 modules are introduced, each of which projects 3D point cloud onto 2D planes, extracts features using a 2D CNN backbone, and back-projects features onto the original 3D point cloud. To demonstrate effective 3D feature extraction using 2D CNN, we perform various experiments including comparison to recent methods. We analyze the proposed modules through ablation studies and perform experiments on object part segmentation (ShapeNet-Part dataset) and indoor scene semantic segmentation (S3DIS dataset). The experimental results show that proposed PBP-Net achieves comparable performance to existing state-of-the-art methods.