3d indoor point cloud semantic segmentation using image and voxel

Sang Sik Yeom, Jong Eun Ha

Research output: Contribution to journalArticlepeer-review

2 Scopus citations

Abstract

In this paper, we propose a parallel network architecture that exhibits improved performance by fusing two-dimensional (2D) and three-dimensional (3D) features. A voxel-based and a projection-based method were adopted to derive the results through one scan. Our approach consists of two parallel networks, extracts features along each dimension, and converges them in a fusion network. In the fusion network, the voxel blocks and 2D feature maps extracted from each structure are fused to the voxel grid and then trained through convolution. For effective training of 2D networks, we use data augmentation techniques using coordinate system rotation transformation. In addition, a multi-loss with weights applied to each dimension was employed to effectively enhance the performance of the system, and the results revealed that the system exhibited better performance than when a single loss was used. Our proposed method can achieve better performance by changing the performance of the 2D network and 3D network, which can be generalized using other structures.

Original languageEnglish
Pages (from-to)1000-1007
Number of pages8
JournalJournal of Institute of Control, Robotics and Systems
Volume27
Issue number12
DOIs
StatePublished - 2021

Keywords

  • 3D Vision
  • Point Cloud
  • Semantic Segmentation

Fingerprint

Dive into the research topics of '3d indoor point cloud semantic segmentation using image and voxel'. Together they form a unique fingerprint.

Cite this