期刊文献+

Fusing Geometrical and Visual Information via Superpoints for the Semantic Segmentation of 3D Road Scenes 被引量:7

Fusing Geometrical and Visual Information via Superpoints for the Semantic Segmentation of 3D Road Scenes
原文传递
导出
摘要 This paper addresses the problem of the semantic segmentation of large-scale 3D road scenes by incorporating the complementary advantages of point clouds and images.To make full use of geometrical and visual information,this paper extracts 3D geometric features from a point cloud using a deep neural network for 3D semantic segmentation and extracts 2D visual features from images using a Convolutional Neural Network(CNN)for 2D semantic segmentation.In order to bridge the features of the two modalities,this paper uses superpoints as an intermediate representation to connect the 2D features with the 3D features.A superpoint-based pooling method is proposed to fuse the features from the two different modalities for joint learning.To evaluate the approach,the paper generates 3D scenes from the Virtual KITTI dataset.The results of the experiments demonstrate that the proposed approach is capable of segmenting large-scale 3D road scenes based on the compact and semantically homogeneous superpoints,and that it achieves considerable improvements over the 2D image and 3D point cloud semantic segmentation methods. This paper addresses the problem of the semantic segmentation of large-scale 3D road scenes by incorporating the complementary advantages of point clouds and images.To make full use of geometrical and visual information, this paper extracts 3D geometric features from a point cloud using a deep neural network for 3D semantic segmentation and extracts 2D visual features from images using a Convolutional Neural Network(CNN)for 2D semantic segmentation.In order to bridge the features of the two modalities, this paper uses superpoints as an intermediate representation to connect the 2D features with the 3D features.A superpoint-based pooling method is proposed to fuse the features from the two different modalities for joint learning.To evaluate the approach, the paper generates 3D scenes from the Virtual KITTI dataset.The results of the experiments demonstrate that the proposed approach is capable of segmenting large-scale 3D road scenes based on the compact and semantically homogeneous superpoints, and that it achieves considerable improvements over the 2D image and 3D point cloud semantic segmentation methods.
出处 《Tsinghua Science and Technology》 SCIE EI CAS CSCD 2020年第4期498-507,共10页 清华大学学报(自然科学版(英文版)
基金 supported by the National Natural Science Foundation of China(No.U1764264/61873165) Shanghai Automotive Industry Science and Technology Development Foundation(No.1807) the International Chair on Automated Driving of Ground Vehicle.
关键词 SCENE understanding point cloud SEMANTIC segmentation MULTI-MODAL information fusion deep learning scene understanding point cloud semantic segmentation multi-modal information fusion deep learning
  • 相关文献

同被引文献49

引证文献7

二级引证文献39

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部