Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2025, Vol. 46 ›› Issue (1): 59-69.DOI: 10.11996/JG.j.2095-302X.2025010059

• Image Processing and Computer Vision • Previous Articles     Next Articles

Point cloud feature enhanced 3D object detection in complex indoor scenes

YUAN Chao(), ZHAO Mingxue, ZHANG Fengyi, FENG Xiaoyong, LI Bing, CHEN Rui()   

  1. Department of Automation, North China Electric Power University, Baoding Hebei 071003, China
  • Received:2024-08-02 Accepted:2024-09-23 Online:2025-02-28 Published:2025-02-14
  • Contact: CHEN Rui
  • About author:First author contact:

    YUAN Chao (1985-), associate professor, Ph.D. His main research interests cover computer vision and intelligent control of multi-axis manipulator, etc. E-mail:chaoyuan@ncepu.edu.cn

  • Supported by:
    The National Natural Science Foundation of China Joint Fund of China is a Key Support Project(U21A20486);Funded by the Central University Basic Research Business Fund(2024MS137)

Abstract:

3D point cloud object detection in complex indoor scenes presents challenges due to large-scale point clouds and dense objects with many details. When dealing with point cloud data, existing detection algorithms lose a significant amount of local features and fail to extract enough spatial and semantic information, resulting in low detection accuracy. To solve this problem, a point cloud features enhanced 3D object detection in complex indoor scenes (PEF) algorithm was proposed based on an improved VoteNet. Firstly, a dynamic feature compensation module was used to simulate the interactive query process between seed point set features and grouping set features, gradually recovering lost features for feature compensation. Secondly, a residual MLP module was introduced into the feature extraction part, and a deeper feature learning network was constructed through a residual structure to mine more detailed point cloud features. Finally, in the proposal stage, a feature self-attention mechanism was introduced to model the semantic relationship between a set of independent object points, generating a new feature map. Experiments conducted on the public datasets SUN RGB-D and ScanNet V2 demonstrated that the improved model enhanced the detection accuracy for indoor objects by 5.0% and 11.5% respectively on mAP@0.25 compared with the baseline model. Extensive ablation experiments confirmed the effectiveness of each improved module.

Key words: indoor scene, 3D point cloud, object detection, feature compensation, interactive query, residual, self-attention, feature map

CLC Number: