Bulletin of Surveying and Mapping ›› 2025, Vol. 0 ›› Issue (4): 9-13.doi: 10.13474/j.cnki.11-2246.2025.0402

Previous Articles    

A visual SLAM method for merging semantic information in indoor dynamic scenes

WANG Yizhe1, ZHANG Ruiju1,2,3, WANG Jian1, XIE Xinrui1, HUANG Qicheng1   

  1. 1. School of Geomatics and Urban Spatial Informatics, Beijing University of Civil Engineering and Architecture, Beijing 102616, China;
    2. Engineering Center of Representative Architecture and Ancient Architecture Database, Ministry of Education, Beijing 102616, China;
    3. Key Laboratory of Fine Reconstruction and Health Monitoring of Architectural Heritage, Beijing 102616, China
  • Received:2024-06-20 Published:2025-04-28

Abstract: Visual SLAM is a core technology for autonomous perception and navigation in intelligent devices, playing a crucial role in AI and robotics. However, traditional visual SLAM algorithms suffer significantly in stability and localization accuracy when scenes contain moving objects. To address this, this paper proposes a SLAM scheme that integrates semantic information for indoor dynamic scenarios. Based on ORB-SLAM2, it introduces the GCNv2 network for deep feature extraction and YOLOv5 for semantic segmentation to identify dynamic objects. Combined with motion consistency analysis, it effectively eliminates dynamic interference, enhancing robustness. Tests on the TUM standard dataset show the improved algorithm significantly outperforms the original ORB-SLAM2 in dynamic indoor environments, with an average positioning accuracy improvement of 55.75%. This result demonstrates the proposed method's effectiveness, significantly boosting SLAM system performance in complex dynamic environments.

Key words: visual SLAM, semantic information, feature extraction, dynamic scenes

CLC Number: