Can Semantic-based Filtering of Dynamic Objects improve Visual SLAM and Visual Odometry?

  • Leonardo Rezende Costa Unicamp
  • Esther Luna Colombini Unicamp


This work introduces a novel approach to improve robot perception in dynamic environments using Semantic Filtering. The goal is to enhance Visual Simultaneous Localization and Mapping (V-SLAM) and Visual Odometry (VO) tasks by excluding feature points associated with moving objects. Four different approaches for semantic extraction, namely YOLOv3, DeepLabv3 with two different backbones, and Mask R-CNN, were evaluated. The framework was tested on various datasets, including KITTI, TUM and a simulated sequence generated on AirSim. The results demonstrated that the proposed semantic filtering significantly reduced estimation errors in VO tasks, with average error reduction ranging from 2.81% to 15.98%, while the results for V-SLAM were similar to the base work, especially for sequences with detected loops. Although fewer keypoints are used, the estimations benefit from the points excluded in VO. More experiments are needed to address the effects in VSLAM due to the presence of loop closure and the nature of the datasets.
Palavras-chave: V-SLAM, Visual Odometry, Semantic
Como Citar

Selecione um Formato
COSTA, Leonardo Rezende; COLOMBINI, Esther Luna. Can Semantic-based Filtering of Dynamic Objects improve Visual SLAM and Visual Odometry?. In: SIMPÓSIO BRASILEIRO DE ROBÓTICA E SIMPÓSIO LATINO AMERICANO DE ROBÓTICA (SBR/LARS), 15. , 2023, Salvador/BA. Anais [...]. Porto Alegre: Sociedade Brasileira de Computação, 2023 . p. 567-572.