Monocular Visual Odometry with YOLO-Based Dynamic Object Filtering
Resumo
This work addresses the challenge of dynamic elements in monocular visual odometry, which traditionally relies on keypoint matching under the assumption of a fully static scene. In real-world environments, however, moving objects violate this assumption and can degrade pose estimation accuracy. To miti-gate this, we propose enhancing a classical feature-based visual odometry pipeline by using the YOLO (You Only Look Once) neural network to detect potentially dynamic objects and remove keypoints located within their bounding boxes. We compare this approach to a baseline method that uses only geometric keypoint matching, using sequences from the KITTI dataset for evaluation. Quantitative evaluation shows an improvement in the accuracy of the estimated trajectory when keypoints on dynamic objects are excluded from the odometry process.
Palavras-chave:
YOLO, Accuracy, Filtering, Robot vision systems, Pose estimation, Pipelines, Neural networks, Trajectory, Odometry, Visual odometry, visual odometry, KITTI, monocular camera
Publicado
13/10/2025
Como Citar
L. FILHO, Márcio G.; FRANÇANI, André O.; MAXIMO, Marcos R. O. A..
Monocular Visual Odometry with YOLO-Based Dynamic Object Filtering. In: SIMPÓSIO BRASILEIRO DE ROBÓTICA E SIMPÓSIO LATINO AMERICANO DE ROBÓTICA (SBR/LARS), 17. , 2025, Vitória/ES.
Anais [...].
Porto Alegre: Sociedade Brasileira de Computação,
2025
.
p. 213-218.
