Moving Object Detection for Visual Odometry in a Dynamic Environment based on Occlusion Accumulation
- Authors
- Kim, Haram; Kim, Pyojin; Kim, H. Jin
- Issue Date
- May-2020
- Publisher
- IEEE Robotics and Automation Society (RAS)
- Citation
- Proceedings - IEEE International Conference on Robotics and Automation, pp 8658 - 8644
- Pages
- -13
- Journal Title
- Proceedings - IEEE International Conference on Robotics and Automation
- Start Page
- 8658
- End Page
- 8644
- URI
- https://scholarworks.sookmyung.ac.kr/handle/2020.sw.sookmyung/151322
- DOI
- 10.1109/ICRA40945.2020.9196767
- ISSN
- 1050-4729
2577-087X
- Abstract
- Detection of moving objects is an essential capability in dealing with dynamic environments. Most moving object detection algorithms have been designed for color images without depth. For robotic navigation where real-time RGB-D data is often readily available, utilization of the depth information would be beneficial for obstacle recognition. Here, we propose a simple moving object detection algorithm that uses RGB-D images. The proposed algorithm does not require estimating a background model. Instead, it uses an occlusion model which enables us to estimate the camera pose on a background confused with moving objects that dominate the scene. The proposed algorithm allows to separate the moving object detection and visual odometry (VO) so that an arbitrary robust VO method can be employed in a dynamic situation with a combination of moving object detection, whereas other VO algorithms for a dynamic environment are inseparable. In this paper, we use dense visual odometry (DVO) as a VO method with a bi-square regression weight. Experimental results show the segmentation accuracy and the performance improvement of DVO in the situations. We validate our algorithm in public datasets and our dataset which also publicly accessible.
- Files in This Item
-
- Appears in
Collections - 공과대학 > 기계시스템학부 > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.