Han XiaoHouxuan LiuYunchao DingLu Yang
Accurate perception of objects in the environment is important for improving the scene understanding capability of SLAM systems. In robotic and augmented reality applications, object maps with semantic and metric information show attractive advantages. In this letter, we present RO-MAP, a novel multi-object mapping pipeline that does not rely on 3D priors. Given only monocular input, we use neural radiance fields to represent objects and couple them with a lightweight object SLAM based on multi-view geometry, to simultaneously localize objects and implicitly learn their dense geometry. We create separate implicit models for each detected object and train them dynamically and in parallel as new observations are added. Experiments on synthetic and real-world datasets demonstrate that our method can generate semantic object map with shape reconstruction, and be competitive with offline methods while achieving real-time performance (25 Hz).
Lingzhi LiZhongshu WangZhen ShenLi ShenPing Tan
Jeffrey S. EiyikeElvis GyaaseVikas Dhiman
Liping ZhuHaibo ZhouSilin WuTianrong ChengHongjun Sun
Peter HedmanPratul P. SrinivasanBen MildenhallJonathan T. BarronPaul Debevec