SOTAVerified

Visual-Inertial Multi-Instance Dynamic SLAM with Object-level Relocalisation

2022-08-08Code Available1· sign in to hype

Yifei Ren, Binbin Xu, Christopher L. Choi, Stefan Leutenegger

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

In this paper, we present a tightly-coupled visual-inertial object-level multi-instance dynamic SLAM system. Even in extremely dynamic scenes, it can robustly optimise for the camera pose, velocity, IMU biases and build a dense 3D reconstruction object-level map of the environment. Our system can robustly track and reconstruct the geometries of arbitrary objects, their semantics and motion by incrementally fusing associated colour, depth, semantic, and foreground object probabilities into each object model thanks to its robust sensor and object tracking. In addition, when an object is lost or moved outside the camera field of view, our system can reliably recover its pose upon re-observation. We demonstrate the robustness and accuracy of our method by quantitatively and qualitatively testing it in real-world data sequences.

Tasks

Reproductions