Robust 3D Object Tracking from Monocular Images using Stable Parts

Mahdi Rad, Vincent Lepetit

Research output: Contribution to journalArticlepeer-review


We present an algorithm for estimating the pose of a rigid object in real-time under challenging conditions. Our method effectively handles poorly textured objects in cluttered, changing environments, even when their appearance is corrupted by large occlusions, and it relies on grayscale images to handle metallic environments on which depth cameras would fail. As a result, our method is suitable for practical Augmented Reality applications including industrial environments. At the core of our approach is a novel representation for the 3D pose of object parts: We predict the 3D pose of each part in the form of the 2D projections of a few control points. The advantages of this representation is three-fold: We can predict the 3D pose of the object even when only one part is visible; when several parts are visible, we can easily combine them to compute a better pose of the object; the 3D pose we obtain is usually very accurate, even when only few parts are visible. We show how to use this representation in a robust 3D tracking framework. In addition to extensive comparisons with the state-of-the-art, we demonstrate our method on a practical Augmented Reality application for maintenance assistance in the ATLAS particle detector at CERN.
Original languageEnglish
Pages (from-to)1465 - 1479
JournalIEEE Transactions on Pattern Analysis and Machine Intelligence
Issue number6
Publication statusPublished - 2017


Dive into the research topics of 'Robust 3D Object Tracking from Monocular Images using Stable Parts'. Together they form a unique fingerprint.

Cite this