Full text

Turn on search term navigation

© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.

Abstract

Recovery of articulated 3D structure from 2D observations is a challenging computer vision problem with many applications. Current learning-based approaches achieve state-of-the-art accuracy on public benchmarks but are restricted to specific types of objects and motions covered by the training datasets. Model-based approaches do not rely on training data but show lower accuracy on these datasets. In this paper, we introduce a model-based method called Structure from Articulated Motion (SfAM), which can recover multiple object and motion types without training on extensive data collections. At the same time, it performs on par with learning-based state-of-the-art approaches on public benchmarks and outperforms previous non-rigid structure from motion (NRSfM) methods. SfAM is built upon a general-purpose NRSfM technique while integrating a soft spatio-temporal constraint on the bone lengths. We use alternating optimization strategy to recover optimal geometry (i.e., bone proportions) together with 3D joint positions by enforcing the bone lengths consistency over a series of frames. SfAM is highly robust to noisy 2D annotations, generalizes to arbitrary objects and does not rely on training data, which is shown in extensive experiments on public benchmarks and real video sequences. We believe that it brings a new perspective on the domain of monocular 3D recovery of articulated structures, including human motion capture.

Details

Title
Structure from Articulated Motion: Accurate and Stable Monocular 3D Reconstruction without Training Data
Author
Kovalenko, Onorina 1   VIAFID ORCID Logo  ; Golyanik, Vladislav 2 ; Malik, Jameel 3 ; Elhayek, Ahmed 4 ; Stricker, Didier 5 

 Department Augmented Vision, German Research Center for Artificial Intelligence (DFKI), 67663 Kaiserslautern, Germany; [email protected] (J.M.); [email protected] (A.E.); [email protected] (D.S.) 
 Department of Computer Graphics, Max Planck Institute for Informatics, 66123 Saarbrücken, Germany; [email protected] 
 Department Augmented Vision, German Research Center for Artificial Intelligence (DFKI), 67663 Kaiserslautern, Germany; [email protected] (J.M.); [email protected] (A.E.); [email protected] (D.S.); Department of Computer Science, University of Kaiserslautern, 67663 Kaiserslautern, Germany; School of Electrical Engineering and Computer Science (SEECS), National University of Sciences and Technology (NUST), 44000 Islamabad, Pakistan 
 Department Augmented Vision, German Research Center for Artificial Intelligence (DFKI), 67663 Kaiserslautern, Germany; [email protected] (J.M.); [email protected] (A.E.); [email protected] (D.S.); Department of Computer Science, University of Prince Mugrin (UPM), 20012 Madinah, Saudi Arabia 
 Department Augmented Vision, German Research Center for Artificial Intelligence (DFKI), 67663 Kaiserslautern, Germany; [email protected] (J.M.); [email protected] (A.E.); [email protected] (D.S.); Department of Computer Science, University of Kaiserslautern, 67663 Kaiserslautern, Germany 
First page
4603
Publication year
2019
Publication date
2019
Publisher
MDPI AG
e-ISSN
14248220
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
2535498473
Copyright
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.