File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: MotioNet: 3D Human Motion Reconstruction from Monocular Video with Skeleton Consistency

TitleMotioNet: 3D Human Motion Reconstruction from Monocular Video with Skeleton Consistency
Authors
Issue Date2021
PublisherAssociation for Computing Machinery, Inc. The Journal's web site is located at http://tog.acm.org
Citation
ACM Transactions on Graphics, 2021, v. 40 n. 1, p. 1-15 How to Cite?
AbstractWe introduce MotioNet, a deep neural network that directly reconstructs the motion of a 3D human skeleton from a monocular video. While previous methods rely on either rigging or inverse kinematics (IK) to associate a consistent skeleton with temporally coherent joint rotations, our method is the first data-driven approach that directly outputs a kinematic skeleton, which is a complete, commonly used motion representation. At the crux of our approach lies a deep neural network with embedded kinematic priors, which decomposes sequences of 2D joint positions into two separate attributes: a single, symmetric skeleton encoded by bone lengths, and a sequence of 3D joint rotations associated with global root positions and foot contact labels. These attributes are fed into an integrated forward kinematics (FK) layer that outputs 3D positions, which are compared to a ground truth. In addition, an adversarial loss is applied to the velocities of the recovered rotations to ensure that they lie on the manifold of natural joint rotations. The key advantage of our approach is that it learns to infer natural joint rotations directly from the training data rather than assuming an underlying model, or inferring them from joint positions using a data-agnostic IK solver. We show that enforcing a single consistent skeleton along with temporally coherent joint rotations constrains the solution space, leading to a more robust handling of self-occlusions and depth ambiguities.
Persistent Identifierhttp://hdl.handle.net/10722/304083
ISSN
2023 Impact Factor: 7.8
2023 SCImago Journal Rankings: 7.766
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorSHI, M-
dc.contributor.authorAberman, K-
dc.contributor.authorAristidou, A-
dc.contributor.authorKomura, T-
dc.contributor.authorLischinski, D-
dc.contributor.authorCohen-Or, D-
dc.contributor.authorChen, B-
dc.date.accessioned2021-09-23T08:55:00Z-
dc.date.available2021-09-23T08:55:00Z-
dc.date.issued2021-
dc.identifier.citationACM Transactions on Graphics, 2021, v. 40 n. 1, p. 1-15-
dc.identifier.issn0730-0301-
dc.identifier.urihttp://hdl.handle.net/10722/304083-
dc.description.abstractWe introduce MotioNet, a deep neural network that directly reconstructs the motion of a 3D human skeleton from a monocular video. While previous methods rely on either rigging or inverse kinematics (IK) to associate a consistent skeleton with temporally coherent joint rotations, our method is the first data-driven approach that directly outputs a kinematic skeleton, which is a complete, commonly used motion representation. At the crux of our approach lies a deep neural network with embedded kinematic priors, which decomposes sequences of 2D joint positions into two separate attributes: a single, symmetric skeleton encoded by bone lengths, and a sequence of 3D joint rotations associated with global root positions and foot contact labels. These attributes are fed into an integrated forward kinematics (FK) layer that outputs 3D positions, which are compared to a ground truth. In addition, an adversarial loss is applied to the velocities of the recovered rotations to ensure that they lie on the manifold of natural joint rotations. The key advantage of our approach is that it learns to infer natural joint rotations directly from the training data rather than assuming an underlying model, or inferring them from joint positions using a data-agnostic IK solver. We show that enforcing a single consistent skeleton along with temporally coherent joint rotations constrains the solution space, leading to a more robust handling of self-occlusions and depth ambiguities.-
dc.languageeng-
dc.publisherAssociation for Computing Machinery, Inc. The Journal's web site is located at http://tog.acm.org-
dc.relation.ispartofACM Transactions on Graphics-
dc.rightsACM Transactions on Graphics. Copyright © Association for Computing Machinery, Inc.-
dc.rights©ACM, YYYY. This is the author's version of the work. It is posted here by permission of ACM for your personal use. Not for redistribution. The definitive version was published in PUBLICATION, {VOL#, ISS#, (DATE)} http://doi.acm.org/10.1145/nnnnnn.nnnnnn-
dc.titleMotioNet: 3D Human Motion Reconstruction from Monocular Video with Skeleton Consistency-
dc.typeArticle-
dc.identifier.emailKomura, T: taku@cs.hku.hk-
dc.identifier.authorityKomura, T=rp02741-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1145/3407659-
dc.identifier.hkuros325510-
dc.identifier.volume40-
dc.identifier.issue1-
dc.identifier.spage1-
dc.identifier.epage15-
dc.identifier.isiWOS:000604780700001-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats