File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: GIMO: Gaze-Informed Human Motion Prediction in Context

TitleGIMO: Gaze-Informed Human Motion Prediction in Context
Authors
Issue Date2022
PublisherSpringer
Citation
17th European Conference on Computer Vision (ECCV 2022), Tel Aviv, Israel, 23-27 October 2022. In Avidan, S, Brostow, G, Cissé, M, et al. (Eds.), Computer Vision – ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23-27, 2022, Proceedings, Part XIII, p. 676-694. Cham: Springer, 2022 How to Cite?
AbstractPredicting human motion is critical for assistive robots and AR/VR applications, where the interaction with humans needs to be safe and comfortable. Meanwhile, an accurate prediction depends on understanding both the scene context and human intentions. Even though many works study scene-aware human motion prediction, the latter is largely underexplored due to the lack of ego-centric views that disclose human intent and the limited diversity in motion and scenes. To reduce the gap, we propose a large-scale human motion dataset that delivers high-quality body pose sequences, scene scans, as well as ego-centric views with the eye gaze that serves as a surrogate for inferring human intent. By employing inertial sensors for motion capture, our data collection is not tied to specific scenes, which further boosts the motion dynamics observed from our subjects. We perform an extensive study of the benefits of leveraging the eye gaze for ego-centric human motion prediction with various state-of-the-art architectures. Moreover, to realize the full potential of the gaze, we propose a novel network architecture that enables bidirectional communication between the gaze and motion branches. Our network achieves the top performance in human motion prediction on the proposed dataset, thanks to the intent information from eye gaze and the denoised gaze feature modulated by the motion. Code and data can be found at https://github.com/y-zheng18/GIMO.
Persistent Identifierhttp://hdl.handle.net/10722/325584
ISBN
ISSN
2023 SCImago Journal Rankings: 0.606
ISI Accession Number ID
Series/Report no.Lecture Notes in Computer Science ; 13673

 

DC FieldValueLanguage
dc.contributor.authorZheng, Yang-
dc.contributor.authorYang, Yanchao-
dc.contributor.authorMo, Kaichun-
dc.contributor.authorLi, Jiaman-
dc.contributor.authorYu, Tao-
dc.contributor.authorLiu, Yebin-
dc.contributor.authorLiu, C. Karen-
dc.contributor.authorGuibas, Leonidas J.-
dc.date.accessioned2023-02-27T07:34:33Z-
dc.date.available2023-02-27T07:34:33Z-
dc.date.issued2022-
dc.identifier.citation17th European Conference on Computer Vision (ECCV 2022), Tel Aviv, Israel, 23-27 October 2022. In Avidan, S, Brostow, G, Cissé, M, et al. (Eds.), Computer Vision – ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23-27, 2022, Proceedings, Part XIII, p. 676-694. Cham: Springer, 2022-
dc.identifier.isbn9783031197772-
dc.identifier.issn0302-9743-
dc.identifier.urihttp://hdl.handle.net/10722/325584-
dc.description.abstractPredicting human motion is critical for assistive robots and AR/VR applications, where the interaction with humans needs to be safe and comfortable. Meanwhile, an accurate prediction depends on understanding both the scene context and human intentions. Even though many works study scene-aware human motion prediction, the latter is largely underexplored due to the lack of ego-centric views that disclose human intent and the limited diversity in motion and scenes. To reduce the gap, we propose a large-scale human motion dataset that delivers high-quality body pose sequences, scene scans, as well as ego-centric views with the eye gaze that serves as a surrogate for inferring human intent. By employing inertial sensors for motion capture, our data collection is not tied to specific scenes, which further boosts the motion dynamics observed from our subjects. We perform an extensive study of the benefits of leveraging the eye gaze for ego-centric human motion prediction with various state-of-the-art architectures. Moreover, to realize the full potential of the gaze, we propose a novel network architecture that enables bidirectional communication between the gaze and motion branches. Our network achieves the top performance in human motion prediction on the proposed dataset, thanks to the intent information from eye gaze and the denoised gaze feature modulated by the motion. Code and data can be found at https://github.com/y-zheng18/GIMO.-
dc.languageeng-
dc.publisherSpringer-
dc.relation.ispartofComputer Vision – ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23-27, 2022, Proceedings, Part XIII-
dc.relation.ispartofseriesLecture Notes in Computer Science ; 13673-
dc.titleGIMO: Gaze-Informed Human Motion Prediction in Context-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1007/978-3-031-19778-9_39-
dc.identifier.scopuseid_2-s2.0-85142707323-
dc.identifier.spage676-
dc.identifier.epage694-
dc.identifier.eissn1611-3349-
dc.identifier.isiWOS:000897100100039-
dc.publisher.placeCham-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats