File Download

There are no files associated with this item.

Conference Paper: Learning Transferable Spatiotemporal Representations from Natural Script Knowledge

TitleLearning Transferable Spatiotemporal Representations from Natural Script Knowledge
Authors
Issue Date17-Jun-2023
Abstract

t Pre-training on large-scale video data has become a common recipe for learning transferable spatiotemporal representations in recent years. Despite some progress, existing methods are mostly limited to highly curated datasets (e.g., K400) and exhibit unsatisfactory out-of-the-box representations. We argue that it is due to the fact that they only capture pixel-level knowledge rather than spatiotemporal semantics, which hinders further progress in video understanding. Inspired by the great success of imagetext pre-training (e.g., CLIP), we take the first step to exploit language semantics to boost transferable spatiotemporal representation learning. We introduce a new pretext task, Turning to Video for Transcript Sorting (TVTS), which sorts shuffled ASR scripts by attending to learned video representations. We do not rely on descriptive captions and learn purely from video, i.e., leveraging the natural transcribed speech knowledge to provide noisy but useful semantics over time. Our method enforces the vision model to contextualize what is happening over time so that it can re-organize the narrative transcripts, and can seamlessly apply to large-scale uncurated video data in the real world. Our method demonstrates strong out-of-the-box spatiotemporal representations on diverse benchmarks, e.g., +13.6% gains over VideoMAE on SSV2 via linear probing. The code is available at https://github.com/ TencentARC/TVT.


Persistent Identifierhttp://hdl.handle.net/10722/337771

 

DC FieldValueLanguage
dc.contributor.authorZeng, Ziyun-
dc.contributor.authorGe, Yuying-
dc.contributor.authorLiu, Xihui-
dc.contributor.authorChen, Bin-
dc.contributor.authorLuo, Ping-
dc.contributor.authorXia, Shu-Tao-
dc.contributor.authorGe, Yixiao-
dc.date.accessioned2024-03-11T10:23:46Z-
dc.date.available2024-03-11T10:23:46Z-
dc.date.issued2023-06-17-
dc.identifier.urihttp://hdl.handle.net/10722/337771-
dc.description.abstract<p>t Pre-training on large-scale video data has become a common recipe for learning transferable spatiotemporal representations in recent years. Despite some progress, existing methods are mostly limited to highly curated datasets (e.g., K400) and exhibit unsatisfactory out-of-the-box representations. We argue that it is due to the fact that they only capture pixel-level knowledge rather than spatiotemporal semantics, which hinders further progress in video understanding. Inspired by the great success of imagetext pre-training (e.g., CLIP), we take the first step to exploit language semantics to boost transferable spatiotemporal representation learning. We introduce a new pretext task, Turning to Video for Transcript Sorting (TVTS), which sorts shuffled ASR scripts by attending to learned video representations. We do not rely on descriptive captions and learn purely from video, i.e., leveraging the natural transcribed speech knowledge to provide noisy but useful semantics over time. Our method enforces the vision model to contextualize what is happening over time so that it can re-organize the narrative transcripts, and can seamlessly apply to large-scale uncurated video data in the real world. Our method demonstrates strong out-of-the-box spatiotemporal representations on diverse benchmarks, e.g., +13.6% gains over VideoMAE on SSV2 via linear probing. The code is available at https://github.com/ TencentARC/TVT.<br></p>-
dc.languageeng-
dc.relation.ispartof2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (17/06/2023-24/06/2023, Vancouver, BC, Canada)-
dc.titleLearning Transferable Spatiotemporal Representations from Natural Script Knowledge-
dc.typeConference_Paper-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats