File Download

There are no files associated with this item.

Supplementary

Conference Paper: Bridging video-text retrieval with multiple choice questions

TitleBridging video-text retrieval with multiple choice questions
Authors
Issue Date2022
PublisherIEEE.
Citation
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (Hybrid), New Orleans, United States, June 21, 2022, p. 16167-16176 How to Cite?
AbstractPre-training a model to learn transferable video-text representation for retrieval has attracted a lot of attention in recent years. Previous dominant works mainly adopt two separate encoders for efficient retrieval, but ignore local associations between videos and texts. Another line of research uses a joint encoder to interact video with texts, but results in low efficiency since each text-video pair needs to be fed into the model. In this work, we enable fine-grained video-text interactions while maintaining high efficiency for retrieval via a novel pretext task, dubbed as Multiple Choice Questions (MCQ), where a parametric module BridgeFormer is trained to answer the 'questions' constructed by the text features via resorting to the video features. Specifically, we exploit the rich semantics of text (i.e., nouns and verbs) to build questions, with which the video encoder can be trained to capture more regional content and temporal dynamics. In the form of questions and answers, the semantic associations between local video-text features can be properly established. BridgeFormer is able to be removed for downstream retrieval, rendering an efficient and flexible model with only two encoders. Our method outperforms state-of-the-art methods on the popular text-to-video retrieval task in five datasets with different experimental setups (i.e., zero-shot and fine-tune), including HowTo100M (one million videos). We further conduct zero-shot action recognition, which can be cast as video-to-text retrieval, and our approach also significantly surpasses its counterparts. As an additional benefit, our method achieves competitive results with much shorter pre-training videos on single-modality downstream tasks, e.g., action recognition with linear evaluation.
DescriptionOral presentation
Persistent Identifierhttp://hdl.handle.net/10722/315552

 

DC FieldValueLanguage
dc.contributor.authorGe, Y-
dc.contributor.authorGe, Y-
dc.contributor.authorLiu, D-
dc.contributor.authorLi, D-
dc.contributor.authorShan, Y-
dc.contributor.authorQie, X-
dc.contributor.authorLuo, P-
dc.date.accessioned2022-08-19T09:00:01Z-
dc.date.available2022-08-19T09:00:01Z-
dc.date.issued2022-
dc.identifier.citationIEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (Hybrid), New Orleans, United States, June 21, 2022, p. 16167-16176-
dc.identifier.urihttp://hdl.handle.net/10722/315552-
dc.descriptionOral presentation-
dc.description.abstractPre-training a model to learn transferable video-text representation for retrieval has attracted a lot of attention in recent years. Previous dominant works mainly adopt two separate encoders for efficient retrieval, but ignore local associations between videos and texts. Another line of research uses a joint encoder to interact video with texts, but results in low efficiency since each text-video pair needs to be fed into the model. In this work, we enable fine-grained video-text interactions while maintaining high efficiency for retrieval via a novel pretext task, dubbed as Multiple Choice Questions (MCQ), where a parametric module BridgeFormer is trained to answer the 'questions' constructed by the text features via resorting to the video features. Specifically, we exploit the rich semantics of text (i.e., nouns and verbs) to build questions, with which the video encoder can be trained to capture more regional content and temporal dynamics. In the form of questions and answers, the semantic associations between local video-text features can be properly established. BridgeFormer is able to be removed for downstream retrieval, rendering an efficient and flexible model with only two encoders. Our method outperforms state-of-the-art methods on the popular text-to-video retrieval task in five datasets with different experimental setups (i.e., zero-shot and fine-tune), including HowTo100M (one million videos). We further conduct zero-shot action recognition, which can be cast as video-to-text retrieval, and our approach also significantly surpasses its counterparts. As an additional benefit, our method achieves competitive results with much shorter pre-training videos on single-modality downstream tasks, e.g., action recognition with linear evaluation.-
dc.languageeng-
dc.publisherIEEE.-
dc.relation.ispartofProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022-
dc.rightsProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Copyright © IEEE.-
dc.titleBridging video-text retrieval with multiple choice questions-
dc.typeConference_Paper-
dc.identifier.emailLuo, P: pluo@hku.hk-
dc.identifier.authorityLuo, P=rp02575-
dc.identifier.hkuros335566-
dc.identifier.spage16167-
dc.identifier.epage16176-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats