File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: TransRAC: Encoding Multi-scale Temporal Correlation with Transformers for Repetitive Action Counting

TitleTransRAC: Encoding Multi-scale Temporal Correlation with Transformers for Repetitive Action Counting
Authors
KeywordsAction and event recognition
Datasets and evaluation
Face and gestures
Others
Pose estimation and tracking
Issue Date2022
Citation
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2022, v. 2022-June, p. 18991-19000 How to Cite?
AbstractCounting repetitive actions are widely seen in human activities such as physical exercise. Existing methods focus on performing repetitive action counting in short videos, which is tough for dealing with longer videos in more realistic scenarios. In the data-driven era, the degradation of such generalization capability is mainly attributed to the lack of long video datasets. To complement this margin, we introduce a new large-scale repetitive action counting dataset covering a wide variety of video lengths, along with more realistic situations where action interruption or action inconsistencies occur in the video. Besides, we also provide a fine-grained annotation of the action cycles instead of just counting annotation along with a numerical value. Such a dataset contains 1,451 videos with about 20,000 annotations, which is more challenging. For repetitive action counting towards more realistic scenarios, we further propose encoding multi-scale temporal correlation with transformers that can take into account both performance and efficiency. Furthermore, with the help of fine-grained annotation of action cycles, we propose a density map regression-based method to predict the action period, which yields better performance with sufficient interpretability. Our proposed method outperforms state-of-the-art methods on all datasets and also achieves better performance on the unseen dataset without fine-tuning. The dataset and code are available 11https://svip-lab.github.io/dataset/RepCount_dataset.html.
Persistent Identifierhttp://hdl.handle.net/10722/345287
ISSN
2023 SCImago Journal Rankings: 10.331

 

DC FieldValueLanguage
dc.contributor.authorHu, Huazhang-
dc.contributor.authorDong, Sixun-
dc.contributor.authorZhao, Yiqun-
dc.contributor.authorLian, Dongze-
dc.contributor.authorLi, Zhengxin-
dc.contributor.authorGao, Shenghua-
dc.date.accessioned2024-08-15T09:26:24Z-
dc.date.available2024-08-15T09:26:24Z-
dc.date.issued2022-
dc.identifier.citationProceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2022, v. 2022-June, p. 18991-19000-
dc.identifier.issn1063-6919-
dc.identifier.urihttp://hdl.handle.net/10722/345287-
dc.description.abstractCounting repetitive actions are widely seen in human activities such as physical exercise. Existing methods focus on performing repetitive action counting in short videos, which is tough for dealing with longer videos in more realistic scenarios. In the data-driven era, the degradation of such generalization capability is mainly attributed to the lack of long video datasets. To complement this margin, we introduce a new large-scale repetitive action counting dataset covering a wide variety of video lengths, along with more realistic situations where action interruption or action inconsistencies occur in the video. Besides, we also provide a fine-grained annotation of the action cycles instead of just counting annotation along with a numerical value. Such a dataset contains 1,451 videos with about 20,000 annotations, which is more challenging. For repetitive action counting towards more realistic scenarios, we further propose encoding multi-scale temporal correlation with transformers that can take into account both performance and efficiency. Furthermore, with the help of fine-grained annotation of action cycles, we propose a density map regression-based method to predict the action period, which yields better performance with sufficient interpretability. Our proposed method outperforms state-of-the-art methods on all datasets and also achieves better performance on the unseen dataset without fine-tuning. The dataset and code are available 11https://svip-lab.github.io/dataset/RepCount_dataset.html.-
dc.languageeng-
dc.relation.ispartofProceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition-
dc.subjectAction and event recognition-
dc.subjectDatasets and evaluation-
dc.subjectFace and gestures-
dc.subjectOthers-
dc.subjectPose estimation and tracking-
dc.titleTransRAC: Encoding Multi-scale Temporal Correlation with Transformers for Repetitive Action Counting-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/CVPR52688.2022.01843-
dc.identifier.scopuseid_2-s2.0-85141796445-
dc.identifier.volume2022-June-
dc.identifier.spage18991-
dc.identifier.epage19000-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats