File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Crowd Counting with Partial Annotations in an Image

TitleCrowd Counting with Partial Annotations in an Image
Authors
Issue Date2021
Citation
Proceedings of the IEEE International Conference on Computer Vision, 2021, p. 15550-15559 How to Cite?
AbstractTo fully leverage the data captured from different scenes with different view angles while reducing the annotation cost, this paper studies a novel crowd counting setting, i.e. only using partial annotations in each image as training data. Inspired by the repetitive patterns in the annotated and unannotated regions as well as the ones between them, we design a network with three components to tackle those unannotated regions: i) in an Unannotated Regions Characterization (URC) module, we employ a memory bank to only store the annotated features, which could help the visual features extracted from these annotated regions flow to these unannotated regions; ii) For each image, Feature Distribution Consistency (FDC) regularizes the feature distributions of annotated head and unannotated head regions to be consistent; iii) a Cross-regressor Consistency Regularization (CCR) module is designed to learn the visual features of unannotated regions in a self-supervised style. The experimental results validate the effectiveness of our proposed model under the partial annotation setting for several datasets, such as ShanghaiTech, UCF-CC-50, UCF-QNRF, NWPU-Crowd and JHU-CROWD++. With only 10% annotated regions in each image, our proposed model achieves better performance than the recent methods and baselines under semi-supervised or active learning settings on all datasets. The code is https://github.com/svip-lab/CrwodCountingPAL.
Persistent Identifierhttp://hdl.handle.net/10722/345169
ISSN
2023 SCImago Journal Rankings: 12.263

 

DC FieldValueLanguage
dc.contributor.authorXu, Yanyu-
dc.contributor.authorZhong, Ziming-
dc.contributor.authorLian, Dongze-
dc.contributor.authorLi, Jing-
dc.contributor.authorLi, Zhengxin-
dc.contributor.authorXu, Xinxing-
dc.contributor.authorGao, Shenghua-
dc.date.accessioned2024-08-15T09:25:40Z-
dc.date.available2024-08-15T09:25:40Z-
dc.date.issued2021-
dc.identifier.citationProceedings of the IEEE International Conference on Computer Vision, 2021, p. 15550-15559-
dc.identifier.issn1550-5499-
dc.identifier.urihttp://hdl.handle.net/10722/345169-
dc.description.abstractTo fully leverage the data captured from different scenes with different view angles while reducing the annotation cost, this paper studies a novel crowd counting setting, i.e. only using partial annotations in each image as training data. Inspired by the repetitive patterns in the annotated and unannotated regions as well as the ones between them, we design a network with three components to tackle those unannotated regions: i) in an Unannotated Regions Characterization (URC) module, we employ a memory bank to only store the annotated features, which could help the visual features extracted from these annotated regions flow to these unannotated regions; ii) For each image, Feature Distribution Consistency (FDC) regularizes the feature distributions of annotated head and unannotated head regions to be consistent; iii) a Cross-regressor Consistency Regularization (CCR) module is designed to learn the visual features of unannotated regions in a self-supervised style. The experimental results validate the effectiveness of our proposed model under the partial annotation setting for several datasets, such as ShanghaiTech, UCF-CC-50, UCF-QNRF, NWPU-Crowd and JHU-CROWD++. With only 10% annotated regions in each image, our proposed model achieves better performance than the recent methods and baselines under semi-supervised or active learning settings on all datasets. The code is https://github.com/svip-lab/CrwodCountingPAL.-
dc.languageeng-
dc.relation.ispartofProceedings of the IEEE International Conference on Computer Vision-
dc.titleCrowd Counting with Partial Annotations in an Image-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/ICCV48922.2021.01528-
dc.identifier.scopuseid_2-s2.0-85125689863-
dc.identifier.spage15550-
dc.identifier.epage15559-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats