File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Locality-constrained spatial transformer network for video crowd counting

TitleLocality-constrained spatial transformer network for video crowd counting
Authors
KeywordsConvolutional neural network
Locality constrained
Spatial transformer network
Video crowd counting
Issue Date2019
Citation
Proceedings - IEEE International Conference on Multimedia and Expo, 2019, v. 2019-July, p. 814-819 How to Cite?
AbstractCompared with single image based crowd counting, video provides the spatial-temporal information of the crowd that would help improve the robustness of crowd counting. But translation, rotation and scaling of people lead to the change of density map of heads between neighbouring frames. Meanwhile, people walking in/out or being occluded in dynamic scenes leads to the change of head counts. To alleviate these issues in video crowd counting, a Locality-constrained Spatial Transformer Network (LSTN) is proposed. Specifically, we first leverage a Convolutional Neural Networks to estimate the density map for each frame. Then to relate the density maps between neighbouring frames, a Locality-constrained Spatial Transformer (LST) module is introduced to estimate the density map of next frame with that of current frame. To facilitate the performance evaluation, a large-scale video crowd counting dataset is collected, which contains 15K frames with about 394K annotated heads captured from 13 different scenes. As far as we know, it is the largest video crowd counting dataset. Extensive experiments on our dataset and other crowd counting datasets validate the effectiveness of our LSTN for crowd counting. All our dataset are released in https://github.com/sweetyy83/Lstn-fdst-dataset.
Persistent Identifierhttp://hdl.handle.net/10722/344990
ISSN
2020 SCImago Journal Rankings: 0.368

 

DC FieldValueLanguage
dc.contributor.authorFang, Yanyan-
dc.contributor.authorZhan, Biyun-
dc.contributor.authorCai, Wandi-
dc.contributor.authorGao, Shenghua-
dc.contributor.authorHu, Bo-
dc.date.accessioned2024-08-15T09:24:32Z-
dc.date.available2024-08-15T09:24:32Z-
dc.date.issued2019-
dc.identifier.citationProceedings - IEEE International Conference on Multimedia and Expo, 2019, v. 2019-July, p. 814-819-
dc.identifier.issn1945-7871-
dc.identifier.urihttp://hdl.handle.net/10722/344990-
dc.description.abstractCompared with single image based crowd counting, video provides the spatial-temporal information of the crowd that would help improve the robustness of crowd counting. But translation, rotation and scaling of people lead to the change of density map of heads between neighbouring frames. Meanwhile, people walking in/out or being occluded in dynamic scenes leads to the change of head counts. To alleviate these issues in video crowd counting, a Locality-constrained Spatial Transformer Network (LSTN) is proposed. Specifically, we first leverage a Convolutional Neural Networks to estimate the density map for each frame. Then to relate the density maps between neighbouring frames, a Locality-constrained Spatial Transformer (LST) module is introduced to estimate the density map of next frame with that of current frame. To facilitate the performance evaluation, a large-scale video crowd counting dataset is collected, which contains 15K frames with about 394K annotated heads captured from 13 different scenes. As far as we know, it is the largest video crowd counting dataset. Extensive experiments on our dataset and other crowd counting datasets validate the effectiveness of our LSTN for crowd counting. All our dataset are released in https://github.com/sweetyy83/Lstn-fdst-dataset.-
dc.languageeng-
dc.relation.ispartofProceedings - IEEE International Conference on Multimedia and Expo-
dc.subjectConvolutional neural network-
dc.subjectLocality constrained-
dc.subjectSpatial transformer network-
dc.subjectVideo crowd counting-
dc.titleLocality-constrained spatial transformer network for video crowd counting-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/ICME.2019.00145-
dc.identifier.scopuseid_2-s2.0-85071039008-
dc.identifier.volume2019-July-
dc.identifier.spage814-
dc.identifier.epage819-
dc.identifier.eissn1945-788X-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats