File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Multi-level feature fusion based Locality-Constrained Spatial Transformer network for video crowd counting

TitleMulti-level feature fusion based Locality-Constrained Spatial Transformer network for video crowd counting
Authors
KeywordsConvolutional neural network
Locality-constrained spatial transformer network
Multi-level feature fusion
Video crowd counting
Issue Date2020
Citation
Neurocomputing, 2020, v. 392, p. 98-107 How to Cite?
AbstractVideo-based crowd counting can leverage the spatial-temporal information between neighboring frames, and thus this information would improve the robustness of crowd counting. Therefore, this solution is more practical than single image-based crowd counting in real applications. Since severe occlusions, translation, rotation, and scaling of persons will give rise to the change of density map of heads between neighboring frames, video-based crowd counting is a very challenging task. To alleviate these issues in video crowd counting, a Multi-Level Feature Fusion Based Locality-Constrained Spatial Transformer Network (MLSTN) is proposed, which consists of two components, namely density map regression module and Locality-Constrained Spatial Transformer (LST) module. Specifically, we first estimate the density map of each frame by utilizing the combination of the low-level, middle-level and high-level features of the Convolutional Neural Networks. This is because the low-level features may be more effective in the extraction of small head information, while the middle and high level features are more effective in the extraction of medium and large head information. Then to measure the relationship of the density maps between neighboring frames, the LST module is proposed, which estimates the density map of the next frame by concatenating several regression density maps. To facilitate the performance evaluation for video crowd counting, we have collected and labeled a large-scale video crowd counting dataset which includes 100 five-second-long sequences with 394,081 annotated heads from 13 different scenes. As far as we know, it is the largest video crowd counting dataset. Extensive experiments show the effectiveness of our proposed approach for crowd counting on our dataset and other video-based crowd counting datasets. All our dataset are released online.1
Persistent Identifierhttp://hdl.handle.net/10722/345110
ISSN
2023 Impact Factor: 5.5
2023 SCImago Journal Rankings: 1.815

 

DC FieldValueLanguage
dc.contributor.authorFang, Yanyan-
dc.contributor.authorGao, Shenghua-
dc.contributor.authorLi, Jing-
dc.contributor.authorLuo, Weixin-
dc.contributor.authorHe, Linfang-
dc.contributor.authorHu, Bo-
dc.date.accessioned2024-08-15T09:25:19Z-
dc.date.available2024-08-15T09:25:19Z-
dc.date.issued2020-
dc.identifier.citationNeurocomputing, 2020, v. 392, p. 98-107-
dc.identifier.issn0925-2312-
dc.identifier.urihttp://hdl.handle.net/10722/345110-
dc.description.abstractVideo-based crowd counting can leverage the spatial-temporal information between neighboring frames, and thus this information would improve the robustness of crowd counting. Therefore, this solution is more practical than single image-based crowd counting in real applications. Since severe occlusions, translation, rotation, and scaling of persons will give rise to the change of density map of heads between neighboring frames, video-based crowd counting is a very challenging task. To alleviate these issues in video crowd counting, a Multi-Level Feature Fusion Based Locality-Constrained Spatial Transformer Network (MLSTN) is proposed, which consists of two components, namely density map regression module and Locality-Constrained Spatial Transformer (LST) module. Specifically, we first estimate the density map of each frame by utilizing the combination of the low-level, middle-level and high-level features of the Convolutional Neural Networks. This is because the low-level features may be more effective in the extraction of small head information, while the middle and high level features are more effective in the extraction of medium and large head information. Then to measure the relationship of the density maps between neighboring frames, the LST module is proposed, which estimates the density map of the next frame by concatenating several regression density maps. To facilitate the performance evaluation for video crowd counting, we have collected and labeled a large-scale video crowd counting dataset which includes 100 five-second-long sequences with 394,081 annotated heads from 13 different scenes. As far as we know, it is the largest video crowd counting dataset. Extensive experiments show the effectiveness of our proposed approach for crowd counting on our dataset and other video-based crowd counting datasets. All our dataset are released online.1-
dc.languageeng-
dc.relation.ispartofNeurocomputing-
dc.subjectConvolutional neural network-
dc.subjectLocality-constrained spatial transformer network-
dc.subjectMulti-level feature fusion-
dc.subjectVideo crowd counting-
dc.titleMulti-level feature fusion based Locality-Constrained Spatial Transformer network for video crowd counting-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1016/j.neucom.2020.01.087-
dc.identifier.scopuseid_2-s2.0-85079630205-
dc.identifier.volume392-
dc.identifier.spage98-
dc.identifier.epage107-
dc.identifier.eissn1872-8286-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats