Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

Spatiotemporal Dilated Convolution With Uncertain Matching for Video-Based Crowd Estimation

Published: 01 January 2022 Publication History

Abstract

In this paper, we propose a novel SpatioTemporal convolutional Dense Network (STDNet) to address the video-based crowd counting problem, which contains the decomposition of 3D convolution and the 3D spatiotemporal dilated dense convolution to alleviate the rapid growth of the model size caused by the Conv3D layer. Moreover, since the dilated convolution extracts the multiscale features, we combine the dilated convolution with the channel attention block to enhance the feature representations. Due to the error that occurs from the difficulty of labeling crowds, especially for videos, imprecise or standard-inconsistent labels may lead to poor convergence for the model. To address this issue, we further propose a new patch-wise regression loss (PRL) to improve the original pixel-wise loss. Experimental results on three video-based benchmarks, i.e., the UCSD, Mall and WorldExpo&#x2019;10 datasets, show that STDNet outperforms both image- and video-based state-of-the-art methods. The source codes are released at <uri>https://github.com/STDNet/STDNet</uri>.

Cited By

View all

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image IEEE Transactions on Multimedia
IEEE Transactions on Multimedia  Volume 24, Issue
2022
2475 pages

Publisher

IEEE Press

Publication History

Published: 01 January 2022

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 29 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Multi-scale fusion public gathering recognition based on residual networkJournal of Intelligent & Fuzzy Systems: Applications in Engineering and Technology10.3233/JIFS-23681146:2(3881-3893)Online publication date: 14-Feb-2024
  • (2024)Efficient Crowd Counting via Dual Knowledge DistillationIEEE Transactions on Image Processing10.1109/TIP.2023.334360933(569-583)Online publication date: 1-Jan-2024
  • (2024)MetaUSACCExpert Systems with Applications: An International Journal10.1016/j.eswa.2024.123228247:COnline publication date: 1-Aug-2024
  • (2024)Dual-branch counting method for dense crowd based on self-attention mechanismExpert Systems with Applications: An International Journal10.1016/j.eswa.2023.121272236:COnline publication date: 1-Feb-2024
  • (2023)Language-guided Residual Graph Attention Network and Data Augmentation for Visual GroundingACM Transactions on Multimedia Computing, Communications, and Applications10.1145/360455720:1(1-23)Online publication date: 14-Jun-2023
  • (2023)Striking a Balance: Unsupervised Cross-Domain Crowd Counting via Knowledge DiffusionProceedings of the 31st ACM International Conference on Multimedia10.1145/3581783.3611797(6520-6529)Online publication date: 26-Oct-2023
  • (2023)DAOT: Domain-Agnostically Aligned Optimal Transport for Domain-Adaptive Crowd CountingProceedings of the 31st ACM International Conference on Multimedia10.1145/3581783.3611793(4319-4329)Online publication date: 26-Oct-2023
  • (2023)DilateFormer: Multi-Scale Dilated Transformer for Visual RecognitionIEEE Transactions on Multimedia10.1109/TMM.2023.324361625(8906-8919)Online publication date: 1-Jan-2023
  • (2023)Need Only One More Point (NOOMP): Perspective Adaptation Crowd Counting in Complex ScenesIEEE Transactions on Multimedia10.1109/TMM.2022.323033725(1414-1426)Online publication date: 1-Jan-2023
  • (2023)Transportation Object Counting With Graph-Based Adaptive Auxiliary LearningIEEE Transactions on Intelligent Transportation Systems10.1109/TITS.2022.322650424:3(3422-3437)Online publication date: 1-Mar-2023
  • Show More Cited By

View Options

View options

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media