Smart City Gnosys

Smart city article details

Title Crowdcl: Unsupervised Crowd Counting Network Via Contrastive Learning
ID_Doc 16685
Authors Hu Y.; Liu Y.; Cao G.; Wang J.
Year 2025
Published IEEE Internet of Things Journal, 12, 12
DOI http://dx.doi.org/10.1109/JIOT.2025.3547898
Abstract With the continuous growth of the population, crowd counting plays a crucial role in intelligent monitoring systems for the Internet of Things (IoT) and smart city development. Accurate monitoring of crowd density not only helps maintain public safety but also effectively promotes the development of smart cities. Currently, supervised crowd counting techniques have made significant progress in improving accuracy, but these methods rely on expensive manual annotations and have limited generalization performance. To address these challenges, this article proposes an unsupervised crowd counting network based on contrastive learning, named CrowdCL. CrowdCL primarily leverages image-image contrastive learning and text-image contrastive learning to achieve unsupervised crowd counting. Specifically, in image-image contrastive learning, we strengthen the network’s ability to distinguish crowd features by designing progressive occlusion strategies and patch matching strategies, effectively differentiating crowd information from background information. In text-image contrastive learning, we construct ordered textual prompts to match ordered feature maps and use modality matching loss (Lm) to guide the image encoder. Additionally, to reduce the loss of fine details and alleviate the interference of complex backgrounds, we design a coarse-grained filtering strategy during the testing phase, assigning higher weights to crowd patches with greater potential. Experiments on multiple public datasets show that CrowdCL not only achieves outstanding performance but also outperforms some fully supervised methods in cross-dataset testing. © 2014 IEEE.
Author Keywords Contrastive learning; contrastive pretrained vision-language model; crowd counting; unsupervised learning


Similar Articles


Id Similarity Authors Title Published
16111 View0.914Ilyas N.; Shahzad A.; Kim K.Convolutional-Neural Network-Based Image Crowd Counting: Review, Categorization, Analysis, And Performance EvaluationSensors (Switzerland), 20, 1 (2020)
16658 View0.891Guo X.; Song K.; Gao M.; Zhai W.; Li Q.; Jeon G.Crowd Counting In Smart City Via Lightweight Ghost Attention Pyramid NetworkFuture Generation Computer Systems, 147 (2023)
16692 View0.884Savner S.S.; Kanhangad V.Crowdnext: Boosting Weakly Supervised Crowd Counting With Dual-Path Feature Aggregation And A Robust Loss FunctionIEEE Transactions on Instrumentation and Measurement, 74 (2025)
17778 View0.87Mansouri W.; Alohali M.A.; Alqahtani H.; Alruwais N.; Alshammeri M.; Mahmud A.Deep Convolutional Neural Network-Based Enhanced Crowd Density Monitoring For Intelligent Urban Planning On Smart CitiesScientific Reports, 15, 1 (2025)
26337 View0.87Pang Y.; Ni Z.; Zhong X.Federated Learning For Crowd Counting In Smart Surveillance SystemsIEEE Internet of Things Journal, 11, 3 (2024)
11921 View0.868Xie Z.; Shao F.; Mu B.; Chen H.; Jiang Q.; Lu C.; Ho Y.-S.Bgdfnet: Bidirectional Gated And Dynamic Fusion Network For Rgb-T Crowd Counting In Smart City SystemIEEE Transactions on Instrumentation and Measurement, 73 (2024)
38647 View0.867Guo X.; Gao M.; Zhai W.; Li Q.; Pan J.; Zou G.Multiscale Aggregation Network Via Smooth Inverse Map For Crowd CountingMultimedia Tools and Applications, 83, 22 (2024)
41571 View0.867Tomar A.; Verma K.K.; Kumar P.People Counting Via Supervised Learning-Based 2D Cnn-Lr Model In Complex Crowd ImagesLecture Notes in Electrical Engineering, 1231 LNEE (2024)
16657 View0.867Dong L.; Zhang H.; Yang K.; Zhou D.; Shi J.; Ma J.Crowd Counting By Using Top-K Relations: A Mixed Ground-Truth Cnn FrameworkIEEE Transactions on Consumer Electronics, 68, 3 (2022)
61671 View0.861Wang L.; Li H.; Hu W.; Zhang X.; Qiu H.; Meng F.; Wu Q.What Happens In Crowd Scenes: A New Dataset About Crowd Scenes For Image CaptioningIEEE Transactions on Multimedia, 25 (2023)