What is it about?

Crowd counting is a popular topic with widespread applications. Currently, the biggest challenge to crowd counting is large-scale variation in objects. In this paper, we focus on overcoming this challenge by proposing a novel Attentive Encoder-Decoder Network (AEDN), which is supervised on multiple feature scales to conduct crowd counting via density estimation.

Featured Image

Why is it important?

This work has three main contributions. First, we augment the traditional encoder-decoder architecture with our proposed residual attention blocks, which, beyond skip-connected encoded features, further extend the decoded features with attentive features. AEDN is better at establishing long-range dependencies between the encoder and decoder, therefore promoting more effective fusion of multi-scale features for handling scale-variations. Second, we design a new KL-divergence based distribution loss to supervise the scale-aware structural differences between two density maps, which complements the pixel-isolated MSE loss and better optimizes AEDN to generate high-quality density maps. Third, we adopt a multi-scale supervision scheme, such that multiple KL divergences and MSE losses are deployed at all decoding stages, providing more thorough supervisions for different feature scales.

Perspectives

The task of crowd counting aims to accurately estimate the number of people contained within a crowded image. Due to its widespread application to numerous significant applications, such as crowd behavior analysis, urban planning, and visual surveillance, tremendous attention has been drawn to this line of research which resulted in fruitful solutions.

anran zhang
Beihang University

Read the Original

This page is a summary of: Multi-scale Supervised Attentive Encoder-Decoder Network for Crowd Counting, ACM Transactions on Multimedia Computing Communications and Applications, April 2020, ACM (Association for Computing Machinery),
DOI: 10.1145/3356019.
You can read the full text:

Read

Resources

Contributors

The following have contributed to this page