Learning a Dynamic Cross-Modal Network for Multispectral Pedestrian Detection
MM 2022 - Proceedings of the 30th ACM International Conference on Multimedia
Multispectral pedestrian detection that enables continuous (day and night) localization of pedestrians has numerous applications. Existing approaches typically aggregate multispectral features by a simple element-wise operation. However, such a local feature aggregation scheme ignores the rich non-local contextual information. Further, we argue that a local tight correspondence across modalities is desired for multi-modal feature aggregation. To address these issues, we introduce a multispectral pedestrian detection framework that comprises a novel dynamic cross-modal network (DCMNet), which strives to adaptively utilize the local and non-local complementary information between multi-modal features. The proposed DCMNet consists of a local and a non-local feature aggregation module. The local module employs dynamically learned convolutions to capture local relevant information across modalities. On the other hand, the non-local module captures non-local cross-modal information by first projecting features from both modalities into the latent space and then obtaining dynamic latent feature nodes for feature aggregation. Comprehensive experiments are performed on two challenging benchmarks: KAIST and LLVIP. Experiments reveal the benefits of the proposed DCMNet, leading to consistently improved detection performance on diverse detection paradigms and backbones. When using the same backbone, our proposed detector achieves absolute gains of 1.74% and 1.90% over the baseline Cascade RCNN on the KAIST and LLVIP datasets.
dynamic learning, multi-modal fusion, pedestrian detection
Xie, Jin; Anwer, Rao Muhammad; Cholakkal, Hisham; Nie, Jing; Cao, Jiale; Laaksonen, Jorma; and Khan, Fahad Shahbaz, "Learning a Dynamic Cross-Modal Network for Multispectral Pedestrian Detection" (2022). Computer Vision Faculty Publications. 275.