Rethinking feature aggregation for deep RGB-D salient object detection
Document Type
Article
Publication Title
Neurocomputing
Abstract
Two-stream UNet based architectures are widely used in deep RGB-D salient object detection (SOD) models. However, UNet only adopts a top-down decoder network to progressively aggregate high-level features with low-level ones. In this paper, we propose to enrich feature aggregation via holistic aggregation paths and an extra bottom-up decoder network. The former aggregates multi-level features holistically to learn abundant feature interactions while the latter aggregates improved low-level features with high-level features, thus promoting their representation ability. Aiming at the two-stream architecture, we propose another early aggregation scheme to aggregate and propagate multi-modal encoder features at each level, thereby improving the encoder capability. We also propose a factorized attention module to efficiently modulate the feature aggregation action for each feature node with multiple learned attention factors. Experimental results demonstrate that all of the proposed components can gradually improve RGB-D SOD results. Consequently, our final SOD model performs favorably against other state-of-the-art methods.
First Page
463
Last Page
473
DOI
10.1016/j.neucom.2020.10.079
Publication Date
1-29-2021
Keywords
Feature aggregation, Gated attention, RGB-D saliency detection, UNet
Recommended Citation
Y. Zhang et al., "Rethinking feature aggregation for deep RGB-D salient object detection," Neurocomputing, vol. 423, pp. 463 - 473, Jan 2021.
The definitive version is available at https://doi.org/10.1016/j.neucom.2020.10.079