Deep RGB-D Saliency Detection Without Depth
Document Type
Article
Publication Title
IEEE Transactions on Multimedia
Abstract
The existing saliency detection models based on RGB colors only leverage appearance cues to detect salient objects. Depth information also plays a very important role in visual saliency detection and can supply complementary cues for saliency detection. Although many RGB-D saliency models have been proposed, they require to acquire depth data, which is expensive and not easy to get. In this paper, we propose to estimate depth information from monocular RGB images and leverage the intermediate depth features to enhance the saliency detection performance in a deep neural network framework. Specifically, we first use an encoder network to extract common features from each RGB image and then build two decoder networks for depth estimation and saliency detection, respectively. The depth decoder features can be fused with the RGB saliency features to enhance their capability. Furthermore, we also propose a novel dense multiscale fusion model to densely fuse multiscale depth and RGB features based on the dense ASPP model. A new global context branch is also added to boost the multiscale features. Experimental results demonstrate that the added depth cues and the proposed fusion model can both improve the saliency detection performance. Finally, our model not only outperforms state-of-the-art RGB saliency models, but also achieves comparable results compared with state-of-the-art RGB-D saliency models.
First Page
755
Last Page
767
DOI
10.1109/TMM.2021.3058788
Publication Date
1-1-2022
Keywords
Convolutional neural network, depth estimation, feature fusion, saliency detection
Recommended Citation
Y. -f. Zhang et al., "Deep RGB-D Saliency Detection Without Depth," in IEEE Transactions on Multimedia, vol. 24, pp. 755-767, 2022, doi: 10.1109/TMM.2021.3058788.
Comments
IR Deposit conditions:
OA version (pathway a) Accepted version
No embargo
When accepted for publication, set statement to accompany deposit (see policy)
Must link to publisher version with DOI
Publisher copyright and source must be acknowledged