RGB-D salient object detection based on BC2 FNet network

In the face of complex scene images, the introduction of depth information can greatly improve the performance of salient object detection. However, up-sampling and down-sampling operations in neural networks maybe blur the boundaries of objects in the saliency map, thereby reducing the performance...

Full description

Saved in:
Bibliographic Details
Main Authors: WANG Feng, CHENG Yongmei
Format: Article
Language:zho
Published: EDP Sciences 2024-12-01
Series:Xibei Gongye Daxue Xuebao
Subjects:
Online Access:https://www.jnwpu.org/articles/jnwpu/full_html/2024/06/jnwpu2024426p1135/jnwpu2024426p1135.html
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:In the face of complex scene images, the introduction of depth information can greatly improve the performance of salient object detection. However, up-sampling and down-sampling operations in neural networks maybe blur the boundaries of objects in the saliency map, thereby reducing the performance of salient object detection. Aiming at this problem, a boundary-driven cross-modal and cross-layer fusion network (BC2FNet) for RGB-D salient object detection is proposed in this paper, which preserves the boundary of the object by adding the guidance of boundary information to the cross-modal and cross-layer fusion, respectively. Firstly, a boundary generation module is designed to extract two kinds of boundary information from low-level features of RGB and depth modalities, respectively. Secondly, a boundary-driven feature selection module is designed, which is dedicated to simultaneously focusing on important feature information and preserving boundary details in the process of RGB and depth modality fusion. Finally, a boundary-driven cross-layer fusion module is proposed which simultaneously adds two kinds of boundary information in the process of up-sampling fusion on adjacent layers. By embedding this module into the top-down information fusion flow, the predicted saliency map can contain accurate objects and sharp boundaries. Simulation results on five standard RGB-D data sets show that the proposed model can achieve better performance.
ISSN:1000-2758
2609-7125