Next Article in Journal
GNSS Spoofing Detection via the Intersection Angle between Two Directions of Arrival in a Single Rotating Antenna
Previous Article in Journal
Intelligent Evacuation Sign Control Mechanism in IoT-Enabled Multi-Floor Multi-Exit Buildings
Previous Article in Special Issue
Remote Photoplethysmography and Motion Tracking Convolutional Neural Network with Bidirectional Long Short-Term Memory: Non-Invasive Fatigue Detection Method Based on Multi-Modal Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SLMSF-Net: A Semantic Localization and Multi-Scale Fusion Network for RGB-D Salient Object Detection

School of Information and Electronic Engineering, Zhejiang University of Science and Technology, Hangzhou 310023, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(4), 1117; https://doi.org/10.3390/s24041117
Submission received: 9 January 2024 / Revised: 2 February 2024 / Accepted: 6 February 2024 / Published: 8 February 2024
(This article belongs to the Special Issue Deep Learning for Information Fusion and Pattern Recognition)

Abstract

:
Salient Object Detection (SOD) in RGB-D images plays a crucial role in the field of computer vision, with its central aim being to identify and segment the most visually striking objects within a scene. However, optimizing the fusion of multi-modal and multi-scale features to enhance detection performance remains a challenge. To address this issue, we propose a network model based on semantic localization and multi-scale fusion (SLMSF-Net), specifically designed for RGB-D SOD. Firstly, we designed a Deep Attention Module (DAM), which extracts valuable depth feature information from both channel and spatial perspectives and efficiently merges it with RGB features. Subsequently, a Semantic Localization Module (SLM) is introduced to enhance the top-level modality fusion features, enabling the precise localization of salient objects. Finally, a Multi-Scale Fusion Module (MSF) is employed to perform inverse decoding on the modality fusion features, thus restoring the detailed information of the objects and generating high-precision saliency maps. Our approach has been validated across six RGB-D salient object detection datasets. The experimental results indicate an improvement of 0.20~1.80%, 0.09~1.46%, 0.19~1.05%, and 0.0002~0.0062, respectively in maxF, maxE, S, and MAE metrics, compared to the best competing methods (AFNet, DCMF, and C2DFNet).

1. Introduction

Salient Object Detection (SOD) plays a crucial role in the field of computer vision, with its primary objective being the identification and accentuation of the most visually engaging objects within a scene [1,2]. These objects typically draw the majority of observer attention and play a vital role in image and video processing tasks, such as object tracking [3,4], image segmentation [5,6], and scene understanding [7,8]. With the rapid advancement of depth sensor technology, RGB-D salient object detection has elicited significant interest among researchers. Compared to using only RGB images, RGB-D datasets offer a richer array of information, including color and depth details, which are invaluable in enhancing the performance of salient object detection. However, the achievement of accurate salient object detection under complex scenarios, with multi-scale objects and noise interference, continues to present a substantial challenge. Current research is confronted with two main issues [9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37]:
  • The Modality Fusion Problem: Undoubtedly, depth information opens up significant possibilities for enhancing detection performance. The distance information it provides between objects aids in clearly distinguishing the foreground from the background, thereby endowing the algorithm with robustness when dealing with complex scenarios. However, an urgent challenge that remains to be solved is how to fully exploit this depth information and effectively integrate it with the color, texture, and other features of RGB images to extract richer and more discriminative features. This challenge becomes particularly pressing when dealing with issues of incomplete depth information and noise interference, which necessitate further exploration and research.
  • The Multi-level Feature Integration Problem: To more effectively integrate multi-level features, it’s vital to fully consider the characteristics of both high-level and low-level features. High-level features contain discriminative semantic information, which aids in the localization of salient objects, while low-level features are rich in detailed information, beneficial for optimizing object edges. Traditional RGB-D salient object detection methods often fuse features from different levels directly, disregarding their inherent differences. This approach can lead to semantic information loss and make the method vulnerable to noise and background interference. Therefore, there is a need to explore more refined feature fusion techniques that fully take into account the characteristics of different levels of features, aiming to boost the performance of salient object detection.
To address the aforementioned challenges, we propose a Semantic Localization and Multi-Scale Fusion Network (SLMSF-Net) for RGB-D salient object detection. SLMSF-Net constitutes two stages: encoding and decoding. During the encoding phase, SLMSF-Net utilizes the ResNet50 network to separately extract features from RGB and depth images and employs a depth attention module for modal feature fusion. In the decoding phase, SLMSF-Net first accurately localizes salient objects through a semantic localization module, and then constructs a reverse decoder using a Multi-Scale Fusion Module to restore the detailed information of the salient objects. Our main contributions can be summarized as follows:
  • We propose a depth attention module that leverages channel and spatial attention mechanisms to fully explore the effective information of depth images and enhance the matching ability between RGB and depth feature maps.
  • We propose a semantic localization module that constructs a global view for the precise localization of salient objects.
  • We propose a reverse decoding network based on multi-scale fusion, which implements reverse decoding on modal fusion features and generates detailed information on salient objects through multi-scale feature fusion.
The design of the SLMSF-Net is poised to address key issues in the current RGB-D salient object detection domain and provide new research insights for other tasks within the field of computer vision. Extensive experimental results fully demonstrate that the SLMSF-Net exhibits excellent performance in RGB-D SOD tasks, enhancing the accuracy and effectiveness of salient object detection.

2. Related Works

In this section, we will review research works [17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37] related to the RGB-D salient object detection method that we propose. These related studies can be broadly divided into two categories: salient object detection based on RGB images and salient object detection based on RGB-D images.

2.1. Salient Object Detection Based on RGB Images

Salient object detection based on RGB images mainly focuses on visual cues such as color, texture, and contrast. Early saliency detection methods primarily depended on handcrafted features and heuristic rules. For instance, Itti et al. [17] proposed a saliency detection model based on the biological visual system, which estimates saliency by calculating the local contrast of color, brightness, and directional features. Achanta et al. [18] introduced a frequency-tuned salient region detection method, which extracts global contrast features in the frequency domain of the image to detect salient regions. Tong et al. [19] combined global and local cues for salient object detection, using a variety of cues (such as color, texture, and contrast) to handle complex scenarios.
In recent years, deep learning technology has achieved significant success in the field of salient object detection. Models such as the deep learning saliency model proposed by Chen et al. [20] accomplish hierarchical representation of saliency features to realize end-to-end salient object detection. Cong et al. [21] proposed a salient object detection method based on a Fully Convolutional Network (FCN), which uses global contextual information and local detail information for saliency prediction. Hou et al. [22] developed a deeply supervised network for salient object detection, improving upon the Holistically Nested Edge Detector (HED) architecture. They introduced short connections between network layers, enhancing salient object detection by combining low-level and high-level features. Zhao et al. [23] proposed GateNet, a new network architecture for salient object detection. This model introduced multilevel gate units to balance encoder block contributions, suppressing non-salient features and contextualizing for the decoder. They also included Fold-ASPP to gather multiscale semantic information, enhancing atrous convolution for better feature extraction. Zhang et al. [24] combined neural network layer features to improve salient object detection accuracy in images. Their approach used both coarse and fine image details and incorporated edge-aware maps to enhance boundary detection. Wu et al. [25] proposed a cascaded partial decoder that discarded low-level features to reduce computational complexity while refining high-level features for accuracy.
Moreover, some researchers have applied attention mechanisms to RGB-based salient object detection models, such as [26,27,28]. These methods enable the models to concentrate their attention on the visually prominent regions of the image. Chen et al. [26] presented an approach for enhancing salient object detection through the use of reverse attention and side-output residual learning. This method aimed to refine saliency maps with a particular focus on improving resolution and reducing the model’s size. Wang et al. [27] presented PAGE-Net, a model for salient object detection. The model utilized a pyramid attention module to enhance saliency representation by incorporating multi-scale information, thereby effectively boosting detection accuracy. Additionally, it featured a salient edge detection module, which sharpened the detection of salient object boundaries. Wang et al. [28] introduced PiNet, a salient object detection model designed for enhancing feature extraction and the progressive refinement of saliency. The model incorporated level-specific feature extraction mechanisms and employed a coarse-to-fine process for refining saliency features, which helped in overcoming common issues in existing methods like noise accumulation and spatial detail dilution. Although methods based on RGB images can achieve good performance in many situations, they lack the ability to handle depth information.

2.2. Salient Object Detection Based on RGB-D Images

With the advancement of depth sensors, RGB-D images (which contain both color and depth information) have been widely applied in salient object detection. For instance, Lang et al. [29] investigated the impact of depth cues on saliency detection, where they found that depth information holds significant value for salient object detection. Based on this, many researchers have begun to explore how to fully utilize depth information for salient object detection.
Peng et al. [30] proposed a multi-modal fusion framework that improves saliency detection performance by fusing local and global depth features with color and texture features. Zhang et al. [31] presented a new RGB-D salient object detection model, addressing challenges with depth image quality and foreground–background consistency. The model introduced a two-stage approach: firstly, an image generation stage that created high-quality, foreground-consistent pseudo-depth images, and secondly, a saliency reasoning stage that utilized these images for enhanced depth feature calibration and cross-modal fusion. Ikeda et al. [32] introduced a model for RGB-D salient object detection that integrated saliency and edge features with reverse attention. This approach effectively enhanced object boundary detection and saliency in complex scenes. The model also incorporated a Multi-Scale Interactive Module for improved global image information understanding and utilized supervised learning to enhance accuracy in salient object and boundary areas. Xu et al. [33] introduced a new approach to RGB-D salient object detection, addressing the object-part relationship dilemma in Salient Object Detection (SOD). The proposed CCNet model utilized a Convolutional Capsule Network based on Feature Extraction and Integration (CCNet) to efficiently explore the object-part relationship in RGB-D SOD with reduced computational demand. Cong et al. [34] presented a comprehensive approach to RGB-D salient object detection, focusing on enhancing the interaction and integration of features from both RGB and depth modalities. It introduced a new network architecture that efficiently combined these modalities, addressing challenges in feature representation and fusion. However, these methods overlook the feature differences between different modalities, resulting in insufficient information fusion.
To address this issue, Qu et al. [35] introduced a simple yet effective deep learning model, which learns the interaction mechanism between RGB and depth-induced saliency features. Yi et al. [36] proposed a Cross-stage Multi-scale Interaction Network (CMINet), which intertwines features at different stages with the use of a Multi-scale Spatial Pooling (MSP) module and a Cross-stage Pyramid Interaction (CPI) module. They then designed an Adaptive Weight Fusion (AWF) module for balancing the importance of multi-modal features and fusing them. Liu et al. [37] proposed a cross-modal edge-guided salient object detection model for RGB-D images. This model extracts edge information from cross-modal color and depth information and integrates the edge information into cross-modal color and depth features, generating a saliency map with clear boundaries. Sun et al. [38] introduced an RGB-D salient object detection method that combined cross-modal interactive fusion with global awareness. This method embedded a transformer network within a U-Net structure to merge global attention mechanisms with local convolution, aiming for enhanced feature extraction. It utilized a U-shaped structure for extracting dual-stream features from RGB and depth images, employing a multi-level information reconstruction approach to suppress lower-layer disturbances and minimize redundant details. Peng et al. [39] introduced MFCG-Net, an RGB-D salient object detection method that leveraged multimodal fusion and contour guidance to improve detection accuracy. It incorporated attention mechanisms for feature optimization and designed an interactive feature fusion module to effectively integrate RGB and depth image features. Additionally, the method utilized contour features to guide the detection process, achieving clearer boundaries for salient objects. Sun et al. [40] introduced a new approach for RGB-D salient object detection, leveraging a cascaded and aggregated Transformer Network structure to enhance feature extraction and fusion. They employed three key modules: the Attention Feature Enhancement Module (AFEM) for multi-scale semantic information, the Cross-Modal Fusion Module (CMFM) to address depth map quality issues, and the Cascaded Correction Decoder (CCD) to refine feature scale differences and suppress noise. Although some significant results have been achieved in existing research, it remains a formidable challenge to achieve accurate salient object detection in complex scenes through cross-modal and cross-level feature fusion.

3. Proposed Method

In this section, we first provide an overview of our method in Section 3.1. Following that, in Section 3.2, we elaborate on the depth attention module we propose, which is used to mine valuable depth information. In Section 3.3 and Section 3.4, we introduce the semantic localization module and the reverse decoding network based on the Multi-Scale Fusion Module, respectively. Finally, in Section 3.5, we discuss the loss function.

3.1. Overview of SLMSF-Net

Figure 1 displays the overall network structure of SLMSF-Net. Without loss of generality, we adopt Resnet50 [41] as the backbone network to extract features from both RGB images and depth images separately. Resnet50 encompasses five convolution stages; we removed the final pooling layer and the fully connected layer, resulting in a fully convolutional neural network, and use the outputs of the intermediate five convolution blocks as feature outputs. These output feature maps are denoted as M1, M2, M3, M4, and M5, with their sizes being 1/2, 1/4, 1/8, 1/16, and 1/32 of the original image, respectively.
  • Modal Feature Fusion: As shown in Figure 1, we proposed a Depth Attention Module. This module performs a modal fusion of RGB image features and depth image features, forming the modal fusion features F 1 F u s e , F 2 F u s e , F 3 F u s e , F 4 F u s e and F 5 F u s e .
  • Semantic Localization: We proposed a Semantic Localization Module. This module first downsamples the top-level modal fusion feature to compute a global view. It then performs coordinate localization on the global view and ultimately fuses the localization information with the global view, thereby precisely locating the salient object. Assuming the semantic localization module is represented as the SLM function, its output result can be written as: F o f = SLM ( F 5 F u s e ) .
  • Multi-Scale Fusion Decoding: After performing semantic localization, we predicted the clear boundaries of the salient object through reverse multi-level feature integration from front to back. To accomplish this multi-level feature integration, we constructed a Multi-Scale Fusion Module, which effectively fuses features at all levels.

3.2. Depth Attention Module

In the process of fusing RGB and depth features, we need to address two main issues. The first one is the modal mismatch problem, which requires us to resolve the modal differences between the two types of features. The second one is the information complementarity problem; since RGB and depth features often capture different aspects of object information, we need to consider how to let these two types of features complement each other’s information, aiming to enhance the accuracy and robustness of object detection. Inspired by [42], we designed a depth attention module to improve the matching and complementarity of multi-modal features.
Specifically, F i R G B represents the ith RGB image feature and F i D e p represents the ith depth image feature, where i is a natural number from 1 to 5. As shown in Figure 2, the depth attention module first enhances the depth image feature through channel attention. The enhanced result is then multiplied element-wise with the RGB image feature to obtain the channel-enhanced fusion feature. Following this, the channel-enhanced fusion feature undergoes spatial attention enhancement, and the enhanced result is multiplied element-wise with the RGB image feature, thus obtaining the modal fusion feature. To enhance the matching of depth features, we stacked a depth attention module behind each depth feature branch. By introducing attention units, we can enhance the saliency representation ability of depth features. The fusion process of the two modal features can be expressed as follows:
F i Fuse = F i RGB × SA ( F i RGB × CA ( F i Dep ) )
Herein, CA ( ) symbolizes the channel attention operation, SA ( ) indicates the spatial attention operation, and × represents the element-wise multiplication operation.

3.3. Semantic Localization Module

In the process of salient object localization, high-level features play a crucial role. Compared to low-level features, high-level features are capable of capturing more abstract information, which aids in highlighting the location of salient objects. Therefore, we introduced a semantic localization module designed to effectively learn the global view of the entire image, thereby achieving more precise salient object localization. As depicted in Figure 3, the semantic localization process is divided into three stages: initially, the first stage downsamples the top-level modal fusion features to compute a global view; subsequently, the second stage carries out coordinate localization on the global view; finally, the third stage fuses the localization information with the global view.
In the first stage, we implement a 1/2 scale downsample operation on the top-level modal fusion features F 5 Fuse , followed by two ConvBR 3 × 3 operations, thereby obtaining the first layer of the global feature map F 1 ov . Subsequently, we perform the same 1/2 scale downsample and two ConvBR 3 × 3 operations on the first layer of the global feature map, resulting in the second layer of the global feature map F 2 ov . As observed, these two global feature maps possess a significantly large receptive field, enabling them to serve as the global view of the entire image. The computation process for the global view can be described as follows:
F 1 ov = ConvBR 3 × 3 ( ConvBR 3 × 3 ( DownS 1 / 2 ( F 5 Fuse ) ) )
F 2 ov = ConvBR 3 × 3 ( ConvBR 3 × 3 ( DownS 1 / 2 ( F 1 ov ) ) )
Herein, DownS 1 / 2 ( ) denotes a 1/2 scale downsample operation on the input feature map. ConvBR 3 × 3 ( ) represents a convolution operation performed on the input feature map using a kernel size of 3 × 3 , followed by batch normalization and activation operations, where the activation function is Relu. This can be expressed as:
ConvBR 3 × 3 ( X ) = Relu ( BN ( Conv 3 × 3 ( X ) ) )
Herein, Conv ( ) symbolizes the convolution operation, BN ( ) denotes the batch normalization operation, and Relu ( ) represents the Relu activation function.
In the second stage, for the second layer of the global feature map F 2 ov , we utilize a pooling kernel of size (1, W) to perform average pooling along the vertical coordinate of the feature map, followed by a convolution operation with a kernel size of 1 × 1 , resulting in the height-oriented feature map T H . Simultaneously, we use a pooling kernel of size (H, 1) to conduct average pooling along the horizontal coordinate of the feature map F 2 ov , then perform a convolution operation with a kernel size of 1 × 1 , yielding the width-oriented feature map T W . This can be described as:
T H = ConvBR 1 × 1 ( 1 W 0 i < W F 2 ov ( H , i ) )
T W = ConvBR 1 × 1 ( 1 H 0 j < H F 2 ov ( j , W ) )
Herein, feature map T H extends in the width direction, while feature map T W expands in the height direction. The two expanded feature maps undergo pixel-wise multiplication, and then through a Sigmoid activation function, a coordinate localization feature map F C is formed. This can be described as:
F C = Sigmoid ( K ( T H ) × K ( T W ) )
Herein, the K ( ) operation refers to expanding the input feature map in the width or height direction to match the size of feature map A, while Sigmoid ( ) signifies the Sigmoid activation function.
In the third stage, we view the localization feature map as a self-attention mechanism for calibrating the global view. Specifically, we perform a pixel-wise multiplication operation between the localization feature map F C and the second layer of the global feature map F 2 ov , followed by a pyramid feature fusion operation on the multiplication results, yielding feature map F * of . Subsequently, we upscale the localization feature map F C twice and perform a pixel-wise multiplication operation with the first layer of the global feature map F 1 o v . The result is stacked with F * of , and then the stacked result is subjected to a pyramid feature fusion operation to finally obtain the global localization fusion feature F of . This can be described as:
F * of = Pyramid ( F C × F 2 ov )
F of = Pyramid ( concat ( UP ( F C ) × F 1 ov , F * of ) )
Herein, Pyramid ( ) represents the pyramid feature fusion operation, concat ( ) signifies the stacking operation along the channel, and UP ( ) denotes the operation of upscaling by a factor of two.
The pyramid feature fusion operation is depicted in Figure 4. Initially, we conduct a convolution operation with a kernel size 1 × 1 , adjusting the number of channels in the input feature map X to 32, which yields the feature map Y . Following this, we execute feature extraction on Y , with the specific extraction method detailed as follows:
Y = ConvBR 1 × 1 ( X )
P 1 = ConvBR 3 × 3 ( ConvBR 3 × 3 ( Y ) )
P i = ConvBR 3 × 3 2 i 1 ( ConvBR 3 × 3 ( Y ) + P i 1 ) ) ,   i ( i { 2 , 3 } )
Herein, ConvBR 3 × 3 2 i 1 ( ) represents a dilated convolution with a kernel size of 3 × 3 and a dilation rate of 2 i 1 . We perform a concatenation operation along the channel with the three extracted features. Subsequently, we conduct a convolution operation on the concatenation result with a kernel size of 1 × 1 , adjusting the channel count to match that of the input feature map. Finally, a residual connection is established with the input feature map. This process can be described as follows:
Pyramid ( X ) = ConvBR 1 × 1 ( concat ( P 1 , P 2 , P 3 , P 4 ) ) + X

3.4. Multi-Scale Fusion Module and the Reverse Decoding Process

Following semantic localization, we integrate multi-layer features in a forward-to-backward manner to delineate intricate details of the salient object. To achieve this multi-layer feature integration, we designed and constructed a Multi-Scale Fusion Module. The reverse decoder operates in five stages, each accepting the output from the preceding stage for reverse multi-scale fusion decoding. Importantly, the input for the fifth stage of the decoder is the global localization fusion feature F of . The process of the reverse decoder can be described as follows:
D e c o d e 5 * = UP ( ConvBR 1 × 1 ( F of ) )
D e c o d e 5 = MSF ( concat ( ConvBR 1 × 1 ( F 5 Fuse ) , D e c o d e 5 * ) )
D e c o d e i * = UP ( ConvBR 1 × 1 ( D e c o d e i + 1 ) )
D e c o d e i = MSF ( concat ( ConvBR 1 × 1 ( F i Fuse ) , D e c o d e i * ) ) ,   i ( i { 1 , 2 , 3 , 4 } )
Herein, MSF ( ) stands for the Multi-Scale Fusion Module. We upscale the output from the first stage of the decoder to the size of the input image, thereby obtaining the final saliency prediction map. The specific formula used to generate the saliency prediction map is as follows:
S = Sigmoid ( Conv 1 × 1 ( UP in ( D e c o d e 1 ) ) )
Herein, S represents the saliency prediction map, UP in ( ) denotes the upscaling of the feature map to the size of the input image, while Conv 1 × 1 ( ) signifies a single-channel convolution with a kernel size of 1 × 1 . The primary purpose of Conv 1 × 1 ( ) is to adjust the channel count of the feature map to 1.
As illustrated in Figure 5, the multi-scale feature fusion module comprises four parallel branches and a residual connection. Initially, we employ a convolutional operation with a kernel of size 1 × 1 to reduce the number of channels in the input feature map to 64. Following this, in the first branch, we sequentially execute a convolution with a kernel also of size 1 × 1 , followed by another with a kernel of size 3 × 3 . For the i - th ( i { 2 , 3 , 4 } ) branch of the module, the procedure commences with a convolution involving a kernel of size ( 2 i 1 ) × 1 , proceeded by another convolution with a kernel of size 1 × ( 2 i 1 ) . Finally, a dilated convolution operation with a kernel of size 3 × 3 and a dilation rate of 2 i 1 is applied. This design strategy is aimed at extracting multi-scale information from the multi-modal fusion features, thereby enriching the representational power of the model. Next, the outputs from the four branches are stacked along the channel dimension, and the channel count of the stacked output is adjusted to match the input feature map’s channel count, using a convolution operation with a kernel size of 1 × 1 . Finally, the adjusted result is connected to the input feature map via a residual connection. The entire fusion process can be described as follows:
b r a n c h 1 ( x ) = ConvBR 3 × 3 ( ConvBR 1 × 1 ( x ) )
b r a n c h i ( x ) = ConvBR 3 × 3 2 i 1 ( ConvBR 1 × ( 2 i 1 ) ( ConvBR ( 2 i 1 ) × 1 ( ConvBR 1 × 1 ( x ) ) ) ) ,   i { 2 , 3 , 4 }
MSF ( x ) = ConvBR 1 × 1 ( concat ( b r a n c h 1 ( x ) , b r a n c h 2 ( x ) , b r a n c h 3 ( x ) , b r a n c h 4 ( x ) ) ) + x
Herein, b r a n c h i ( x ) denotes the ith parallel branch, while x symbolizes the input feature map.

3.5. Loss Function

As depicted in Figure 1, at each stage of the decoder, the decoded output is upsampled to the size of the input image. Following this, a convolution operation with a single-channel convolution kernel of 1 × 1 is performed, and then a prediction saliency map is generated through a sigmoid activation function. The saliency maps predicted at each of the five stages are denoted as Oi (i = 1, 2, ⋯, 5). Following the same process, we can also generate the predicted saliency map Oof corresponding to the output of the semantic localization module. This process can be described as follows:
O i = Sigmoid ( Conv 1 × 1 ( UP in ( D e c o d e i ) ) )
O of = Sigmoid ( Conv 1 × 1 ( UP in ( F of ) ) )
Assuming the predicted saliency map is denoted as O, and the real saliency map is denoted as GT, the formula for calculating the loss value of the prediction results is as follows:
Loss ( O , G T ) = Bce ( O , G T ) + D i c e ( O , G T )
Bce ( O , G T ) = GT logO + ( 1 G T ) l o g ( 1 O )
D i c e ( O , G T ) = 1 2 GT O | | G T | | + | | O | |
Herein, Bce ( ) represents the binary cross-entropy loss function, D i c e ( ) denotes the Dice loss function [43], and | | | | represents the L 1 norm. The total loss function during the training phase is described as follows:
L = α i = 1 5 Loss ( O i , G T ) + ( 1 α ) Loss ( O of , GT )
wherein, α represents the weight coefficients. During the testing phase, O1 is the final prediction result of the model.

4. Experiments

Section 4.1 provides a detailed description of the implementation details, Section 4.2 discusses the sources of the datasets used, Section 4.3 introduces the setup of the evaluation metrics, Section 4.4 presents the comparison with the current state-of-the-art (SOTA) methods, and Section 4.5 is dedicated to the discussion of the ablation experiments. Together, these sections form the experimental analysis and evaluation part of the paper, comprehensively demonstrating the effectiveness and reliability of the research method.

4.1. Implementation Details

The salient object detection method proposed in this paper is implemented based on the Pytorch framework [44,45], and all experimental procedures were carried out on a single NVIDIA RTX A6000 GPU(NVIDIA, Santa Clara City, CA, USA). The initialization parameters of the backbone model, ResNet50, are derived from a pre-trained model on ImageNet [46]. Specifically, both the RGB image branch and the depth image branch use a ResNet50 model for feature extraction, with the only difference being that the input channel number for the depth image branch is 1. To enhance the model’s generalization capability, various augmentation strategies, such as random flipping, rotation, and boundary cropping, were applied to all training images. Throughout the training process, the Adam optimizer was employed, with parameters set to β1 = 0.9 and β2 = 0.999, and a batch size of 10. The initial learning rate was set to 1 × 10−4 and was divided by 10 every 50 rounds. The dimensions of the input images were all adjusted to 768 × 768. The model converged within 200 rounds. In order to show the training process of our model more clearly, we report the training and validation loss curve of our network in Figure 6.

4.2. Datasets

In this study, SLMSF-Net was extensively evaluated across six widely used datasets, including NJU2K [47], NLPR [30], STERE [48], SSD [49], SIP [50], and DES [51]. These datasets contain 1985, 1000, 1000, 80, 929, and 135 images, respectively. For the training phase, we utilized 1485 images from the NJU2K dataset and 700 images from the NLPR dataset. During the testing phase, the remaining images from the NJU2K and NLPR datasets, as well as the entire STERE, SSD, SIP, and DES datasets were used.

4.3. Evaluation Metrics

We employed four widely used evaluation metrics to compare SLMSF-Net with previous state-of-the-art methods, namely E-Measure, F-measure, S-measure, and MAE.
E-Measure ( E ξ ) is a saliency map evaluation method based on cognitive vision, capable of integrating statistical information at both the image level and local pixel level. This measurement strategy was proposed by [52] and is defined as follows:
E ξ = l W × H i = 1 W j = 1 H ξ ( i , j )
Here, W and H represent the width and height of the saliency map, respectively, while ξ signifies the enhanced alignment matrix. E-measure has three different variants: maximum E-measure, adaptive E-measure, and average E-measure. In our experiments, we used the maximum E-measure (maxE) as the evaluation criterion.
F-measure ( F β ) serves as a weighted harmonic mean of precision and recall. It is defined as follows:
F β = ( 1 + β 2 ) Precision × Recall β 2 × Precision + Recall
Here, β is a parameter used to balance Precision and Recall. In this study, we set β² to 0.3. Similar to E-measure, F-measure also has three different variants: maximum F-measure, adaptive F-measure, and average F-measure. In our experiments, we reported the results of the maximum F-measure (maxF).
S-measure ( S α ) is a method for evaluating structural similarity. It assesses from two perspectives: region awareness ( S r ) and object awareness ( S o ). It is defined as follows:
S α = α × S o + ( 1 α ) S r
Here, α [ 0 , 1 ] is a hyperparameter used to balance between S o and S r . In our experiments, α is set to 0.5.
MAE (Mean Absolute Error) represents the average per-pixel absolute error between the predicted saliency map S and the ground truth map GT. It is defined as follows:
MAE = 1 W × H i = 1 W j = 1 H | S ( i , j ) G T ( i , j ) |
Here, W and H, respectively, denote the width and height of the saliency map. The MAE is normalized to a value in the [0, 1] interval.

4.4. Comparison with SOTA Methods

We compared the SLMSF-Net model proposed in this study with ten deep learning-based RGB-D saliency detection methods, including AFNet [53], HINet [54], C2DFNet [55], DCMF [56], CFIDNet [57], CIR-Net [58], DCF [59], DASNet [60], D3Net [50], and ICNet [61]. To ensure a fair comparison, we used the saliency maps provided by the authors. If the saliency maps were not provided, we computed them using the source code and model files provided by the authors.

4.4.1. Quantitative Comparison

Figure 7 presents the comparison results of PR curves from different methods, while Table 1 presents the quantitative comparison results for four evaluation metrics. As shown in the figure and table, our PR curve outperforms all other comparison methods, whether on the NJU2K, NLPR, DES, SIP, SSD, or STERE datasets. This advantage is largely attributed to our designed semantic localization and multi-scale fusion strategies, which, respectively, achieve precise localization of salient objects and capture of detailed boundary information. Additionally, our designed depth attention module can effectively utilize depth information to enhance the model’s segmentation performance. Concurrently, the table data reflects the same conclusion, i.e., our method outperforms all comparison methods in performance on the NJU2K, NLPR, DES, SIP, SSD, and STERE datasets. Compared with the best comparison methods (AFNet, C2DFNet, and DCMF), we have improved the MAE, max F β , max E ξ , S α evaluation metrics by 0.0002~0.0062, 0.2~1.8%, 0.09~1.46%, and 0.19~1.05%, respectively. Therefore, both the PR curves and evaluation metrics affirm the effectiveness and superiority of our method proposed for the RGB-D SOD task.

4.4.2. Qualitative Comparison

For a qualitative comparison, we present a selection of representative visual examples in Figure 8. Upon observation, our method demonstrates superior performance in several challenging scenarios compared to other methods. Examples of these scenarios include situations where the foreground and background colors are similar (rows 1–2), in complex environments (rows 3–4), in scenes with multiple objects present (rows 5–6), for small object detection (rows 7–8), and under conditions of low-quality depth images (rows 9–10). These visual examples show that our method can more precisely locate salient objects and generate more accurate saliency maps.

4.5. Ablation Studies

As shown in Table 2, we conducted an in-depth ablation analysis to verify the effectiveness of each module. DAM represents the Deep Attention Module, SLM is the Semantic Localization Module, and MSFM stands for Multi-Scale Fusion Module. “Without DAM”, “without SLM”, and “without MSFM” refer to the models obtained after removing the DAM, SLM, and MSFM modules from the SLMSF-Net model, respectively. By comparing the data in the third column with the sixth column, we can clearly see that the introduction of the DAM module significantly improves the performance of the model. Similarly, by comparing the data in the fourth and sixth columns, we can see that the introduction of the SLM module can significantly enhance the performance of the model. Comparing the data in the fifth and sixth columns, we can see that adding the MSFM module will enhance the model’s performance. These results prove the importance of the three modules: the DAM module introduces depth image information, the SLM module realizes the precise semantic location of salient objects, and the MSFM module can fuse multi-scale features to refine the boundaries of salient objects. Each of these three functional modules resulted in a significant increase in model performance. In the last column, we can see that the SLMSF-Net model that incorporates these three modules achieved the best results.

5. Conclusions

In complex scenarios, achieving precise RGB-D salient object detection against multiple scales of objects and noisy backgrounds remains a daunting task. Current research primarily faces two major challenges: modality fusion and multi-level feature integration. To address these challenges, we propose an innovative RGB-D salient object detection network, the Semantic Localization and Multi-Scale Fusion Network (SLMSF-Net). This network comprises two main stages: encoding and decoding. In the encoding stage, SLMSF-Net utilizes ResNet50 to extract features from RGB and depth images and employs a depth attention module for the effective fusion of modal features. In the decoding stage, the network precisely locates salient objects through the semantic localization module and restores the detailed information of salient objects in the reverse decoder via the Multi-Scale Fusion Module. Rigorous experimental validation shows that SLMSF-Net exhibits superior accuracy and robustness on multiple RGB-D salient object detection datasets, outperforming existing technologies. In the future, we plan to further optimize the model, improve the attention mechanism, delve into refining edge details, and explore its application in RGB-T salient object detection tasks.

Author Contributions

Conceptualization, Y.P.; Data curation, Z.Z.; Formal analysis, M.F.; Funding acquisition, Y.P.; Investigation, M.F.; Methodology, Y.P.; Project administration, Y.P.; Resources, Z.Z.; Software, Y.P. and Z.Z.; Supervision, Y.P.; Validation, Y.P., Z.Z. and M.F.; Visualization, Z.Z.; Writing—original draft, Y.P.; Writing—review and editing, Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 61972357), the basic public welfare research program of Zhejiang Province (No. LGF22F020017 and No. GG21F010013), and the Natural Science Foundation of Zhejiang Province (No. Y21F020030).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Liu, J.J.; Hou, Q.; Liu, Z.A.; Cheng, M.M. Poolnet+: Exploring the potential of pooling for salient object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 887–904. [Google Scholar] [CrossRef] [PubMed]
  2. Liang, Y.; Qin, G.; Sun, M.; Qin, J.; Yan, J.; Zhang, Z.H. Multi-modal interactive attention and dual progressive decoding network for RGB-D/T salient object detection. Neurocomputing 2022, 490, 132–145. [Google Scholar] [CrossRef]
  3. Zakharov, I.; Ma, Y.; Henschel, M.D.; Bennett, J.; Parsons, G. Object Tracking and Anomaly Detection in Full Motion Video. In Proceedings of the IGARSS 2022, 2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 7910–7913. [Google Scholar]
  4. Zhang, Y.; Sun, P.; Jiang, Y.; Yu, D.D.; Weng, F.C.; Yuan, Z.H.; Luo, P.; Liu, W.Y.; Wang, X.G. Bytetrack: Multi-object tracking by associating every detection box. In Proceedings of the Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Springer Nature: Cham, Switzerland, 2022; pp. 1–21. [Google Scholar]
  5. Wang, R.; Lei, T.; Cui, R.; Zhang, B.T.; Meng, H.Y.; Nandi, A.K. Medical image segmentation using deep learning: A survey. IET Image Process. 2022, 16, 1243–1267. [Google Scholar] [CrossRef]
  6. He, B.; Hu, W.; Zhang, K.; Yuan, S.D.; Han, X.L.; Su, C.; Zhao, J.M.; Wang, G.Z.; Wang, G.X.; Zhang, L.Y. Image segmentation algorithm of lung cancer based on neural network model. Expert Syst. 2022, 39, e12822. [Google Scholar] [CrossRef]
  7. Fan, J.; Zheng, P.; Li, S. Vision-based holistic scene understanding towards proactive human–robot collaboration. Robot. Comput.-Integr. Manuf. 2022, 75, 102304. [Google Scholar] [CrossRef]
  8. Gong, T.; Zhou, W.; Qian, X.; Lei, J.S.; Yu, L. Global contextually guided lightweight network for RGB-thermal urban scene understanding. Eng. Appl. Artif. Intell. 2023, 117, 105510. [Google Scholar] [CrossRef]
  9. Chen, G.; Shao, F.; Chai, X.; Chen, H.; Jiang, Q.; Meng, X.; Ho, Y.S. Modality-Induced Transfer-Fusion Network for RGB-D and RGB-T Salient Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 1787–1801. [Google Scholar] [CrossRef]
  10. Gao, L.; Fu, P.; Xu, M.; Wang, T.; Liu, B. UMINet: A unified multi-modality interaction network for RGB-D and RGB-T salient object detection. Vis. Comput. 2023, 1–18. [Google Scholar] [CrossRef]
  11. Wu, Y.H.; Liu, Y.; Xu, J.; Bian, J.W.; Gu, Y.C.; Cheng, M.M. MobileSal: Extremely efficient RGB-D salient object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 10261–10269. [Google Scholar] [CrossRef]
  12. Liu, S.; Huang, D. Receptive field block net for accurate and fast object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 385–400. [Google Scholar]
  13. Zhang, N.; Han, J.; Liu, N. Learning implicit class knowledge for rgb-d co-salient object detection with transformers. IEEE Trans. Image Process. 2022, 31, 4556–4570. [Google Scholar] [CrossRef]
  14. Wu, Y.H.; Liu, Y.; Zhang, L.; Cheng, M.M.; Ren, B. EDN: Salient object detection via extremely-downsampled network. IEEE Trans. Image Process. 2022, 31, 3125–3136. [Google Scholar] [CrossRef]
  15. Wu, Z.; Li, S.; Chen, C.; Hao, A.; Qin, H. Recursive multi-model complementary deep fusion for robust salient object detection via parallel sub-networks. Pattern Recognit. 2022, 121, 108212. [Google Scholar] [CrossRef]
  16. Fan, D.P.; Zhai, Y.; Borji, A.; Yang, J.; Shao, L. BBS-Net: RGB-D salient object detection with a bifurcated backbone strategy network. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer International Publishing: Cham, Switzerland, 2020; pp. 275–292. [Google Scholar]
  17. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef]
  18. Achanta, R.; Hemami, S.; Estrada, F.; Susstrunk, S. Frequency-tuned salient region detection. In Proceedings of the 2009 IEEE Conference on Computer vision And Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1597–1604. [Google Scholar]
  19. Tong, N.; Lu, H.; Zhang, Y.; Ruan, X. Salient object detection via global and local cues. Pattern Recognit. 2015, 48, 3258–3267. [Google Scholar] [CrossRef]
  20. Chen, C.; Wei, J.; Peng, C.; Qin, H. Depth-quality-aware salient object detection. IEEE Trans. Image Process. 2021, 30, 2350–2363. [Google Scholar] [CrossRef] [PubMed]
  21. Cong, R.; Yang, N.; Li, C.; Fu, H.; Zhao, Y.; Huang, Q.; Kwong, S. Global-and-local collaborative learning for co-salient object detection. IEEE Trans. Cybern. 2022, 53, 1920–1931. [Google Scholar] [CrossRef] [PubMed]
  22. Hou, Q.; Cheng, M.M.; Hu, X.; Borji, A.; Tu, Z.; Torr, P. Deeply supervised salient object detection with short connections. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  23. Zhao, X.; Pang, Y.; Zhang, L.; Lu, H. Suppress and balance: A simple gated network for salient object detection. In Proceedings of the Computer Vision–ECCV, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  24. Zhang, P.; Wang, D.; Lu, H.; Wang, H.; Ruan, X. Amulet: Aggregating multi-level convolutional features for salient object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  25. Wu, Z.; Su, L.; Huang, Q. Cascaded partial decoder for fast and accurate salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 July 2019. [Google Scholar]
  26. Chen, S.; Tan, X.L.; Wang, B.; Hu, X.L. Reverse attention for salient object detection. In Proceedings of the European Conference on Computer Vision ECCV, Munich, Germany, 8–14 September 2018. [Google Scholar]
  27. Wang, W.; Zhao, S.Y.; Shen, J.B.; Hoi, S.C.H.; Borji, A. Salient object detection with pyramid attention and salient edges. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  28. Wang, X.; Liu, Z.; Liesaputra, V.; Huang, Z. Feature specific progressive improvement for salient object detection. Pattern Recognit. 2024, 147, 110085. [Google Scholar] [CrossRef]
  29. Lang, C.; Nguyen, T.V.; Katti, H.; Yadati, K.; Kankanhalli, M.; Yan, S. Depth matters: Influence of depth cues on visual saliency. In Proceedings of the Computer Vision–ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 101–115. [Google Scholar]
  30. Peng, H.; Li, B.; Xiong, W.; Hu, W.; Ji, R. RGBD salient object detection: A benchmark and algorithms. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Springer International Publishing: Berlin/Heidelberg, Germany, 2014; pp. 92–109. [Google Scholar]
  31. Zhang, Q.; Qin, Q.; Yang, Y.; Jiao, Q.; Han, J. Feature Calibrating and Fusing Network for RGB-D Salient Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2023, 1–15. [Google Scholar] [CrossRef]
  32. Ikeda, T.; Masaaki, I. RGB-D Salient Object Detection Using Saliency and Edge Reverse Attention. IEEE Access 2023, 11, 68818–68825. [Google Scholar] [CrossRef]
  33. Xu, K.; Guo, J. RGB-D salient object detection via convolutional capsule network based on feature extraction and integration. Sci. Rep. 2023, 13, 17652. [Google Scholar] [CrossRef]
  34. Cong, R.; Liu, H.; Zhang, C.; Zhang, W.; Zheng, F.; Song, R.; Kwong, S. Point-aware interaction and cnn-induced refinement network for RGB-D salient object detection. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023. [Google Scholar]
  35. Qu, L.; He, S.; Zhang, J.; Tang, Y. RGBD salient object detection via deep fusion. IEEE Trans. Image Process. 2017, 26, 2274–2285. [Google Scholar] [CrossRef]
  36. Yi, K.; Zhu, J.; Guo, F.; Xu, J. Cross-Stage Multi-Scale Interaction Network for RGB-D Salient Object Detection. IEEE Signal Process. Lett. 2022, 29, 2402–2406. [Google Scholar] [CrossRef]
  37. Liu, Z.; Wang, K.; Dong, H.; Wang, Y. A cross-modal edge-guided salient object detection for RGB-D image. Neurocomputing 2021, 454, 168–177. [Google Scholar] [CrossRef]
  38. Sun, F.; Hu, X.H.; Wu, J.Y.; Sun, J.; Wang, F.S. RGB-D Salient Object Detection Based on Cross-modal Interactive Fusion and Global Awareness. J. Softw. 2023, 1–15. [Google Scholar] [CrossRef]
  39. Peng, Y.; Feng, M.; Zheng, Z. RGB-D Salient Object Detection Method Based on Multi-modal Fusion and Contour Guidance. IEEE Access 2023, 11, 145217–145230. [Google Scholar] [CrossRef]
  40. Sun, F.; Ren, P.; Yin, B.; Wang, F.; Li, H. CATNet: A cascaded and aggregated transformer network for RGB-D salient object detection. IEEE Trans. Multimed. 2023, 26, 2249–2262. [Google Scholar] [CrossRef]
  41. Theckedath, D.; Sedamkar, R.R. Detecting affect states using VGG16, ResNet50 and SE-ResNet50 networks. SN Comput. Sci. 2020, 1, 79. [Google Scholar] [CrossRef]
  42. Li, H.; Wu, X.J.; Durrani, T. NestFuse: An infrared and visible image fusion architecture based on nest connection and spatial/channel attention models. IEEE Trans. Instrum. Meas. 2020, 69, 9645–9656. [Google Scholar] [CrossRef]
  43. Milletari, F.; Navab, N.; Ahmadi, S.A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  44. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 1–12. [Google Scholar]
  45. Ketkar, N.; Moolayil, J. Introduction to pytorch. In Deep Learning with Python: Learn Best Practices of Deep Learning Models with PyTorch; Apress: New York, NY, USA, 2021; pp. 27–91. [Google Scholar]
  46. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  47. Ju, R.; Ge, L.; Geng, W.; Ren, T.; Wu, G. Depth saliency based on anisotropic center-surround difference. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 1115–1119. [Google Scholar]
  48. Niu, Y.; Geng, Y.; Li, X.; Liu, F. Leveraging stereopsis for saliency analysis. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 454–461. [Google Scholar]
  49. Zhu, C.; Li, G. A three-pathway psychobiological framework of salient object detection using stereoscopic technology. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 3008–3014. [Google Scholar]
  50. Fan, D.P.; Lin, Z.; Zhang, Z.; Zhu, M.; Cheng, M.M. Rethinking RGB-D salient object detection: Models, data sets, and large-scale benchmarks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 2075–2089. [Google Scholar] [CrossRef]
  51. Cheng, Y.; Fu, H.; Wei, X.; Xiao, J.; Cao, X. Depth enhanced saliency detection method. In Proceedings of the International Conference on Internet Multimedia Computing and Service, Xiamen, China, 10–12 July 2014; pp. 23–27. [Google Scholar]
  52. Fan, D.P.; Gong, C.; Cao, Y.; Ren, B.; Cheng, M.M.; Borji, A. Enhanced-alignment measure for binary foreground map evaluation. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 16 July 2018; pp. 698–704. [Google Scholar]
  53. Chen, T.; Xiao, J.; Hu, X.; Zhang, G.; Wang, S. Adaptive fusion network for RGB-D salient object detection. Neurocomputing 2023, 522, 152–164. [Google Scholar] [CrossRef]
  54. Bi, H.; Wu, R.; Liu, Z.; Zhu, H.; Zhang, C.; Xiang, T.Z. Cross-modal hierarchical interaction network for RGB-D salient object detection. Pattern Recognit. 2023, 136, 109194. [Google Scholar] [CrossRef]
  55. Zhang, M.; Yao, S.; Hu, B.; Piao, Y. C2DFNet: Criss-Cross Dynamic Filter Network for RGB-D Salient Object Detection. IEEE Trans. Multimed. 2022, 25, 5142–5154. [Google Scholar] [CrossRef]
  56. Wang, F.; Pan, J.; Xu, S.; Tang, J. Learning discriminative cross-modality features for RGB-D saliency detection. IEEE Trans. Image Process. 2022, 31, 1285–1297. [Google Scholar] [CrossRef]
  57. Chen, T.; Hu, X.; Xiao, J.; Zhang, G.; Wang, S. CFIDNet: Cascaded feature interaction decoder for RGB-D salient object detection. Neural Comput. Appl. 2022, 34, 7547–7563. [Google Scholar] [CrossRef]
  58. Cong, R.; Lin, Q.; Zhang, C.; Li, C.; Cao, X.; Huang, Q.; Zhao, Y. CIR-Net: Cross-modality interaction and refinement for RGB-D salient object detection. IEEE Trans. Image Process. 2022, 31, 6800–6815. [Google Scholar] [CrossRef] [PubMed]
  59. Ji, W.; Li, J.; Yu, S.; Zhang, M.; Piao, Y.; Yao, S.; Bi, Q.; Ma, K.; Zheng, Y.; Lu, H.; et al. Calibrated RGB-D salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 9471–9481. [Google Scholar]
  60. Zhao, J.; Zhao, Y.; Li, J.; Chen, X. Is depth really necessary for salient object detection? In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 1745–1754. [Google Scholar]
  61. Li, G.; Liu, Z.; Ling, H. ICNet: Information conversion network for RGB-D based salient object detection. IEEE Trans. Image Process. 2020, 29, 4873–4884. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The overall network architecture of SLMSF-Net.
Figure 1. The overall network architecture of SLMSF-Net.
Sensors 24 01117 g001
Figure 2. Depth attention module.
Figure 2. Depth attention module.
Sensors 24 01117 g002
Figure 3. Semantic localization module.
Figure 3. Semantic localization module.
Sensors 24 01117 g003
Figure 4. Pyramid feature fusion operation.
Figure 4. Pyramid feature fusion operation.
Sensors 24 01117 g004
Figure 5. Multi-Scale Fusion Module.
Figure 5. Multi-Scale Fusion Module.
Sensors 24 01117 g005
Figure 6. Training and validation loss curve.
Figure 6. Training and validation loss curve.
Sensors 24 01117 g006
Figure 7. Comparison of precision-recall (P-R) curves for different methods across six RGB-D datasets. Our SLMSF-Net method is represented by a solid red line.
Figure 7. Comparison of precision-recall (P-R) curves for different methods across six RGB-D datasets. Our SLMSF-Net method is represented by a solid red line.
Sensors 24 01117 g007aSensors 24 01117 g007b
Figure 8. Visual comparison between SLMSF-Net and state-of-the-art RGB-D models.
Figure 8. Visual comparison between SLMSF-Net and state-of-the-art RGB-D models.
Sensors 24 01117 g008
Table 1. Comparison of results for four evaluation metrics—mean absolute error (MAE), maximum F-measure (maxF), maximum E-measure (maxE), and S-measure (S)—across six datasets. The symbol “↑” indicates that a higher value is better for the metric, while “↓” indicates that a lower value is better. The best performance in each row is highlighted in bold.
Table 1. Comparison of results for four evaluation metrics—mean absolute error (MAE), maximum F-measure (maxF), maximum E-measure (maxE), and S-measure (S)—across six datasets. The symbol “↑” indicates that a higher value is better for the metric, while “↓” indicates that a lower value is better. The best performance in each row is highlighted in bold.
DatasetsEvaluation MetricsDeep Learning-Based RGB-D Saliency Detection Methods
DASNet
ICMM2020
D3Net
TNNLS
2020
ICNet
TIP2020
DCF
CVPR2021
CIRNet
TIP2022
CFIDNet
NCA2022
DCMF
TIP2022
C2DFNet
TMM2022
HINet
PR2023
AFNet
NC2023
Ours
NJU2K
[47]
MAE↓0.04180.04670.05190.03570.03500.03780.03570.03870.03850.03170.0315
maxF↑0.90150.89930.89050.91470.92810.91480.92520.90890.91380.92820.9352
maxE↑0.93930.93810.92640.95040.95470.94640.95820.94250.94470.95780.9615
S↑0.90250.90.89410.91160.92520.91420.92470.90820.91530.92620.9306
NLPR
[30]
MAE↓0.02120.02980.02810.02170.02800.02560.02900.02170.02570.02010.0199
maxF↑0.92180.89680.90790.91180.90710.90540.90570.91660.90620.92490.9298
maxE↑0.96410.95290.95240.96280.95540.95530.95410.96050.95650.96840.9693
S↑0.92940.91180.92270.92390.92080.92190.92200.92790.92230.93620.9388
DES
[51]
MAE↓0.02460.03140.02660.02410.02870.02330.02320.01990.02150.02210.0176
maxF↑0.90250.88420.91320.89350.89170.91080.92390.91590.92200.92250.9307
maxE↑0.93900.94510.95980.95140.94070.93960.96790.95900.96700.95290.9739
S↑0.90470.89730.92010.90490.90670.91690.93240.92170.92740.92520.9403
SIP
[50]
MAE↓0.05080.06320.06950.05180.06850.06010.06230.05290.06560.04340.0422
maxF↑0.88640.8610.85710.88440.86620.86990.87190.87700.85500.90890.9114
maxE↑0.92470.90850.90330.92170.90470.90880.91110.91600.89930.93890.9408
S↑0.87670.86030.85380.87560.86150.86380.87000.87150.85610.89590.9045
SSD
[49]
MAE↓0.04230.05850.06370.04980.05230.05040.07310.04780.04880.03830.0321
maxF↑0.87250.8340.84140.85090.85470.87070.81080.85980.85240.88480.9007
maxE↑0.92980.91050.90250.90900.91190.92610.89700.91710.91600.94270.9565
S↑0.88460.85660.84840.86440.87250.87910.83820.87180.86520.89680.9062
STERE
[48]
MAE↓0.03680.04620.04460.03890.04570.04260.04330.03850.04900.03360.0331
maxF↑0.90430.89110.89780.90090.89660.89710.90610.89730.88280.91770.9195
maxE↑0.94360.93820.94150.94470.93880.94200.94630.94290.93250.95720.9584
S↑0.91040.89850.90250.90220.90130.90120.90970.90230.89190.91840.9201
Table 2. Comparison of ablation study results. The symbol “↑” indicates that a higher value is better for the metric, while “↓” indicates that a lower value is better. The best performance in each row is highlighted in bold.
Table 2. Comparison of ablation study results. The symbol “↑” indicates that a higher value is better for the metric, while “↓” indicates that a lower value is better. The best performance in each row is highlighted in bold.
DatasetsEvaluation MetricsWithout DAMWithout SLMWithout MSFMSLMSF-Net
NJU2K
[47]
MAE↓0.03620.03520.03930.0315
maxF↑0.92140.92150.91650.9352
maxE↑0.95160.95080.94780.9615
S↑0.9210.92250.91850.9306
NLPR
[30]
MAE↓0.02350.02340.02840.0199
maxF↑0.92260.91930.9110.9298
maxE↑0.96270.9640.95930.9693
S↑0.93280.93270.92380.9388
DES
[51]
MAE↓0.01910.02280.02310.0176
maxF↑0.92840.91740.92630.9307
maxE↑0.97040.96180.96870.9739
S↑0.93420.92600.93170.9403
SIP
[50]
MAE↓0.05690.05280.06000.0422
maxF↑0.88270.89160.87480.9114
maxE↑0.91540.92020.91190.9408
S↑0.87760.88300.87390.9045
SSD
[49]
MAE↓0.05370.05340.05480.0321
maxF↑0.83780.83810.83950.9007
maxE↑0.90930.90420.90450.9565
S↑0.86610.8650.86580.9062
STERE
[48]
MAE↓0.04430.03760.05080.0331
maxF↑0.89190.91000.89060.9195
maxE↑0.93810.94790.93300.9584
S↑0.90140.91430.89860.9201
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peng, Y.; Zhai, Z.; Feng, M. SLMSF-Net: A Semantic Localization and Multi-Scale Fusion Network for RGB-D Salient Object Detection. Sensors 2024, 24, 1117. https://doi.org/10.3390/s24041117

AMA Style

Peng Y, Zhai Z, Feng M. SLMSF-Net: A Semantic Localization and Multi-Scale Fusion Network for RGB-D Salient Object Detection. Sensors. 2024; 24(4):1117. https://doi.org/10.3390/s24041117

Chicago/Turabian Style

Peng, Yanbin, Zhinian Zhai, and Mingkun Feng. 2024. "SLMSF-Net: A Semantic Localization and Multi-Scale Fusion Network for RGB-D Salient Object Detection" Sensors 24, no. 4: 1117. https://doi.org/10.3390/s24041117

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop