Next Article in Journal
TactiGraph: An Asynchronous Graph Neural Network for Contact Angle Prediction Using Neuromorphic Vision-Based Tactile Sensing
Next Article in Special Issue
A Novel Approach for Apple Freshness Prediction Based on Gas Sensor Array and Optimized Neural Network
Previous Article in Journal
Sollerman Hand Function Sub-Test “Write with a Pen”: A Computer-Vision-Based Approach in Rehabilitation Assessment
Previous Article in Special Issue
Traversable Region Detection and Tracking for a Sparse 3D Laser Scanner for Off-Road Environments Using Range Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CoSOV1Net: A Cone- and Spatial-Opponent Primary Visual Cortex-Inspired Neural Network for Lightweight Salient Object Detection

by
Didier Ndayikengurukiye
* and
Max Mignotte
Département d’Informatique et de Recherche Opérationnelle, Université de Montréal, Montreal, QC H3C 3J7, Canada
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(14), 6450; https://doi.org/10.3390/s23146450
Submission received: 12 June 2023 / Revised: 12 July 2023 / Accepted: 14 July 2023 / Published: 17 July 2023
(This article belongs to the Special Issue Object Detection Based on Vision Sensors and Neural Network)

Abstract

:
Salient object-detection models attempt to mimic the human visual system’s ability to select relevant objects in images. To this end, the development of deep neural networks on high-end computers has recently achieved high performance. However, developing deep neural network models with the same performance for resource-limited vision sensors or mobile devices remains a challenge. In this work, we propose CoSOV1net, a novel lightweight salient object-detection neural network model, inspired by the cone- and spatial-opponent processes of the primary visual cortex (V1), which inextricably link color and shape in human color perception. Our proposed model is trained from scratch, without using backbones from image classification or other tasks. Experiments on the most widely used and challenging datasets for salient object detection show that CoSOV1Net achieves competitive performance (i.e., F β = 0.931 on the ECSSD dataset) with state-of-the-art salient object-detection models while having a low number of parameters ( 1.14 M), low FLOPS ( 1.4 G) and high FPS ( 211.2 ) on GPU (Nvidia GeForce RTX 3090 Ti) compared to the state of the art in lightweight or nonlightweight salient object-detection tasks. Thus, CoSOV1net has turned out to be a lightweight salient object-detection model that can be adapted to mobile environments and resource-constrained devices.

1. Introduction

The human visual system (HVS) has the ability to select and process relevant information from among the large amount that is received. This relevant information in an image is called salient objects [1]. Salient object-detection models in computer vision try to mimic this phenomenon by detecting and segmenting salient objects in images. This is an important task, given its many applications in computer vision, such as object tracking, recognition and detection [2], advertisement optimization [3], image/video compression [4], image correction [5], analysis of iconographic illustrations [6], image retrieval [7], aesthetic evaluation [8], image quality evaluation [9], image retargeting [10], image editing [11] and image collages [12], to name a few. Thus, it has been the subject of intensive research in recent years and is still being investigated [13]. Salient object-detection models generally fall into two categories, namely conventional and deep learning-based models, which differ by their feature extraction process. The former use hand-crafted features, while the latter use features learned from a neural network. Thanks to powerful representation learning methods, deep learning-based salient object-detection models have recently shown superior performance over conventional models [13,14]. The high performance of these models is undeniable; however, generally, they are also heavy if we consider their number of parameters and the amount of memory occupied, in addition to their high computational cost and slow detection speed. This makes these models less practical for resource-limited vision sensors or mobile devices that have many constraints on their memory and computational capabilities, as well as for real-time applications [15,16]. Hence, there is a need for lightweight salient object-detection models whose performance is comparable to state-of-the-art models, with the advantages of being deployed on resource-limited vision sensors or mobile devices and having a detection speed that allows them to be used in real-time applications. Existing lightweight salient object-detection models have used different methodologies, such as backbones from nonlightweight classification models [17,18], the imitation of primate hierarchical visual perception [19], human attention mechanisms [16,19], etc.
In this work, we propose an original approach for a new lightweight neural network model, namely CoSOV1Net, for salient object detection, that can therefore be adapted to mobile environments and resource-limited or -constrained devices, with the additional properties of being able to be trained from scratch without having to use backbones developed from image-classification tasks and having few parameters, but with comparable performance with state-of-the-art models.
Given that detecting salient objects is a capability of the human visual system and that a normal human visual system performs this quickly and correctly, we used images or scenes encoding mechanism research advances in neuroscience, especially for the early stage of the human visual system [20,21,22]. Our strategy in this model is therefore inspired by two neuroscience discoveries in human color perception, namely:
1.
The color-opponent encoding in the early stage of the HVS (human visual system) [23,24,25,26];
2.
The fact that color and pattern are linked inextricably in human color perception [20,27].
Inspired by these neuroscience discoveries, we propose a cone- and spatial-opponent primary visual cortex (CoSOV1) module that extracts features at the spatial level and between color channels at the same time to integrate color in the patterns. This process is applied first on opposing color pair channels two by two and then to grouped feature maps through our deep neural network. Thus, based on the CoSOV1 module, we build a novel lightweight encoder–decoder deep neural network for salient object detection: CoSOV1Net, which has only 1.14 M parameters but comparable performance with state-of-the-art salient object-detection models. CoSOV1Net predicts salient maps at a speed of 4.4 FPS on an Intel CPU, i7-11700F and 211.2 FPS on a Nvidia GeForce RTX 3090 Ti GPU for 384 × 384 images and it has a low FLOPS = 1.4 G. Therefore, CoSOV1net is a lightweight salient object-detection model that can be adapted for mobile environments and limited-resource devices.
Our contribution is threefold:
  • We propose a novel approach to extract features from opposing color pairs in a neural network to exploit the strength of the color-opponent principle from human color perception. This approach permits the acceleration of neural network learning;
  • We propose a novel strategy to integrate color in patterns in a neural network by extracting features locally and between color channels at the same time in successively grouped feature maps, which results in a reduction in the number of parameters and the depth of the neural network, while keeping good performance;
  • We propose—for the first time, to our knowledge—a novel lightweight salient object-detection neural network architecture based on the proposed approach for learning opposing color pairs along with the strategy of integrating color in patterns. This model has few parameters, but its performance is comparable to state-of-the-art methods.
The rest of this work is organized as follows: Section 2 presents some lightweight models related to this approach; Section 3 presents our proposed lightweight salient object-detection model; Section 4 describes the datasets used, evaluation metrics, our experimental results and the comparison of our model with state-of-the-art models; Section 5 discusses our results; Section 6 concludes this work.

2. Related Work

Many salient object-detection models have been proposed and most of the influential advances in image-based salient object detection have been reviewed by Gupta et al. [13]. Herein, we present some conventional models and lightweight neural network models related to this approach.

2.1. Lightweight Salient Object Detection

In recent years, lightweight salient object-detection models have been proposed with different strategies and architectures. Qin et al. [28] designed U 2 n e t , a lightweight salient object-detection model with a two-level nested Unet [29] neural network able to capture more contextual information from different scales, thanks to the mixture of receptive fields of different sizes. Its advantages are threefold: first, it increases the depth of the whole architecture without increasing the computational cost; second, it is trained from scratch without using pretrained backbones, thus being able to keep feature maps high-resolution; third, it has high accuracy. Its disadvantage is its number of parameters. Other models are based on streamlined architecture to build lightweight deep neural networks. MobileNets [30,31] and ShuffleNets [32,33], along with their variants, are among the latter models. MobileNets [30] uses architecture based on depthwise separable convolution. ShuffleNets [32] uses architecture based on pointwise group convolution and channel shuffle, as well as depthwise convolution, to greatly reduce computational cost while maintaining accuracy. Their advantages are their computational cost, accuracy and speed, while their disadvantages are their number of parameters and their input resolution. Other authors have been inspired by primate or human visual system processes. Thus, Liu et al. [19] designed HVPNet, a lightweight salient object-detection network based on a hierarchical visual perception (HVP) module that mimics the primate visual cortex for hierarchical perception learning, whereas Liu et al. [16] were inspired by human perception attention mechanisms in designing SAMNet, another lightweight salient object-detection network, based on a stereoscopically attentive multiscale (SAM) module that adopts a stereoscopic attention mechanism for effective and efficient multiscale learning. Their advantages are their computational cost and accuracy, while their disadvantages are their number of parameters and their input resolution.

2.2. Color-Opponent Models

Color opponency, which is a human color perception propriety, has inspired many authors who have defined channels or feature maps to tackle their image-processing tasks. Frintrop et al. [34] used three opponent channels— R G , B Y and I—to extract features for their salient object-detection model.
To extract features for salient object detection, Ndayikengurukiye and Mignotte [1] used nine (9) opponent channels for RGB color space (RR: red–red; RG: red–green; RB: red–blue; GR: green–red; GG: green–green; GB: green–blue; BR: blue–red; BG: blue–green; BB: blue–blue) with a nonlinear combination, thanks to the OCLTP (opponent color local ternary pattern) texture descriptor, which is an extension of the OCLBP (opponent color local binary pattern) [35,36] and Fastmap [37], which is a fast version of MDS (multidimensional scaling).
Most authors apply the opponent color mechanism to the input image color space channels and not on the resulting feature maps. However, Jain and Healey [38] used opponent features computed from Gabor filter outputs. They computed opponent features by combining information across different spectral bands at different scales obtained via Gabor filters for color texture recognition [38]. Yang et al. [39] proposed a framework based on the color-opponent mechanisms of color-sensitive double-opponent (DO) cells in the human visual system’s primary visual cortex (V1) in order to combine brightness and color to maximize the boundary-detection reliability in natural scenes. The advantages of hand-crafted models are their computational cost, number of parameters, speed and input resolution, while their disadvantage is accuracy.
In this work, we propose a model inspired by the human visual system but different from other models, because our model uses the primary visual cortex (V1) cone- and spatial-opponent principle to extract features at channels’ spatial levels and between color channels at the same time to integrate color into patterns in a manner allowing for a lightweight deep neural network design with performance comparable with state-of-the-art lightweight salient object-detection models.

3. Materials and Methods

3.1. Introduction

Our model for tackling the challenge of lightweight salient object detection is inspired by the human visual system (HVS)’s early visual color process, especially its cone opponency and spatial opponency in the primary visual cortex (V1). The human retina (located in the inner surface of the eye) has two types of photoreceptors, namely rods and cones. Rods are responsible for monochromatic vision under low levels of illumination, while cones are responsible for color vision at normal levels of illumination. There are three classes of cones: L, M and S. When light is absorbed by cone photoreceptors, the L, M and S cones absorb long-, middle- and short-wavelength visible light, respectively [24,25,27].
The cone signals are then processed by single-opponent retina ganglion cells. The single opponent operates an antagonistic comparison of the cone signals [23,25,26,40]:
  • L − M opponent for red–green;
  • S − (L + M) opponent for blue–yellow.
The red–green and blue–yellow signals are carried by specific cells (different cells each for red–green and blue–yellow) through the lateral geniculate nucleus (LGN) to the primary visual cortex (V1).
Shapley [27] and Shapley and Hawken [20] showed that the primary visual cortex (V1) plays an important role in color perception through the combined activity of two kinds of color-sensitive cortical neurons, namely single-opponent and double-opponent cells. Single-opponent cells in V1 operate in the same manner as those of retina ganglion cells and provide neuronal signals that can be used for estimating the color of the illumination [27]. Double-opponent cells in V1 compare cone signals across space as well as between cones [21,22,24,27]. Double-opponent cells thus have two opponencies: spatial opponency and cone opponency. These properties permit them to be sensitive to color edges and spatial patterns. They are thus able to inextricably link color and pattern in human color perception [20,27].
As the primary visual cortex (V1) is known to play a major role in visual color perception, as highlighted above, in this work, we propose a deep neural network based on the primary visual cortex (V1) to tackle the challenge of lightweight salient object detection. In particular, we use two neuroscience discoveries in human color perception, namely:
1.
The color-opponent encoding in the early stage of the HVS;
2.
The fact that color and pattern are inextricably linked in human color perception
These two discoveries in neuroscience inspired us to design a neural network architecture for lightweight salient object detection, which hinges on two main ideas. First, at the beginning of the neural network, our model opposes color channels two by two by grouping them (R-R, R-G, R-B, G-G, G-B, B-B) then extracting the features at the channels’ spatial levels and between the color channels from each channel pair at the same time, to integrate color into patterns. Therefore, instead of performing a subtractive comparison or an OCLTP (opponent color linear ternary pattern) like Ndayikengurukiye and Mignotte [1], we let the neural network learn the features that represent the comparison of the two color pairs. Second, this idea of grouping and then extracting the features at the channels’ spatial levels and between the color channels at the same time is applied on feature maps at each neural network level until the saliency maps are obtained. This process allows the proposed model to mimic the human visual system’s capability of inextricably linking color and pattern in color perception [20,27].
It is this idea that differentiates our model from other models that use depthwise convolution followed by pointwise convolution [30,31] to extract features at each individual color channel level (or feature map) first, not through a group of color channels (or feature maps) at the same time, as our model does. This idea also differentiates our model from models that combine a group of color channels (or feature maps) pixel by pixel first and apply depthwise convolution afterwards [32,33]. The idea of grouping color channels in pairs (or feature map groups) differentiates our model from models that consider all color channels (or feature maps) as a single group while extracting features at color channels’ spatial levels and between color channels at the same time.
Our model takes into account nonlinearities in the image at the beginning as well as through our neural network. For this purpose, we use an encoder–decoder neural network type whose core is a module that we call CoSOV1 (cone- and spatial-opponent primary visual cortex).

3.2. CoSOV1: Cone- and Spatial-Opponent Primary Visual Cortex Module

The CoSOV1 (cone- and spatial-opponent primary visual cortex) module is composed of two parts (see Figure 1).
In the first part, input color channels (or input feature maps) are split into groups of equal depth. Convolution ( 3 × 3 ) operations are then applied to each group of channels (or feature maps) in order to extract features from each group as opposing color channels (or opposing feature maps). This is performed thanks to a set of filters that convolve the group of color channels (or feature maps). Each filter is applied to the color channels (or input feature maps) through a convolution operation that detects local features at all locations on the input. Let I g R W × H × S be an input group of feature maps, where W and H are the width and the height of each group’s feature map, respectively, and W R 3 × 3 × S , a filter with learned weights, with S being the depth of each group or the number of the channels in each group g, with g { 1 , , G } (where G is the number of groups). The output feature map O g R W × H for this group g has a pixel value in the ( k , l ) position, defined as follows:
O k , l g = s = 1 S i = 0 2 j = 0 2 W i , j , s I k + i 1 , l + j 1 , s g
The weight matrix W R 3 × 3 × S is the same across the whole group of channels or feature maps. Therefore, each resulting output feature map represents a particular feature at all locations in the input color channels (or input feature maps) [41]. We call the 3 × 3 convolution on grouped channels (or grouped feature maps) groupwise convolution. The zero padding is applied during the convolution process to keep the input channel size for the output feature maps. After groupwise convolution, we apply the batch normalization transform, which is known to enable faster and more stable training of deep neural networks [42,43]. Let B = { X 1 , , X K } be a minibatch that contains K examples from a dataset. The minibatch mean is
μ B = 1 K k = 1 K X k
and the minibatch variance is
σ B 2 = 1 K k = 1 K ( X k μ B ) 2
The batch normalization transform B N γ , β : { X 1 , , X K } { Y 1 , , Y K } ( γ and β are parameters to be learned):
Y k = γ X k ^ + β
where k { 1 , , K } and
X k ^ = X k μ B σ B 2 + ϵ
and ϵ is a very small constant to avoid division by zero.
In order to take into account the nonlinearities present in the color channel input (or feature map input), given that groupwise convolution is a linear transformation, batch normalization is followed by a nonlinear function, exponential linear unit (ELU), defined as follows:
ELU ( x ) = x if x 0 , α × ( exp ( x ) 1 ) otherwise
where α = 1 by default.
The nonlinear function, which is the activation function, is placed after batch normalization, as recommended by Chollet [44].
The second part of the module searches for the best representation of the obtained feature maps. It is similar to the first part of the module, except for the groupwise convolution, which is replaced by point-wise convolution, but the input feature maps for pointwise convolution in this model are not grouped. Pointwise convolution allows us to learn the filters’ weights and thus obtain feature maps that best represent the input channels (or input feature maps) for the salient object-detection task, while having few parameters.
Let O R W × H × M be the output of the first part of the module, with M being the number of feature maps in this output and W and H being the width and the height, respectively. Let a filter of the learned weights V R M and FM R W × H be its output feature map by pointwise convolution. Its pixel value FM k , l in ( k , l ) position is:
FM k , l = m = 1 M V m O k , l , m
Thus, V R M is a vector of learned weights that associates the input feature maps O R W × H × M to the feature map FM R W × H , which is the best representation of the latter-mentioned input feature maps. The pointwise convolution in this module uses many filters and thus it outputs many feature maps that are the best representation of the input feature map O . As pointwise convolution is a linear combination, we again apply batch normalization followed by a exponential linear unit function (ELU) on the feature map FM to obtain the best representation of the input feature maps for the learned weights V R M , which takes into account nonlinearities in the feature maps O R W × H × M .
Our scheme is different from depthwise separable convolution in that it uses the convolution of a group of channels instead of each channel individually [30,45].
In addition, after the nonlinear function, noise is injected in the resulting feature maps during the neural network learning stage thanks to the dropout process (but not in the prediction stage) to facilitate the learning process. In this model, we use DropBlock [46] if the width of the feature map is greater than 5; otherwise, we use the common dropout [47].
The CoSOV1 module allows our neural network to have few parameters but good performance.

3.3. CoSOV1Net Neural Network Model Architecture

Our proposed model is built on the CoSOV1 module (see Figure 1). It is a neural network of the U-net encoder–decoder type [29] and is illustrated in Figure 2. Thus, our model consists of three main blocks:
1.
The input RGB color channel pairing;
2.
The encoder;
3.
The decoder.

3.3.1. Input RGB Color Channel Pairing

At this stage, through Pairing_Color_Unit, the input RGB image is paired in six opposing color channel pairs: R-R, R-G, R-B, G-G, G-B and B-B [1,35,48]. These pairs are then concatenated, which gives 12 channels, R, R, R, G, R, B, G, G, G, B, B, B, as illustrated in Figure 3. This is the step for choosing the color channels to oppose. The set of concatenated color channels is then fed to the encoder.

3.3.2. Encoder

The encoder, in our proposed neural network model, is a convolutional neural network (CNN) [49] where an encoder unit (see Figure 2) is repeated eight times. Each encoder unit is followed by a max pooling ( 2 × 2 ) with strides = 2, except for the eighth neural network level, where the max pooling is 3 × 3 with strides = 3 (the max pooling is a downsampling operation, like a filtering with a maximum filter). While the size of each feature map is reduced by half, the depth of the feature maps is doubled, except for the first level, where it is kept at 12 and the last two levels, where it is kept at 128 to have few parameters.
The encoder unit (see Figure 4a) is composed of a residual block (Figure 4b) repeated three (3) times.
We used the residual block because this kind of block is known to improve the training of deeper neural networks [50]. The residual block consists of two CoSOV1 modules with a residual link. The reason for all these repetitions is to encode more information and thus allow our network performance to increase.
In the encoder, schematically, as explained above (Section 3.2), the CoSOV1 module (Figure 4c) splits the input channels into groups and applies groupwise convolution ( 3 × 3 convolution). Then, pointwise convolution is applied to the outputs of the concatenated groups (see Figure 5 for the first-level input illustration). Each of these convolutions is followed by batch normalization and a nonlinear function (ELU: exponential linear unit activation). After these layers, during the model training, regularization is performed in the CoSOV1 module using the dropout [47] method for small feature maps (dimensions smaller than 5 × 5 ) and DropBlock [46]—which is a variant of dropout that zeroes a block instead of pixels individually as dropout does—for feature maps with dimensions greater than 5 × 5 .
At its end, the encoder is followed by the middle unit (see Figure 6a), which is the CoSOV1 module (see Figure 6b), where we remove the groupwise convolution—since at this stage, the feature maps are 1 × 1 × 128 in size—and add a residual link.

3.3.3. Decoder

The decoder transforms the features from the encoder to obtain the estimate of the salient object(s) present in the input image. This transformation is achieved through a repeating block, namely the decoder unit (see Figure 7a). The decoder unit consists of two parts: the decoder residual block (see Figure 7b) and the decoder deconvolution block (see Figure 7c). The decoder residual block is a modified CoSOV1 module that allows the model to take into account the output of the corresponding level in the encoder. The output of the decoder residual block takes two directions. On the one hand, it is passed to the next level of the decoder; and on the other, to the second part of the decoder unit, which is the decoder deconvolution block. The latter deconvolves this output, obtaining two feature maps having the size of the input image ( 384 × 384 × 2 in our case). At the last level of the decoder, all the outputs from the deconvolution blocks are concatenated and fed to a convolution layer followed by a softmax activation layer, which gives the estimation of the salient object-detection map.

4. Experimental Results

4.1. Implementation Details

For our proposed model implementation, we used the deep learning platform TensorFlow with the Keras deep learning application programming interface (API) [51]. All input images were resized to 384 × 384 and pixel values were normalized (each pixel channel value [ 0.0 , , 1.0 ] and ground truth pixels { 0 , 1 } ). Experiments were conducted on a single GPU, Nvidia GeForce RTX 3090 Ti (24 GB) and an Intel CPU, i7-11700F.

4.2. Datasets

Our proposed model’s experiments were conducted on public datasets, which are the most widely used in the field of salient object detection [52]. Thus, we used the Extended Complex Scene Saliency dataset (ECSSD) [53] and the DUT-OMRON (Dalian University of Technology—OMRON Corporation) [54], DUTS [55], HKU-IS [56] and THUR15K [57] datasets.
ECSSD [53] contains 1000 natural images and their ground truths. Many of its images are semantically meaningful but structurally complex for saliency detection [53].
DUT-OMRON [54] contains 5168 images and their binary masks, with diverse variations and complex backgrounds.
The DUTS dataset [55] is divided into DUTS-TR (10,553 training images) and DUTS-TE (5019 test images). We trained and validated our proposed model on the DUTS-TR and DUTS-TE was used for tests.
HKU-IS [56] is composed of 4447 complex images, which contain many disconnected objects with different spatial distributions. Furthermore, it is very challenging for similar foreground/background appearances [58].
THUR15K is a dataset of images taken from the “Flickr” website, divided into five categories (butterfly, coffee mug, dog jump, giraffe, plane), which contains 3000 images. The images of this dataset represent real-world scenes and are considered complex for obtaining salient objects [57] (6232 images with ground truths).

4.3. Model Training Settings

For the reproducibility of the experiments, we set the seed = 123. We trained our proposed model on DUTS-TR (10,553 training images). We split the DUTS-TR dataset into a train set (9472 images) and a validation set (1056 images); that is, approximately 90 % of the dataset for the training set and 10 % for the validation set. We did not use 25 images because we wanted the training set and the validation set to be divisible by batch size, which is 32.
Our proposed model was trained on scratch without pretrained backbones from image classification (i.e., VGG [59], etc.) or lightweight backbones (i.e., MobileNets [30,31] or ShuffleNets [32,33]). As DUTS-TR is not a big dataset, we used data augmentation during training and many epochs in order to overcome this problem. Indeed, the more epochs, the more the data-augmentation process transforms data. Thus, our proposed model training has two successive stages:
  • The first stage is with data augmentation, which is applied to each batch with random transformation ( 40 % zoom in or horizontal flip or vertical flip). This stage has 480 epochs: 240 epochs with learning rate = 0.001 and 240 epochs with learning rate = 0.0001 ;
  • The second stage is without data augmentation. It has 620 epochs: 240 epochs with learning rate = 0.001 , followed by 140 epochs with learning rate = 0.0001 and 240 epochs with learning rate = 0.00005 .
We also used the same initializer for all layers in the neural network: the HeUniform Keras initializer [60], which draws samples from a uniform distribution within [−limit, limit], where limit = 6 f a n _ i n ( f a n _ i n is the number of input units in the weight tensor). The dropout rate was set to 0.2 . We used the RMSprop [61] Keras optimizer with default values except for the learning rate; the centered, which was set to true; and the clipnorm = 1. The loss function used was the “sparse_categorical_crossentropy” Keras function; the Keras metric was “SparseCategoricalAccuracy; the Keras check point monitor was “val_sparse_categorical_accuracy”.

4.4. Hyperparameters

Hyperparameters such as the ELU activation function, the optimizer, the batch size, the filter size and the learning rates were chosen experimentally by observing the results.
The other hyperparameters were chosen as follows:
  • Image size: The best image size was 384 × 384 . We did not choose a small size because we expected to have a small salient object. As we also wanted to have a low computational cost, we did not go beyond this size.
  • Number of levels for the encoder: We empirically obtained eight levels as the best number. The choice of image size permitted us to have a maximum of eight levels for the encoder part, given that 384 = 2 7 × 3 . The size of the feature maps of each level corresponds to the size of those of the previous level divided by 2, except the last level, where the division is by 3.
  • Number of levels for the decoder: Eight levels. The number of levels is the same for the encoder part and the decoder part.
  • Number of layers: At each level, we chose to use an encoder unit that has an equal number of layers for all levels and a decoder unit that has an equal number of layers for all levels. The number of layers was obtained experimentally.
  • Number of filters: We also experimentally chose the number of filters keeping in mind the minimum parameters; the encoder’s number of filters was 12, 16, 32, 64, 128, 128, 128 and 128, respectively, for the first, second, , seventh and eighth levels; the decoder residual bloc number of filters was 128, 128, 128, 128, 64, 32, 16 and 8, respectively, for the eighth, seventh, sixth, , second and first levels. For the decoder deconvolution blocs, at each level, the number of filters was 2.
  • The use of batch normalization: Batch normalization is known to enable faster and more stable training for deep neural networks [42,43]. So, we decided to use it.
  • Use of dropout: The dropout process injects noise in the resulting feature maps during the neural network learning stage (but not in the prediction stage) to facilitate the learning process. In this model, we used DropBlock [46] if the width of the feature map was greater than 5; otherwise, we used the common dropout [47]. The best results were obtained for DropBlock size = 5 × 5 and rate = 0.1 (the authors’ paper suggested a value between 0.05 and 0.25). For the common dropout, the best rate was 0.2, obtained experimentally.
As our proposed model, CoSOV1Net does not use pretrained backbones and the input image is resized to 384 × 384 ; it has the advantage of good resolution.

4.5. Evaluation Metrics

4.5.1. Accuracy

The metrics used to evaluate our proposed model accuracy were F β measure, M A E (mean absolute error) and weighted F β w measure [62]. We also used precision, precision–recall and F β measure curves.
Let M be the binary mask obtained for the predicted saliency probability map, given a threshold in the range of [ 0 , 1 ) and with G being the corresponding ground truth:
Precision = | M G | | M |
Recall = | M G | | G |
∩: set intersection symbol; |.|: the number of pixels whose values are not zeros.
The F β -measure ( F β ) is the weighted harmonic mean of precision and recall:
F β = ( 1 + β 2 ) × Precision × Recall β 2 × Precision + Recall
During evaluation, β 2 = 0.3 , as it is often suggested [16,58].
Let S ¯ be the saliency map estimation with pixel values normalized in order to be in [ 0.0 , , 1.0 ] and G ¯ ; its ground truth also normalized in { 0 ; 1 } . The M A E (mean absolute error) is:
M A E = 1 W × H x = 1 W y = 1 H | S ¯ ( x , y ) G ¯ ( x , y ) |
where W and H are the width and the height, respectively, of the above maps ( S ¯ and G ¯ ).
The F β w measure [62] fixes the interpolation flaw, dependence flaw and equal importance flaw in traditional evaluation metrics and its value is:
F β w = ( 1 + β 2 ) Precision w × Recall w β 2 × Precision w + Recall w
Precision w and Recall w are the weighted precision and the weighted recall, respectively.

4.5.2. Lightweight Measures

Since we propose a lightweight salient object-detection model in this work, we therefore also evaluate the model with lightweight measures: the number of parameters, the saliency map estimation speed (FPS: frames per second) and the computational cost by measuring the FLOPS (the number of floating-point operations). The FLOPS is related to the device’s energy consumption (the higher the FLOPS, the higher the energy consumption). The floating-point operation numbers are computed as follows [63]:
  • For a convolution layer with n filters of size k × k applied to W × H × C feature maps (W: width; H: height; C: channels), with P: number of parameters:
    FLOPS = W × H × P
  • For a max-pooling layer or an upsampling layer with a window of size s z × s z on W × H × C feature maps (W: width; H: height; C: channels):
    FLOPS = W × H × C × s z × s z

4.6. Comparison with State of the Art

We compare our proposed model with 20 state-of-the-art salient object detection and 10 state-of-the-art lightweight salient object-detection models. We divided these methods because the lightweight methods outperform others with respect to lightweight measures. However, the lightweight methods’ accuracy is lower than the accuracy of those with huge parameters. We mainly used the salient object-detection results provided by Liu et al. [16], except for the F β measure and precision–recall curves, where we used saliency maps provided by these authors. We also used saliency maps provided by the HVPNet authors [19] to compute HVPNet F β ω measures.
In this section, we describe the comparison with the 20 salient object-detection models, namely DRFI [64], DCL [65], DHSNet [66], RFCN [67], NLDF [68], DSS [69], Amulet [18], UCF [70], SRM [71], PiCANet [17], BRN [72], C2S [73], RAS [74], DNA [75], CPD [76], BASNet [77], AFNet [78], PoolNet [79], EGNet [80] and BANet [81].
Table 1 shows that our proposed model CoSOV1Net outperforms all 20 state-of-the-art salient object-detection models for lightweight measures (#parameters, FLOPS and FPS) by a large margin (i.e., the best among them for FLOPS is DHSNet [66], with FLOPS = 15.8 G and F β = 0.903 for ECSSD; the worst is EGNet [80], with FLOPS = 270.8 G and F β = 0.938 for ECSSD; meanwhile, our proposed model, CoSOV1Net, has FLOPS = 1.4 G, and its F β = 0.931 for ECSSD) (see Table 1).
Table 1 also shows that CoSOV1Net is among the top 6 models for ECSSD, among the top 7 for DUT-OMRON and around the top 10 for the other three datasets for the F-measure. Table 2 and Table 3 compare our model with the state-of-the-art models for the MAE and F β ω measures, respectively. From this comparison, we see that our model is ranked around the top 10 for all four datasets and is ranked 15th for the HKU-IS dataset. This demonstrates that our model is also competitive with respect to the performance of state-of-the-art models.
Table 1, Table 2 and Table 3 show that our proposed model, CoSOV1Net, clearly has the advantage of the number of parameters, computational cost and speed over salient object detection. They also show that its performance is closer to the best among them.
We also compared CoSOV1Net with the state-of-the-art lightweight salient object-detection models MobileNet [30], MobileNetV2 [31], ShuffleNet [32], ShuffleNetV2 [33], ICNet [82], BiSeNet R18 [83], BiSeNet X39 [83], DFANet [84], HVPNet [19] and SAMNet [16].
For the comparison with state-of-the-art lightweight models, Table 4 shows that our proposed model outperforms these state-of-the-art lightweight models in parameter numbers and the F β measure for the ECSSD dataset and is competitive for other measures and datasets. Table 5 shows that our model outperforms these state-of-the-art lightweight models for the MAE measure for the ECSSD and DUTS-TE datasets and is ranked first ex aequo with HVPNet for DUT-OMRON, first ex aequo with HVPNet and SAMNet for the HKU-IS dataset and second for the THUR15K dataset. Our model also outperforms these state-of-the-art lightweight models for the F β ω measure for ECSSD and DUTS-TE and is competitive for the three other datasets (see Table 6).
Table 4, Table 5 and Table 6 show that CoSOV1Net clearly has the advantage of the number of parameters over the lightweight salient object detection. They also show that its performance is closer to the best among them. Thus, CoSOV1Net has the advantage of performance.
Regarding computational cost, CoSOV1Net has an advantage over half of the state-of-the-art lightweight salient object-detection models. Overall, we can conclude that it has an advantage in terms of computational cost.

4.7. Comparison with SAMNet and HVPNet State of the Art

We chose to compare our CoSOV1Net model specifically with SAMNet [16] and HVPNet [19] because they are among the best state-of-the-art models.
Figure 8 shows that precision curves for ECSSD and HKU-IS datasets highlight that CoSOV1Net slightly dominates the SAMNet and HVPNet state-of-the-art lightweight salient object-detection models and that there is no clear domination for the DUT-OMRON, DUTS-TE and THUR15K precision curves between the three models. Therefore, the proposed model CoSOV1Net is competitive with these two state-of-the-art lightweight salient object-detection models with respect to precision.
Figure 9 shows that the three models’ precision–recall curves (for the five datasets used: ECSSD, DUT-OMRON, DUTS-TE, HKU-IS and THUR15K) are very close to each other. Therefore, the proposed model is competitive with these two state-of-the-art lightweight salient object-detection models with respect to precision–recall.
Figure 10 shows that the three models’ F β measure curves (for the five datasets used: ECSSD, DUT-OMRON, DUTS-TE, HKU-IS and THUR15K) are very close to each other. The CoSOV1Net model slightly dominates the two state-of-the-art lightweight salient object-detection models for thresholds 150 and the two state-of-the-art models slightly dominate for thresholds 150 . Thus, there is no clear dominance for one model among the three. This proves that our CoSOV1Net model is comparable to these state-of-the-art lightweight salient object-detection models while having the advantage of a low number of parameters compared to them.
For qualitative comparison, Figure 11 shows some images highlighting that our proposed model (CoSOV1Net) is competitive with regard to the state-of-the-art SAMNet [16] and HVPNet [19] models, which are among the best ones.
Images from rows 1 and 2 show a big salient object on a cloudy background and a big object on a complex background, respectively: CoSOV1Net (ours) performs better than HVPNet on these saliency maps. Row 3 shows salient objects with the same colors and row 4 shows salient objects with multiple colors: the SAMNet and CoSOV1Net saliency maps are slightly identical and the HVPNet saliency map is slightly better. Row 5 shows n image with three salient objects with different sizes and colors: two are big and one is very small; the CoSOV1Net saliency map is better than SAMNet’s and HVPNet’s. Row 6 shows red salient objects on a black and yellow background; SAMNet’s saliency map is the worst, while CoSOV1Net and HVPNet perform well on that image. Row 7 shows a complex background and multiple salient objects with different colors: CoSOV1Net performs better than SAMNet and HVPNet. Row 8 shows tiny salient objects: the three models perform well. On row 9, SAMNet has the worst performance, while CoSOV1Net is the best. Row 10 shows colored glasses as salient objects: the CoSOV1Net performance is better than SAMNet’s and HVPNet’s. On row 11, SAMNet has the worst performance. On row 12 and 13, CoSOV1Net has the best performance. Row 18 shows a submarine image: CoSOV1Net is better than SAMNet.
Figure 8, Figure 9, Figure 10 and Figure 11 confirm that CoSOV1Net has an advantage on performance.

5. Discussion

The results show the performance of our model, CoSOV1Net, for accuracy measures and lightweight measures. CoSOV1Net’s rank, when compared to state-of-the-art models, shows that it behaves as a lightweight salient object-detection model by dominating lightweight measures and having good performance for accuracy measures (see Table 7).
The results also show that when CoSOV1Net is compared to state-of-the-art lightweight salient object-detection models, its measure results are generally ranked among the best for the datasets and measures used (see Table 8). Thus, we can conclude that CoSOV1Net behaves as a competitive lightweight salient object-detection model.
As we did not use backbones from image classification (i.e., VGG [59], ) or lightweight backbones (i.e., MobileNets [30,31] or ShuffleNets [32,33]), we conclude that CoSOV1Net’s performance is intrinsic to this model itself.
Finally, putting together the measures for salient object-detection models and lightweight salient object-detection models in a graphic, we noticed that the CoSOV1Net model is located for F β measures with respect to FLOPS and for the number of parameters in the top left, while for the FPS measure, it is located in the top right, thus demonstrating its performance as a lightweight salient object-detection model (see Figure 12). This shows that CoSOV1Net is competitive with the best state-of-the-art models used.
The quantitative and the qualitative comparisons with SAMNet [16] and HVPNet [19] showed that our proposed model has good performance, given that these state-of-the-art models are among the best ones.

6. Conclusions

In this work, we present a lightweight salient object-detection deep neural network, CoSOV1Net, with a very low number of parameters (1.14 M), a low floating-point operations number (FLOPS = 1.4 G) and thus low computational cost and respectable speed ( FPS = 211.2 on GPU: Nvidia GeForce RTX 3090 Ti), yet with comparable performance with state-of-the-art salient object-detection models that use significantly more parameters, and other lightweight salient object-detection models such as SAMNet [16] and HVPNet [19].
The novelty of our proposed model (CoSOV1Net) is that it uses the principle of integrating color in pattern in a salient object-detection deep neural network, since according to Shapley [27] and Shapley and Hawken [20], color and pattern are inextricably linked in color human perception. This is implemented by taking inspiration from the primary visual cortex (V1) cells, especially cone- and spatial-opponent cells. Thus, our method extracts features at the color channels’ spatial level and between the color channels at the same time on a pair of opposing color channels. The idea of grouping color pushed us to group feature maps through the neural network and extract features at the spatial level and between feature maps, as carried out for color channels.
Our results showed that this strategy generates a model that is very promising, competitive with most state-of-the-art salient object-detection and lightweight salient object-detection models and practical for mobile environments and limited-resource devices.
In future work, our proposed CoSOV1Net model, based on integrating color into patterns, can be improved by coupling it with the human visual system attention mechanism, which is the basis of many lightweight models, to tackle its speed limitation and thus produce a more efficient lightweight salient object-detection model.

Author Contributions

Conceptualization, D.N. and M.M.; methodology, D.N.; software, D.N.; validation, D.N.; formal analysis, D.N.; investigation, D.N.; resources, D.N. and M.M.; data curation, D.N.; writing—original draft preparation, D.N.; writing—review and editing, D.N. and M.M.; visualization, D.N.; supervision, M.M.; project administration, M.M.; funding acquisition, M.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by individual discovery grant number RGPIN-2022-03654.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The ECSSD dataset is available at 09 June 2023. https://www.cse.cuhk.edu.hk/leojia/projects/hsaliency/dataset.html. The DUT-OMRON dataset is available at 09 June 2023. http://saliencydetection.net/dut-omron/. The DUTS-TR and DUTS-TE datasets are available at 09 June 2023 http://saliencydetection.net/duts/. The HKU-IS dataset is available at 09 June 2023. https://i.cs.hku.hk/~yzyu/research/deep_saliency.html. The THUR15K dataset is available at 09 June 2023. https://mmcheng.net/code-data/. The SAMNet [16] model’s datasets are available at 09 June 2023. https://github.com/yun-liu/FastSaliency/tree/master/SaliencyMaps/SAMNet. The HVPNet [19] model’s datasets are available at 09 June 2023. https://github.com/yun-liu/FastSaliency/tree/master/SaliencyMaps/HVPNet. Our data results (CoSOV1Net saliency maps) are available at 16 July 2023. http://www.iro.umontreal.ca/~mignotte/ResearchMaterial/COSOV1NET-Data/.

Acknowledgments

The authors would like to thank the NSERC (Natural Sciences and Engineering Research Council of Canada) for having supported this research work via the individual discovery grant program (RGPIN-2022-03654).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ndayikengurukiye, D.; Mignotte, M. Salient Object Detection by LTP Texture Characterization on Opposing Color Pairs under SLICO Superpixel Constraint. J. Imaging 2022, 8, 110. [Google Scholar] [CrossRef] [PubMed]
  2. Smeulders, A.W.; Chu, D.M.; Cucchiara, R.; Calderara, S.; Dehghan, A.; Shah, M. Visual tracking: An experimental survey. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 1442–1468. [Google Scholar]
  3. Pieters, R.; Wedel, M. Attention capture and transfer in advertising: Brand, pictorial and text-size effects. J. Mark. 2004, 68, 36–50. [Google Scholar] [CrossRef] [Green Version]
  4. Itti, L. Automatic foveation for video compression using a neurobiological model of visual attention. IEEE Trans. Image Process. 2004, 13, 1304–1318. [Google Scholar] [CrossRef] [Green Version]
  5. Li, J.; Feng, X.; Fan, H. Saliency-based image correction for colorblind patients. Comput. Vis. Media 2020, 6, 169–189. [Google Scholar] [CrossRef]
  6. Pinciroli Vago, N.O.; Milani, F.; Fraternali, P.; da Silva Torres, R. Comparing CAM algorithms for the identification of salient image features in iconography artwork analysis. J. Imaging 2021, 7, 106. [Google Scholar] [CrossRef]
  7. Gao, Y.; Shi, M.; Tao, D.; Xu, C. Database saliency for fast image retrieval. IEEE Trans. Multimed. 2015, 17, 359–369. [Google Scholar] [CrossRef]
  8. Wong, L.K.; Low, K.L. Saliency-enhanced image aesthetics class prediction. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 997–1000. [Google Scholar]
  9. Liu, H.; Heynderickx, I. Studying the added value of visual attention in objective image quality metrics based on eye movement data. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 3097–3100. [Google Scholar]
  10. Chen, L.Q.; Xie, X.; Fan, X.; Ma, W.Y.; Zhang, H.J.; Zhou, H.Q. A visual attention model for adapting images on small displays. Multimed. Syst. 2003, 9, 353–364. [Google Scholar] [CrossRef]
  11. Chen, T.; Cheng, M.M.; Tan, P.; Shamir, A.; Hu, S.M. Sketch2photo: Internet image montage. ACM Trans. Graph. (TOG) 2009, 28, 1–10. [Google Scholar]
  12. Huang, H.; Zhang, L.; Zhang, H.C. Arcimboldo-like collage using internet images. In Proceedings of the 2011 SIGGRAPH Asia Conference, Hong Kong, China, 12–15 December 2011; pp. 1–8. [Google Scholar]
  13. Gupta, A.K.; Seal, A.; Prasad, M.; Khanna, P. Salient object detection techniques in computer vision—A survey. Entropy 2020, 22, 1174. [Google Scholar] [CrossRef]
  14. Wang, W.; Lai, Q.; Fu, H.; Shen, J.; Ling, H.; Yang, R. Salient object detection in the deep learning era: An in-depth survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3239–3259. [Google Scholar] [CrossRef]
  15. Gao, S.H.; Tan, Y.Q.; Cheng, M.M.; Lu, C.; Chen, Y.; Yan, S. Highly efficient salient object detection with 100k parameters. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part VI. Springer: Berlin/Heidelberg, Germany, 2020; pp. 702–721. [Google Scholar]
  16. Liu, Y.; Zhang, X.Y.; Bian, J.W.; Zhang, L.; Cheng, M.M. SAMNet: Stereoscopically attentive multi-scale network for lightweight salient object detection. IEEE Trans. Image Process. 2021, 30, 3804–3814. [Google Scholar] [CrossRef] [PubMed]
  17. Liu, N.; Han, J.; Yang, M.H. Picanet: Learning pixel-wise contextual attention for saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3089–3098. [Google Scholar]
  18. Zhang, P.; Wang, D.; Lu, H.; Wang, H.; Ruan, X. Amulet: Aggregating multi-level convolutional features for salient object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 202–211. [Google Scholar]
  19. Liu, Y.; Gu, Y.C.; Zhang, X.Y.; Wang, W.; Cheng, M.M. Lightweight salient object detection via hierarchical visual perception learning. IEEE Trans. Cybern. 2020, 51, 4439–4449. [Google Scholar] [CrossRef] [PubMed]
  20. Shapley, R.; Hawken, M.J. Color in the cortex: Single-and double-opponent cells. Vis. Res. 2011, 51, 701–717. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Kruger, N.; Janssen, P.; Kalkan, S.; Lappe, M.; Leonardis, A.; Piater, J.; Rodriguez-Sanchez, A.J.; Wiskott, L. Deep hierarchies in the primate visual cortex: What can we learn for computer vision? IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1847–1871. [Google Scholar] [CrossRef]
  22. Nunez, V.; Shapley, R.M.; Gordon, J. Cortical double-opponent cells in color perception: Perceptual scaling and chromatic visual evoked potentials. i-Perception 2018, 9, 2041669517752715. [Google Scholar] [CrossRef] [Green Version]
  23. Conway, B.R. Color vision, cones and color-coding in the cortex. Neuroscientist 2009, 15, 274–290. [Google Scholar] [CrossRef]
  24. Conway, B.R. Spatial structure of cone inputs to color cells in alert macaque primary visual cortex (V-1). J. Neurosci. 2001, 21, 2768–2783. [Google Scholar] [CrossRef] [Green Version]
  25. Hunt, R.W.G.; Pointer, M.R. Measuring Colour; John Wiley & Sons: Hoboken, NJ, USA, 2011. [Google Scholar]
  26. Engel, S.; Zhang, X.; Wandell, B. Colour tuning in human visual cortex measured with functional magnetic resonance imaging. Nature 1997, 388, 68–71. [Google Scholar] [CrossRef]
  27. Shapley, R. Physiology of color vision in primates. In Oxford Research Encyclopedia of Neuroscience; Oxford University Press: Oxford, UK, 2019. [Google Scholar]
  28. Qin, X.; Zhang, Z.; Huang, C.; Dehghan, M.; Zaiane, O.R.; Jagersand, M. U2-Net: Going deeper with nested U-structure for salient object detection. Pattern Recognit. 2020, 106, 107404. [Google Scholar] [CrossRef]
  29. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  30. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  31. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  32. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar]
  33. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
  34. Frintrop, S.; Werner, T.; Martin Garcia, G. Traditional saliency reloaded: A good old model in new shape. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 82–90. [Google Scholar]
  35. Mäenpää, T.; Pietikäinen, M. Classification with color and texture: Jointly or separately? Pattern Recognit. 2004, 37, 1629–1640. [Google Scholar] [CrossRef] [Green Version]
  36. Chan, C.H.; Kittler, J.; Messer, K. Multispectral local binary pattern histogram for component-based color face verification. In Proceedings of the 2007 First IEEE International Conference on Biometrics: Theory, Applications and Systems, Crystal City, Virginia, 27–29 September 2007; pp. 1–7. [Google Scholar]
  37. Faloutsos, C.; Lin, K.I. FastMap: A Fast Algorithm for Indexing, Data-Mining and Visualization of Traditional and Multimedia Datasets; ACM: Rochester, NY, USA, 1995; Volume 24. [Google Scholar]
  38. Jain, A.; Healey, G. A multiscale representation including opponent color features for texture recognition. IEEE Trans. Image Process. 1998, 7, 124–128. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Yang, K.F.; Gao, S.B.; Guo, C.F.; Li, C.Y.; Li, Y.J. Boundary detection using double-opponency and spatial sparseness constraint. IEEE Trans. Image Process. 2015, 24, 2565–2578. [Google Scholar] [CrossRef] [PubMed]
  40. Hurvich, L.M.; Jameson, D. An opponent-process theory of color vision. Psychol. Rev. 1957, 64, 384. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Farabet, C.; Couprie, C.; Najman, L.; LeCun, Y. Learning hierarchical features for scene labeling. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1915–1929. [Google Scholar] [CrossRef] [Green Version]
  42. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning. Pmlr, Lille, France, 7–9 July 2015; pp. 448–456. [Google Scholar]
  43. Santurkar, S.; Tsipras, D.; Ilyas, A.; Madry, A. How does batch normalization help optimization? Adv. Neural Inf. Process. Syst. 2018, 31, 2483–2493. [Google Scholar]
  44. Chollet, F. Deep Learning with Python; Simon and Schuster: Manhattan, NY, USA, 2021. [Google Scholar]
  45. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  46. Ghiasi, G.; Lin, T.Y.; Le, Q.V. Dropblock: A regularization method for convolutional networks. Adv. Neural Inf. Process. Syst. 2018, 31, 10727–10737. [Google Scholar]
  47. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  48. Pietikäinen, M.; Hadid, A.; Zhao, G.; Ahonen, T. Computer Vision Using Local Binary Patterns; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2011; Volume 40. [Google Scholar]
  49. Voulodimos, A.; Doulamis, N.; Doulamis, A.; Protopapadakis, E. Deep learning for computer vision: A brief review. Comput. Intell. Neurosci. 2018, 2018, 7068349. [Google Scholar] [CrossRef]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  51. Chollet, F. Keras. 2015. Available online: https://keras.io (accessed on 9 June 2023).
  52. Borji, A.; Cheng, M.M.; Jiang, H.; Li, J. Salient object detection: A benchmark. IEEE Trans. Image Process. 2015, 24, 5706–5722. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Shi, J.; Yan, Q.; Xu, L.; Jia, J. Hierarchical image saliency detection on extended CSSD. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 717–729. [Google Scholar] [CrossRef] [PubMed]
  54. Yang, C.; Zhang, L.; Lu, H.; Ruan, X.; Yang, M.H. Saliency detection via graph-based manifold ranking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 23–28 June 2013; pp. 3166–3173. [Google Scholar]
  55. Wang, L.; Lu, H.; Wang, Y.; Feng, M.; Wang, D.; Yin, B.; Ruan, X. Learning to detect salient objects with image-level supervision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 136–145. [Google Scholar]
  56. Li, G.; Yu, Y. Visual saliency based on multiscale deep features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5455–5463. [Google Scholar]
  57. Cheng, M.M.; Mitra, N.J.; Huang, X.; Hu, S.M. Salientshape: Group saliency in image collections. Vis. Comput. 2014, 30, 443–453. [Google Scholar] [CrossRef]
  58. Feng, W.; Li, X.; Gao, G.; Chen, X.; Liu, Q. Multi-scale global contrast CNN for salient object detection. Sensors 2020, 20, 2656. [Google Scholar] [CrossRef] [PubMed]
  59. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  60. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  61. Tieleman, T.; Hinton, G. Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude. COURSERA Neural Netw. Mach. Learn. 2012, 4, 26–31. [Google Scholar]
  62. Margolin, R.; Zelnik-Manor, L.; Tal, A. How to evaluate foreground maps? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 23–28 June 2014; pp. 248–255. [Google Scholar]
  63. Varadarajan, V.; Garg, D.; Kotecha, K. An efficient deep convolutional neural network approach for object detection and recognition using a multi-scale anchor box in real-time. Future Internet 2021, 13, 307. [Google Scholar] [CrossRef]
  64. Jiang, H.; Wang, J.; Yuan, Z.; Wu, Y.; Zheng, N.; Li, S. Salient object detection: A discriminative regional feature integration approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2083–2090. [Google Scholar]
  65. Li, G.; Yu, Y. Deep contrast learning for salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 478–487. [Google Scholar]
  66. Liu, N.; Han, J. Dhsnet: Deep hierarchical saliency network for salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 678–686. [Google Scholar]
  67. Wei, J.; Zhong, B. Saliency detection using fully convolutional network. In Proceedings of the 2018 Chinese Automation Congress (CAC), Xi’an, China, 30 November–2 December 2018; pp. 3902–3907. [Google Scholar]
  68. Luo, Z.; Mishra, A.; Achkar, A.; Eichel, J.; Li, S.; Jodoin, P.M. Non-local deep features for salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6609–6617. [Google Scholar]
  69. Hou, Q.; Cheng, M.M.; Hu, X.; Borji, A.; Tu, Z.; Torr, P.H. Deeply supervised salient object detection with short connections. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3203–3212. [Google Scholar]
  70. Zhang, P.; Wang, D.; Lu, H.; Wang, H.; Yin, B. Learning uncertain convolutional features for accurate saliency detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 212–221. [Google Scholar]
  71. Wang, T.; Borji, A.; Zhang, L.; Zhang, P.; Lu, H. A stagewise refinement model for detecting salient objects in images. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4019–4028. [Google Scholar]
  72. Wang, T.; Zhang, L.; Wang, S.; Lu, H.; Yang, G.; Ruan, X.; Borji, A. Detect globally, refine locally: A novel approach to saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3127–3135. [Google Scholar]
  73. Li, X.; Yang, F.; Cheng, H.; Liu, W.; Shen, D. Contour knowledge transfer for salient object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 355–370. [Google Scholar]
  74. Chen, S.; Tan, X.; Wang, B.; Hu, X. Reverse attention for salient object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 234–250. [Google Scholar]
  75. Liu, Y.; Cheng, M.M.; Zhang, X.Y.; Nie, G.Y.; Wang, M. DNA: Deeply supervised nonlinear aggregation for salient object detection. IEEE Trans. Cybern. 2021, 52, 6131–6142. [Google Scholar] [CrossRef]
  76. Wu, Z.; Su, L.; Huang, Q. Cascaded partial decoder for fast and accurate salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3907–3916. [Google Scholar]
  77. Qin, X.; Zhang, Z.; Huang, C.; Gao, C.; Dehghan, M.; Jagersand, M. Basnet: Boundary-aware salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7479–7489. [Google Scholar]
  78. Feng, M.; Lu, H.; Ding, E. Attentive feedback network for boundary-aware salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1623–1632. [Google Scholar]
  79. Liu, J.J.; Hou, Q.; Cheng, M.M.; Feng, J.; Jiang, J. A simple pooling-based design for real-time salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3917–3926. [Google Scholar]
  80. Zhao, J.X.; Liu, J.J.; Fan, D.P.; Cao, Y.; Yang, J.; Cheng, M.M. EGNet: Edge guidance network for salient object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 8779–8788. [Google Scholar]
  81. Su, J.; Li, J.; Zhang, Y.; Xia, C.; Tian, Y. Selectivity or invariance: Boundary-aware salient object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 3799–3808. [Google Scholar]
  82. Zhao, H.; Qi, X.; Shen, X.; Shi, J.; Jia, J. Icnet for real-time semantic segmentation on high-resolution images. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 405–420. [Google Scholar]
  83. Yu, C.; Wang, J.; Peng, C.; Gao, C.; Yu, G.; Sang, N. Bisenet: Bilateral segmentation network for real-time semantic segmentation. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 325–341. [Google Scholar]
  84. Li, H.; Xiong, P.; Fan, H.; Sun, J. Dfanet: Deep feature aggregation for real-time semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9522–9531. [Google Scholar]
Figure 1. The CoSOV1 (cone- and spatial-opponent primary visual cortex) module is the core of our neural network model.
Figure 1. The CoSOV1 (cone- and spatial-opponent primary visual cortex) module is the core of our neural network model.
Sensors 23 06450 g001
Figure 2. Our model CoSOV1 neural network architecture consisting of 5 blocks: Pairing_Color_Unit, Encoder_Unit, Middle_Unit, Dec_Res_Block and Dec_Dconv_Block.
Figure 2. Our model CoSOV1 neural network architecture consisting of 5 blocks: Pairing_Color_Unit, Encoder_Unit, Middle_Unit, Dec_Res_Block and Dec_Dconv_Block.
Sensors 23 06450 g002
Figure 3. Pairing_Color_Unit: input RGB color image is transformed in 6 opposing color channel pairs; these are then concatenated to obtain 12 color channels.
Figure 3. Pairing_Color_Unit: input RGB color image is transformed in 6 opposing color channel pairs; these are then concatenated to obtain 12 color channels.
Sensors 23 06450 g003
Figure 4. Encoder unit: (a) encoder unit; (b) the residual block; (c) CoSOV1 module.
Figure 4. Encoder unit: (a) encoder unit; (b) the residual block; (c) CoSOV1 module.
Sensors 23 06450 g004
Figure 5. Simplified flowchart in CoSOV1 module for processing pairs of opposing color pairs (or group of feature maps).
Figure 5. Simplified flowchart in CoSOV1 module for processing pairs of opposing color pairs (or group of feature maps).
Sensors 23 06450 g005
Figure 6. (a) The middle unit, (b) the CoSOV1 module.
Figure 6. (a) The middle unit, (b) the CoSOV1 module.
Sensors 23 06450 g006
Figure 7. (a) The decoder unit; (b) the decoder residual block; (c) the decoder deconvolution block.
Figure 7. (a) The decoder unit; (b) the decoder residual block; (c) the decoder deconvolution block.
Sensors 23 06450 g007
Figure 8. Precision curves for (a) ECSSD, (b) DUT-OMRON, (c) DUTS-TE, (d) HKU-IS and (e) THUR15K datasets.
Figure 8. Precision curves for (a) ECSSD, (b) DUT-OMRON, (c) DUTS-TE, (d) HKU-IS and (e) THUR15K datasets.
Sensors 23 06450 g008
Figure 9. Precision–recall curves for (a) ECSSD, (b) DUT-OMRON, (c) DUTS-TE, (d) HKU-IS and (e) THUR15K datasets.
Figure 9. Precision–recall curves for (a) ECSSD, (b) DUT-OMRON, (c) DUTS-TE, (d) HKU-IS and (e) THUR15K datasets.
Sensors 23 06450 g009
Figure 10. F β measure curves for (a) ECSSD, (b) DUT-OMRON, (c) DUTS-TE, (d) HKU-IS and (e) THUR15K datasets.
Figure 10. F β measure curves for (a) ECSSD, (b) DUT-OMRON, (c) DUTS-TE, (d) HKU-IS and (e) THUR15K datasets.
Sensors 23 06450 g010
Figure 11. Comparison between SAMNet [16], HVPNet [19] and our proposed model, CoSOV1Net, on some image saliency maps: 1st column: images; 2nd column: ground truth or human visual system saliency map; 3rd column: SAMNet; 4th column: HVPNet; 5th column: CoSOV1Net (ours).
Figure 11. Comparison between SAMNet [16], HVPNet [19] and our proposed model, CoSOV1Net, on some image saliency maps: 1st column: images; 2nd column: ground truth or human visual system saliency map; 3rd column: SAMNet; 4th column: HVPNet; 5th column: CoSOV1Net (ours).
Sensors 23 06450 g011aSensors 23 06450 g011bSensors 23 06450 g011c
Figure 12. Example of trade-off between (a) F β measure and #parameters; (b) F β ω measure and #parameters; (c) F β measure and FLOPS; (d) F β measure and FPS, for ECSSD.
Figure 12. Example of trade-off between (a) F β measure and #parameters; (b) F β ω measure and #parameters; (c) F β measure and FLOPS; (d) F β measure and FPS, for ECSSD.
Sensors 23 06450 g012
Table 1. Our proposed model F-measure ( F β , β 2 = 0.3 ) compared with 20 state-of-the-art models (best value in bold) [# Param: number of parameters, ↑: great is best, ↓: small is the best].
Table 1. Our proposed model F-measure ( F β , β 2 = 0.3 ) compared with 20 state-of-the-art models (best value in bold) [# Param: number of parameters, ↑: great is best, ↓: small is the best].
Methods# Param (M) ↓FLOPS (G) ↓Speed (FPS) ↑ECSSDDUT-OMRONDUTS-TEHKU-ISTHUR15K
DRFI [64]--0.10.7770.6520.6490.7740.670
DCL [65]66.24224.91.40.8950.7330.7850.8920.747
DHSNet [66]94.0415.810.00.903-0.8070.8890.752
RFCN [67]134.69102.80.40.8960.7380.7820.8920.754
NLDF [68]35.49263.918.50.9020.7530.8060.9020.762
DSS [69]62.23114.67.00.9150.7740.8270.9130.770
Amulet [18]33.1545.39.70.9130.7430.7780.8970.755
UCF [70]23.9861.412.00.9010.7300.7720.8880.758
SRM [71]43.7420.312.30.9140.7690.8260.9060.778
PiCANet  [17]32.8537.15.60.9230.7660.8370.9160.783
BRN [72]126.3524.13.60.9190.7740.8270.9100.769
C2S [73]137.0320.516.70.9070.7590.8110.8980.775
RAS [74]20.1335.620.40.9160.7850.8310.9130.772
DNA [75]20.0682.525.00.9350.7990.8650.9300.793
CPD [76]29.2359.568.00.9300.7940.8610.9240.795
BASNet  [77]87.06127.336.20.9380.8050.8590.9280.783
AFNet  [78]37.1138.421.60.9300.7840.8570.9210.791
PoolNet [79]53.63123.439.70.9340.7910.8660.9250.800
EGNet [80]108.07270.812.70.9380.7940.8700.9280.800
BANet  [81]55.90121.612.50.9400.8030.8720.9320.796
CoSOV1Net (OURS)1.141.4211.20.9310.7890.8330.9120.773
Table 2. Our proposed model MAE (↓) compared with 20 state-of-the-art models (best performance in bold) [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Table 2. Our proposed model MAE (↓) compared with 20 state-of-the-art models (best performance in bold) [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Methods# Param (M) ↓FLOPS (G) ↓Speed (FPS) ↑ECSSDDUT-OMRONDUTS-TEHKU-ISTHUR15K
DRFI [64]--0.10.1610.1380.1540.1460.150
DCL [65]66.24224.91.40.0800.0950.0820.0630.096
DHSNet [66]94.0415.810.00.062-0.0660.0530.082
RFCN [67]134.69102.80.40.0970.0950.0890.0800.100
NLDF [68]35.49263.918.50.0660.0800.0650.0480.080
DSS [69]62.23114.67.00.0560.0660.0560.0410.074
Amulet [18]33.1545.39.70.0610.0980.0850.0510.094
UCF [70]23.9861.412.00.0710.1200.1120.0620.112
SRM [71]43.7420.312.30.0560.0690.0590.0460.077
PiCANet  [17]32.8537.15.60.0490.0680.0540.0420.083
BRN [72]126.3524.13.60.0430.0620.0500.0360.076
C2S [73]137.0320.516.70.0570.0720.0620.0460.083
RAS [74]20.1335.620.40.0580.0630.0590.0450.075
DNA [75]20.0682.525.00.0410.0560.0440.0310.069
CPD [76]29.2359.568.00.0440.0570.0430.0330.068
BASNet  [77]87.06127.336.20.0400.0560.0480.0320.073
AFNet  [78]37.1138.421.60.0450.0570.0460.0360.072
PoolNet [79]53.63123.439.70.0480.0570.0430.0370.068
EGNet [80]108.07270.812.70.0440.0560.0440.0340.070
BANet  [81]55.90121.612.50.0380.0590.0400.0310.068
CoSOV1Net (OURS)1.141.4211.20.0510.0640.0570.0450.076
Table 3. Our proposed model weighted F-measure ( F β ω , β 2 = 1 ) compared with 20 state-of-the-art models (best value in bold) [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Table 3. Our proposed model weighted F-measure ( F β ω , β 2 = 1 ) compared with 20 state-of-the-art models (best value in bold) [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Methods# Param (M) ↓FLOPS (G) ↓Speed (FPS) ↑ECSSDDUT-OMRONDUTS-TEHKU-ISTHUR15K
DRFI [64]--0.10.5480.4240.3780.5040.444
DCL [65]66.24224.91.40.7820.5840.6320.7700.624
DHSNet [66]94.0415.810.00.837-0.7050.8160.666
RFCN [67]134.69102.80.40.7250.5620.5860.7070.591
NLDF [68]35.49263.918.50.8350.6340.7100.8380.676
DSS [69]62.23114.67.00.8640.6880.7520.8620.702
Amulet [18]33.1545.39.70.8390.6260.6570.8170.650
UCF [70]23.9861.412.00.8050.5730.5950.7790.613
SRM [71]43.7420.312.30.8490.6580.7210.8350.684
PiCANet  [17]32.8537.15.60.8620.6910.7450.8470.687
BRN [72]126.3524.13.60.8870.7090.7740.8750.712
C2S [73]137.0320.516.70.8490.6630.7170.8350.685
RAS [74]20.1335.620.40.8550.6950.7390.8490.691
DNA [75]20.0682.525.00.8970.7290.7970.8890.723
CPD [76]29.2359.568.00.8890.7150.7990.8790.731
BASNet  [77]87.06127.336.20.8980.7510.8020.8890.721
AFNet  [78]37.1138.421.60.8800.7170.7840.8690.719
PoolNet [79]53.63123.439.70.8750.7100.7830.8640.724
EGNet [80]108.07270.812.70.8860.7270.7960.8760.727
BANet  [81]55.90121.612.50.9010.7360.8100.8890.730
CoSOV1Net (OURS)1.141.4211.20.8610.6960.7310.8340.688
Table 4. Our proposed model’s F-measure ( F β , β 2 = 0.3 ) compared with state-of-the-art lightweight salient object-detection models (best value in bold) [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Table 4. Our proposed model’s F-measure ( F β , β 2 = 0.3 ) compared with state-of-the-art lightweight salient object-detection models (best value in bold) [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Methods# Param (M) ↓FLOPS (G) ↓Speed (FPS) ↑ECSSDDUT-OMRONDUTS-TEHKU-ISTHUR15K
MobileNet * [30]4.272.2295.80.9060.7530.8040.8950.767
MobileNetV2 *  [31]2.370.8446.20.9050.7580.7980.8900.766
ShuffleNet * [32]1.800.7406.90.9070.7570.8110.8980.771
ShuffleNetV2 *  [33]1.600.5452.50.9010.7460.7890.8840.755
ICNet [82]6.706.375.10.9180.7730.8100.8980.768
BiSeNet R18 [83]13.4825.0120.50.9090.7570.8150.9020.776
BiSeNet X39 [83]1.847.3165.80.9010.7550.7870.8880.756
DFANet [84]1.831.791.40.8960.7500.7910.8840.757
HVPNet [19]1.231.1333.20.9250.7990.8390.9150.787
SAMNet [16]1.330.5343.20.9250.7970.8350.9150.785
CoSOV1Net (OURS)1.141.4211.20.9310.7890.8330.9120.773
* SAMNet, where the encoder is replaced by this backbone.
Table 5. Our proposed model MAE (↓) compared with state-of-the art lightweight salient object-detection models (best value in bold) [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Table 5. Our proposed model MAE (↓) compared with state-of-the art lightweight salient object-detection models (best value in bold) [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Methods# Param (M) ↓FLOPS (G) ↓Speed (FPS) ↑ECSSDDUT-OMRONDUTS-TEHKU-ISTHUR15K
MobileNet * [30]4.272.2295.80.0640.0730.0660.0520.081
MobileNetV2 *  [31]2.370.8446.20.0660.0750.0700.0560.085
ShuffleNet * [32]1.800.7406.90.0620.0690.0620.0500.078
ShuffleNetV2 *  [33]1.600.5452.50.0690.0760.0710.0590.086
ICNet [82]6.706.375.10.0590.0720.0670.0520.084
BiSeNet R18 [83]13.4825.0120.50.0620.0720.0620.0490.080
BiSeNet X39 [83]1.847.3165.80.0700.0780.0740.0590.090
DFANet [84]1.831.791.40.0730.0780.0750.0610.089
HVPNet [19]1.231.1333.20.0550.0640.0580.0450.076
SAMNet [16]1.330.5343.20.0530.0650.0580.0450.077
CoSOV1Net (OURS)1.141.4211.20.0510.0640.0570.0450.076
* SAMNet, where the encoder is replaced by this backbone.
Table 6. Our proposed model’s weighted F-measure ( F β ω , β 2 = 1 ) compared with lightweight salient object-detection models (best value in bold) [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Table 6. Our proposed model’s weighted F-measure ( F β ω , β 2 = 1 ) compared with lightweight salient object-detection models (best value in bold) [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Methods# Param (M) ↓FLOPS (G) ↓Speed (FPS) ↑ECSSDDUT-OMRONDUTS-TEHKU-ISTHUR15K
MobileNet * [30]4.272.2295.80.8290.6560.6960.8160.675
MobileNetV2 * [31]2.370.8446.20.8200.6510.6760.7990.660
ShuffleNet * [32]1.800.7406.90.8310.6670.7090.8200.683
ShuffleNetV2 *  [33]1.600.5452.50.8120.6370.6650.7880.652
ICNet [82]6.706.375.10.8380.6690.6940.8120.668
BiSeNet R18 [83]13.4825.0120.50.8290.6480.6990.8190.675
BiSeNet X39 [83]1.847.3165.80.8020.6320.6520.7840.641
DFANet [84]1.831.791.40.7990.6270.6520.7780.639
HVPNet [19]1.231.1333.20.8540.6990.7300.8390.696
SAMNet [16]1.330.5343.20.8550.6990.7290.8370.693
CoSOV1Net (OURS)1.141.4211.20.8610.6960.7310.8340.688
* SAMNet, where the encoder is replaced by this backbone.
Table 7. Our proposed model (CoSOV1Net)’s ranking with respect to existing salient object detection [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Table 7. Our proposed model (CoSOV1Net)’s ranking with respect to existing salient object detection [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Measure# Param (M) ↓FLOPS (G) ↓Speed (FPS) ↑ECSSDDUT-OMRONDUTS-TEHKU-ISTHUR15K
F β 1st1st1st6th7th9th11th11th
MAE1st1st1st10th10th11th11th10th
F β ω 1st1st1st11th9th11th15th11th
Table 8. Our proposed model (CoSOV1Net)’s ranking with respect to lightweight salient object-detection models [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Table 8. Our proposed model (CoSOV1Net)’s ranking with respect to lightweight salient object-detection models [# Param: number of parameters, ↑: great is the best, ↓: small is the best].
Measure# Param (M) ↓FLOPS (G) ↓Speed (FPS) ↑ECSSDDUT-OMRONDUTS-TEHKU-ISTHUR15K
F β 1st6th7th1st3rd3rd3rd4th
MAE1st6th7th1st1st1st1st2nd
F β ω 1st6th7th1st3rd1st3rd3rd
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ndayikengurukiye, D.; Mignotte, M. CoSOV1Net: A Cone- and Spatial-Opponent Primary Visual Cortex-Inspired Neural Network for Lightweight Salient Object Detection. Sensors 2023, 23, 6450. https://doi.org/10.3390/s23146450

AMA Style

Ndayikengurukiye D, Mignotte M. CoSOV1Net: A Cone- and Spatial-Opponent Primary Visual Cortex-Inspired Neural Network for Lightweight Salient Object Detection. Sensors. 2023; 23(14):6450. https://doi.org/10.3390/s23146450

Chicago/Turabian Style

Ndayikengurukiye, Didier, and Max Mignotte. 2023. "CoSOV1Net: A Cone- and Spatial-Opponent Primary Visual Cortex-Inspired Neural Network for Lightweight Salient Object Detection" Sensors 23, no. 14: 6450. https://doi.org/10.3390/s23146450

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop