Next Article in Journal
Multi-Gene Genetic Programming-Based Identification of a Dynamic Prediction Model of an Overhead Traveling Crane
Next Article in Special Issue
Localization and Edge-Based Segmentation of Lumbar Spine Vertebrae to Identify the Deformities Using Deep Learning Models
Previous Article in Journal
An Efficient and Secure Revocation-Enabled Attribute-Based Access Control for eHealth in Smart Society
Previous Article in Special Issue
Wildfire Smoke Classification Based on Synthetic Images and Pixel- and Feature-Level Domain Adaptation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DTS-Net: Depth-to-Space Networks for Fast and Accurate Semantic Object Segmentation

1
Department of Information and Communication Engineering, School of Electrical and Computer Engineering, Chungbuk National University, Cheongju-si 28644, Korea
2
Electrical Engineering Department, Faculty of Engineering, Assiut University, Assiut 71515, Egypt
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(1), 337; https://doi.org/10.3390/s22010337
Submission received: 2 December 2021 / Revised: 23 December 2021 / Accepted: 29 December 2021 / Published: 3 January 2022
(This article belongs to the Special Issue Neural Networks and Deep Learning in Image Sensing)

Abstract

:
We propose Depth-to-Space Net (DTS-Net), an effective technique for semantic segmentation using the efficient sub-pixel convolutional neural network. This technique is inspired by depth-to-space (DTS) image reconstruction, which was originally used for image and video super-resolution tasks, combined with a mask enhancement filtration technique based on multi-label classification, namely, Nearest Label Filtration. In the proposed technique, we employ depth-wise separable convolution-based architectures. We propose both a deep network, that is, DTS-Net, and a lightweight network, DTS-Net-Lite, for real-time semantic segmentation; these networks employ Xception and MobileNetV2 architectures as the feature extractors, respectively. In addition, we explore the joint semantic segmentation and depth estimation task and demonstrate that the proposed technique can efficiently perform both tasks simultaneously, outperforming state-of-art (SOTA) methods. We train and evaluate the performance of the proposed method on the PASCAL VOC2012, NYUV2, and CITYSCAPES benchmarks. Hence, we obtain high mean intersection over union (mIOU) and mean pixel accuracy (Pix.acc.) values using simple and lightweight convolutional neural network architectures of the developed networks. Notably, the proposed method outperforms SOTA methods that depend on encoder–decoder architectures, although our implementation and computations are far simpler.

1. Introduction

Semantic segmentation is an important task in computer vision, as it constitutes pixel-wise classification of an image to mask each object in the scene. Hundreds of applications, such as autonomous driving, robotics, medical diagnostics, image editing, and augmented reality applications, incorporate semantic segmentation. Recent studies on semantic segmentation have achieved promising results using convolutional neural networks (CNNs), particularly encoder–decoder CNN architectures. In such architectures, the semantic segmentation task is modeled in two stages: the encoding stage, in which the image is down-sampled to obtain its deep semantic features, and the decoding stage, in which the semantic features are up-sampled to obtain a semantic segmentation mask of the same size as the input image. Encoder–decoder architecture can achieve highly accurate segmentation results; however, the decoder stage adds considerable computational complexity to the overall model and the input image size of such models is usually small. Furthermore, the decoding-stage reconstruction is inefficient because de-convolution or up-sampling layers are used, which eliminate small details and sometimes propagate noise.
The semantic segmentation problem can be seen as an image-to-image translation problem, as the aim of this task is to construct segmentation masks equivalent to each object in the input image. Further, the objects in the segmentation mask have the same boundaries as the input image; thus, the segmentation mask can be considered as an alternative to the image and, hence, the segmentation task can be performed using image reconstruction techniques. The efficient sub-pixel CNN [1] has shown promising results in image and video super-resolution tasks because of its depth-to-space (DTS) layer, from which a high-resolution image can be reconstructed from many low-resolution images. This layer performs image reconstruction through pixel reordering of the low-resolution feature maps obtained from the CNN network, in order to form super-pixels of the high-resolution image with very accurate borders and object details. Thus, the DTS layer efficiently obtains accurate semantic segmentation masks with clear borders, at far higher speeds than the traditional encoder–decoder architectures usually used for segmentation tasks. The DTS layer has many advantages, including its lower computation count and considerably higher accuracy compared with the decoder stage of the encoder–decoder architecture. Therefore, modeling can be accelerated, while high segmentation accuracy is maintained.
Vision transformers, which apply a transformer network to images in a manner similar to natural language processing, are a recent achievement in computer vision. Vision transformers allow parallelization in the processing of a sequence dataset or a sequence of patches of the target image. This is achieved through positional encoding, which allows the network to learn the position of the patch in the input “big image”. Vision transformer-based methods achieve better accuracy than CNN-based methods for classification, segmentation, and object detection tasks without increased computational cost.
In this study, we propose the DTS-Net deep network for semantic segmentation using a sub-pixel CNN, to address the semantic segmentation complexity problem that arises for encoder–decoder architectures with or without attention methods while retaining high segmentation accuracy. We also present DTS-Net-Lite, a lightweight version of this network. Our contributions can be summarized as follows:
  • Rather than a pixel-wise classification problem, we treat the semantic segmentation problem as an image-to-image translation problem through regression using the DTS layer, and construct segmentation maps using the higher-resolution image reconstruction approach of the super-resolution task;
  • We propose DTS-Net, a deep model that uses Xception architecture as a feature extractor for high-accuracy critical applications, as well as a small lightweight model, DTS-Net-Lite, for high-speed critical applications that uses MobileNetV2 architecture as a feature extractor;
  • We reduce the typical decoding stage complexity for segmentation mask construction to that of the DTS image construction layer and show that this layer can construct segmentation masks with far lower computational cost and much higher precision than conventional CNN-based decoding architectures.
  • We propose a new segmentation improvement technique namely nearest label filtration (NLF) to improve the segmentation by correcting the wrong predicted pixels by DTS-layer in the segmentation mask.
The proposed method achieves higher accuracy and speed than the recent semantic segmentation methods. Further, learning of highly detailed feature maps is possible as we depend on CNN architecture in addition to the DTS layer; this allows the construction of higher-resolution prediction maps. In addition, we explore the joint semantic segmentation and depth estimation task and achieve promising results. A preview of our results is shown in Figure 1. The remainder of this paper is structured as follows. Section 2 summarizes related work and Section 3 presents the proposed method and architectures. Section 4 and Section 5 discuss the training and test datasets and present the results with comparison to state-of-the-art (SOTA) methods, respectively. Section 6 discusses future work and Section 7 presents conclusions.

2. Related Work

Recent semantic segmentation studies have shown that encoder–decoder architecture can efficiently perform segmentation tasks. The first encoder–decoder architecture was the fully convolutional network (FCN) [5], in which the same architecture was used for image classification but the final dense layers were replaced with 1 × 1 convolutional layers having the same weights. The decoder stage was a simple up-sampling layer. The FCN obtained relatively good results, motivating further research on encoder–decoder architectures. Later, SegNet [6] was proposed as a deep encoder–decoder architecture. This architecture features pooling indices shared between the max-pooling layers in the encoder stage and the corresponding max-unpooling layers in the decoder stage. SegNet exhibited impressive semantic segmentation results on outdoor and indoor segmentation datasets. U-Net [7] is another impressive encoder–decoder architecture, which was proposed for microscopy cell segmentation in medical images. U-net suggests residual connections between corresponding layers in the encoder and decoder stages and, hence, achieves considerable segmentation accuracy.
The four versions of DeepLabVx also constitute considerable contributions to the semantic segmentation task. DeepLabV1 [8] tackled the problem of inefficient down-sampling through a wide field of view convolution using the Atrous convolution [8], which increases the spatial field of the convolutional window using the same weights as the normal convolution. In that work, the conditional random field (CRF) was also proposed, which uses an energy function derived from the summation of a unary potential term calculated from the probability distribution of the output label of each pixel, along with a binary potential term calculated from the correlation between pixel labels. In general, the CRF allows the model to learn small image details. With DeepLabV2 [9], Atrous Spatial Pyramid Pooling (ASPP) was proposed to enhance the model learning at multiple scales of the feature maps. In addition, the VGG16 [10] used in DeeplabV1 was replaced with ResNet [11], which yielded better performance. The ASPP was further improved for DeepLabV3 [12], with the use of different sampling rates in the ASPP in a cascaded manner. Finally, for DeepLabV3+ [13], the encoder architecture was replaced with a depth-wise separable convolution-based architecture, and Aligned Xception was adopted. The latter is a modified version of Xception [14] that replaces the max-pooling layers in the original architecture with strided convolutional layers with more Xception blocks; this facilitates higher accuracy and speed during processing.
Zhao et al. proposed the pyramid scene parsing network (PSPNet) [15], which adopts a pyramid parsing module that learns the global context of the image through region-based aggregation. Pyramid pooling is employed to learn the context of the image from the final small feature maps. Later, Zhao et al. presented the image cascade network (ICNet) [16], which performs high-speed semantic segmentation on high-resolution images using cascade feature fusion. The latter approach mixes the features obtained from the input image on different scales in a method called cascade label guidance. Up-sampling layers are then used to resize the image to the input image size. As another approach, the bilateral segmentation network (BiSeNet) [17] performs real-time semantic segmentation using an architecture with two paths. The first is a spatial path for spatial information preservation and the other is a context path for general context learning through down-sampling. The features obtained by the two paths are then combined via a feature fusion technique. ResNeSt, developed by Zhang et al. [18], applies channel-wise attention to different network branches to learn diverse feature representations and cross-feature information. Shi et al. [19] proposed Hierarchical Parsing Net for semantic segmentation which enhances scene parsing by learning the global scene information and the contextual relation between objects in the scene using a deep neural network. Chen et al. [20] proposed a one-shot semantic segmentation method which uses the multiclass labels information during training to encourage the network to learn more accurate semantic features of each category and they also proposed the pyramid feature fusion module to mine the fused features of the objects and a self-prototype guidance branch to support the segmentation task. Although all the previously mentioned methods presented challenging results, they adopted inefficient decoding stages that eliminate the object details and introduce some noise.
Zoph et al. [21] proposed pre-training and self-training techniques using stronger augmentation on ImageNet [22] across the different image sizes using EfficientNet [23] architecture, and showed that pre-training and self-training are mutually beneficial and improve the accuracies of both the semantic segmentation and object detection tasks. Further, Rashwan et al. [24] proposed dilated SpineNet or SpineNet-Seg, which is a neural architecture search discovered network from DeepLabv3. In this approach, the scale permuted networks originally used for object detection in the semantic segmentation task are evaluated, adopting a customized dilation ratio per block. Bai et al. [25] proposed the multiscale deep equilibrium model (MDEQ), MDEQ backpropagates through the equilibrium points of multiple scale features simultaneously using simple differentiation to avoid storing intermediate states, they attained a high segmentation accuracy on CITYSCAPES however, their model has high computational complexity. Termritthikun et al. [26] proposed EEEA-Net in which they employed a neural architecture search method to search an optimized model with the lowest number of parameters by using an early exit population initialization algorithm. They could achieve an average accuracy segmentation model but the number of the parameters of their model was extremely low. Ding et al. [27] recently proposed RepVGG in which they adopted a VGG-like architecture composed of a stack of 3 × 3 convolution and Relu. RepVGG models ran much faster than ResNet-50 or ResNet-101 with higher accuracy on the classification and semantic segmentation tasks. It is important to note that Aich et al. [28] initiated the direction of using depth-to-space for segmentation when they employed it to perform binary segmentation for satellite maps in DeepGlobe dtaset [29] using ResNet and VGG16 backbones but their implementation was not efficient enough and our model can produce better quality segmentation than their model.
Other methods [30,31] employed depth information to support the semantic segmentation task. Kang et al. [30] proposed a depth adaptive deep neural network for semantic segmentation using a depth-adaptive multiscale convolutional layer consisting of the adaptive perception neuron and the in-layer multiscale neuron to adjust the receptive field at each spatial location and to apply the different size of the receptive field at each feature space to learn features at multiple scales, respectively. Gu et al. [31] proposed Hard Pixel Mining for semantic segmentation using a multiscale loss weight map generated by the depth data to enforce the model to pay more attention to the hard pixels in segmentation. They employed the depth data during the training step only and did not use it in the testing step. Other studies have shown the ability of CNN models to perform the joint task of semantic segmentation and depth estimation. For example, Mousavian et al. [32] proposed a multi-scale fully convolutional CNN for simultaneous semantic segmentation and monocular depth estimation. In this architecture, a CNN model is coupled with a fully connected conditional random field (CRF) to obtain the contextual relation and the interactions between the semantics of the image and depth cues. Zhang et al. [33] developed joint task-recursive learning (TRL) for semantic segmentation and depth estimation and showed that TRL can recursively refine the results from both tasks using a task attention module. Further, a hybrid CNN for depth estimation and semantic segmentation (HybridNet) was developed by Lin et al. [34]; this network functions by sharing the parameters that can yield mutual improvement for each task. Finally, He et al. [35] proposed semantic object segmentation and depth estimation (SOSD)-Net for application to monocular images; in this approach, semantic objectness is used to perform image processing based on the geometric relationship between the two tasks. The aligned Xception architecture is employed.
In all of the related work, the studies adopted mainly encoder-decoder approaches with some methods adopted extra attention methods or transformers. The proposed method mainly eliminates the need for a complex decoding stage which is computationally expensive, eliminates some of the image details, and propagates noise. Hence, we propose a simple and fast approach for dense prediction. We use the DTS-layer as the decoding stage to directly construct a dense map out of small feature maps extracted from an encoding stage.

3. Proposed Method

The proposed method, which is illustrated in Figure 2, consists of three main components:
  • The feature extractor CNNs: These CNNs extract the image semantics and deliver deep feature maps. Both proposed architectures, DTS-Net and DTS-Net-Lite, are presented in this section. We also discuss our reasons for choosing depth-wise separable convolution-based architectures.
  • DTS layer: This layer aggregates the small feature maps to form a higher-resolution segmentation map (or depth map).
  • The nearest label filtration (NLF): This filtration employs multi-label classification labels to filter the segmentation mask patch by patch.

3.1. Feature Extractor Architectures

The feature extractor is the most important part of a dense prediction CNN model, as the dense predictions are highly dependent on the extracted features. Depth-wise separable convolution was first presented by Chollet [14] and was clearly shown to be much faster and more efficient than normal convolution. Depth-wise separable convolution-based architectures (such as Xception and MobileNetV2 [36]) can attain high accuracy for ImageNet classification and other tasks (such as object detection and semantic segmentation) with a significantly low parameter number and low floating-point operations count compared with conventional CNN-based architectures.
Xception [14] was also proposed by Chollet. This unique convolution approach consists of two main modules: depth-wise and point-wise convolution. In depth-wise convolution, the convolution is performed for each channel separately, whereas point-wise convolution involves a 1 × 1 convolution through the channels; the latter can be regarded as channel projection or dot-product multiplication between channels. As regards the computational cost of depth-wise separable convolution [36], it is clearly apparent that this convolution has 1 N + 1 D k 2 fewer parameters than conventional convolution, where N is the number of channels and D k is the kernel size.
The original Xception architecture proposed for Imagenet classification consists of three stages. The main building block of the model is the Xception block, which involves two or three depth-wise separable convolutions with a kernel size of 3 × 3 and rectified linear unit (ReLU) activation with or without max-pooling at the end. Residual connections are used to connect the block from start to end. The first stage involves the entry flow, which has three Xception blocks with max-pooling at the end of each block. The second stage involves the middle flow, which has eight Xception blocks without max-pooling. Finally, the third stage pertains to the exit flow, which has one Xception block with max-pooling and then two depth-wise separable convolutions. The Xception block details are shown in Figure 3a. The final feature map depth is 2048. In our proposed approach, we then add a 1 × 1 × 1024 (i.e., 32 × 32) convolution to reduce the depth from 2048 to 1024, so as to aggregate a segmentation map that is 32 times the width and 32 times the height of the final feature maps after the DTS layer. As a second branch, we add global average pooling and then a fully connected layer with sigmoid activation to obtain the multi-label classification predictions. The Xception feature extractor is incorporated in the main network proposed in this work, that is, DTS-Net.
MobileNetV2 [36] is the other architecture employed in this study and is implemented in DTS-Net-Lite to accelerate our approach for real-time application on low-computational-power devices such as embedded devices or mobile phones. Like Xception, MobileNetV2 adopts depth-wise separable convolution as the main building module, to benefit from the cheap computational cost of this module along with other features that further decrease the computation. In MobileNetV2, inverted residuals and linear bottlenecks are implemented. The feature maps can be encoded in a low-dimensional subspace using the bottlenecks with a linear operation. A bottleneck consists of an expansion module (1 × 1 convolution with a higher number of filters with expansion factor t), followed by depth-wise separable convolution with a 3 × 3 kernel and, finally, a projection module (1 × 1 convolution with a lower number of filters with the same factor t). In [36], t = 6 was employed as the expansion factor. Each of the three modules features patch normalization and ReLU6 activation (i.e., ReLU with the maximum value clipped to 6). The inverted residuals form a skip-connection between the bottleneck start and end, which have the same channel depth. This gives the network access to earlier features before entering the bottleneck block. MobileNetV2 originally consisted of seven sequential MobileNet blocks (a MobileNetV2 block is shown in detail in Figure 3b), followed by a 1 × 1 × 1280 convolution. In our proposed approach, we add a 1 × 1 × 1024 convolution to reduce the depth to 1024 (32 × 32) in order to construct a dense map using the DTS layer. We also add a global average pooling layer after the 1 × 1 × 1280 convolution, which is followed by a fully-connected layer with sigmoid activation for multi-label classification. As noted above, we employ MobileNetV2 as the feature extractor of the lightweight version of the proposed method, that is, DTS-Net-Lite.
We believe that our choice of depth-wise separable convolution-based architecture is key to the high accuracy and speed achieved in this work. The depth-wise convolution (DW-Conv) is perfectly compatible with depth-to-space image construction as DW-Conv extracts information from each channel separately, allowing small details in each channel (depth-axis) to be captured. Then, the DTS layer reconstructs the image from the aggregation of the information in each channel, as discussed in detail in the next section. This explains the high accuracy of the depth-wise separable convolution-based architecture combined with the DTS dense map reconstruction.

3.2. Efficient Sub-Pixel Convolutional Neural Networks

Efficient sub-pixel CNN [1] was originally proposed for image and video super-resolution, in which high-resolution images are constructed from lower-resolution images through DTS aggregation. In this approach, pixels with the same location in the low-resolution image channels are mapped and re-ordered to form a super-pixel in the high-resolution image. DTS aggregates small feature maps of dimension h × w × r 2 to form a final high-resolution dense map of size r h × r w (as a segmentation or depth map), where h, w, and r 2 are the feature-map width, height, and depth, respectively. The feature maps extracted from the feature extractor CNN are passed to the 1 × 1 × r 2 convolutional layer, which gives a total activation patterns of r 2 with size W r and H r , where W and H are the width and height of the input image, respectively. The final reconstructed dense map is also produced. The DTS module then rearranges the elements of the tensor W r × H r × r 2 to a tensor of shape r . W r × r . H r , which is equivalent to the input image size of W × H. This operation is illustrated visually in Figure 2. In this method, fast one-step up-sampling is performed, which can be mathematically expressed as
M A P x , y = F [ x / r ] , [ y / r ] , r . m o d ( y , r ) + m o d ( x , r ) ,
where M A P x , y is the constructed dense map, F represents the feature maps extracted from the CNN and passed to the 1 × 1 × r 2 convolutional layer, r is the feature map depth, and operation m o d is the modulus, which maps the pixels from the low-resolution feature maps to the super-pixel in the final dense map when m o d ( x , r ) = 0 or m o d ( y , r ) = 0 .
In our implementation, we use r = 32 , as the Xception or MobileNetV2 feature extractors down-sample the features using the convolution and max-pooling to 1/32 of the input image size. Therefore, to reconstruct the final dense map at the same size as the input image, we use the 1 × 1 × 1024 convolutional layer, where r 2 = 1024 , to obtain a final feature map count equal to r 2 . This reconstruction layer can be trained simply with a regression loss (we use the mean absolute error), while inference (a rounding operation) is used to round the predicted pixels in the dense layer to the nearest integer in the case of semantic segmentation.

3.3. Nearest Label Filtration

As we use an image reconstruction technique, the reconstructed segmentation map introduces a reconstruction noise, particularly at the borders of the objects in the image. Hence, mislabeled pixels are introduced. In detail, these noisy pixels are usually mislabeled with a label neighboring the true label. During the training stage, the network attempts to learn the true label as accurately as possible; however, an abrupt color change occurs at a border, causing mislearning of the border pixels. This problem is particularly notable when the CNN down-samples the image using max-pooling, and some details disappear from the final feature maps. To solve this problem, we propose Nearest Label Filteration (NLF), which aims to correct those noisy pixels by mapping them to the nearest classification label using a patch-by-patch technique. We train the network for multi-label classification besides segmentation; thus, we use the classification labels of the image (global labels) to filter the noisy pixels in the segmentation map. We divide the image into patches (imperially, 10 × 10 and 8 × 8 are the best patch sizes). For each patch, we obtain the unique labels that exist in the patch (patch labels). Then, we find the intersection between the patch labels ( P L ) and global labels ( G L ) of the image to obtain the true labels ( T L ):
T L = G L P L
By measuring the distance between each P L and the T L , we can find the distance array for all P L as follows:
d i s t _ a r r = | P L i T L j |
for i in Ω P L space and j in Ω G L space. Then we can estimate the nearest label in T L corresponding to the minimum distance in d i s t _ a r r and overlay the noisy pixel with the nearest label ( N L ), as follows:
N L i = T L [ a r g m i n ( d i s t _ a r r ) ]
This filtration method efficiently removes the noise from the image and provides clear edges. However, performance degradation occurs in extremely dense cases where the segmentation mask has a large number of classes and almost all pixels have widely differing labels (e.g., for the Cityscapes [4] dataset, in which the scenes are very crowded with small labels and many different classes). However, the method performs very well on the Pascal VOC2012 and NYUV2 datasets, in general. Figure 1 presents sample images that clearly show the difference before and after filtration. Figure 2 presents a visual illustration of the filtration step as a component of the overall proposed method.

3.4. Loss Function

The loss function for the proposed CNNs is divided into two separate loss functions: one for learning the semantic segmentation mask reconstruction and the other for multi-label classification. The loss for semantic segmentation reconstruction is the mean absolute error (MAE), which is expressed as follows:
S e g L o s s = 1 r 2 h w x = 1 r w y = 1 r h | S x , y G T S x , y C o n s t | ,
where S x , y G T is the ground-truth pixel in the semantic segmentation mask and S x , y C o n s t is the pixel in the constructed semantic segmentation mask. The other loss is the multi-label binary cross-entropy (BCE) loss for N-classes of the dataset. This classification loss can be defined as
C l a s s L o s s = i = 1 N [ p i l o g ( q i ) + ( 1 p i ) l o g ( 1 q i ) ] ,
where p is the ground-truth label probability (0 or 1) and q is the predicted label probability. The multi-label binary cross-entropy loss is used to predict the objectness of each class independently with sigmoid activation which outputs a value in the range of 0 to 1 equivalent to the probability of the existence of each object. The employment of MAE for segmentation map regression and the multi-label classification for mask filtration improves the speed, it lowers the number of computations needed for obtaining the segmentation map as it constructs the segmentation map directly like an image with a single channel instead of predicting N channels image (where N is the number of classes) which is computationally expensive since the final segmentation maps have the same size of the input image which is relatively a high-resolution image ( 1024 × 512 in case of cityscapes). Additionally, mask filtration using multi-label classification (NLF) is much cheaper than constructing the N segmentation masks.

4. Experiments

In this section, we report the experiments performed to test the performance of the proposed method. These experiments include a segmentation quality comparison between different CNN backbones as feature extractors. We show the intermediate feature maps used to construct the segmentation maps. We also report semantic segmentation results for our proposed network DTS-Net and DTS-Net-Lite, and compare these results with those for state-of-the-art (SOTA) methods. Finally, we show that our method can be used for the joint task of semantic segmentation and depth estimation, as these two tasks are highly similar. The same training and testing environment and conditions are used for all experiments.

4.1. Semantic Segmentation Evaluation Metrics

To evaluate the proposed models on semantic segmentation, we use two evaluation metrics: mean intersection over union (mIOU ) and mean pixel accuracy (Pix.acc. ). The definition for those metrics is stated as below:
m I O U = 1 N c = 1 N T P c T P c + F P c + F N c
P i x . a c c . = 1 N c = 1 N T P c + T N c T P c + T N c + F P c + F N c ,
where N is the number of class categories, T P (true-positive) represents the pixels that are truly predicted to be belonging to a class (c ), T N (true-negative) represents the pixels that are predicted to not belong to that class, F P (false-positive) the pixels which are falsely predicted to be belonging to a class, and F N (false-negative) the pixels which are falsely predicted to be not belonging to that class.

4.2. DTS-Net and DTS-Net-Lite

DTS-Net is the proposed deep architecture for semantic segmentation and employs Xception as the feature extractor CNN. In the experiment, we used relatively high-resolution images to obtain a segmentation mask with a fine border and high mean intersection over union. DTS-Net-Lite is the fast, lightweight version of the proposed method and employs MobileNetV2 as the feature extractor. In the experiment, we used approximately half the image size for DTS-Net to gain more processing speed; thus, some accuracy was sacrificed.

4.3. Benchmarks

To train and test our approach, we used three popular benchmarks for semantic segmentation, PASCAL VOC2012, NYUV2 Depth, and Cityscapes. Two of these benchmarks contain depth data.
PASCAL VOC2012 [2] is a popular benchmark with 20 commonly seen classes. The dataset contains annotation for several computer vision tasks such as bounding-box object detection, semantic segmentation, instance segmentation, and action recognition. The segmentation data contain 2913 images of varying sizes: 1464 for training and 1449 for validation, the definition of the classes and the color used for each class’ mask is shown in Figure 4a. We resized the images to 480 × 480 and 256 × 256 for DTS-NET and DTS-Net-Lite training and testing, respectively. We trained the network as detailed in the previous section and constructed the segmentation map with the DTS layer at the same size as the image. This was achieved using feature map sizes of 15 × 15 and 8 × 8 , both with a depth of 1024, for DTS-Net and DTS-Net-Lite, respectively (both architectures down-sampled the image to 1/32nd the input image size). In testing both networks, we used the same image size as in training.
NYUV2 Depth [3] is a benchmark for semantic segmentation and depth estimation. The image data are in RGB-D format and captured using a Kinect sensor. The dataset contains 1449 images of 13 indoor object classes, that is, 795 and 654 images for training and testing, respectively, with fixed image sizes, the definition of the classes, and the color used for each class’ mask is shown in Figure 4b. We trained and tested DTS-Net with the original image size ( 640 × 480 ) and used a 320 × 256 image size for DTS-Net-Lite. We trained both networks as described in Section 3, with feature map sizes before the DTS layer of 20 × 15 and 10 × 8 , both with a depth of 1024, for DTS-Net and DTS-Net-Lite, respectively.
Cityscapes [4] is another benchmark for semantic segmentation and depth estimation, which contains urban street scenes in 50 cities, organized as 19 classes grouped into 8 categories. The dataset contains 5000 training images with fine segmentation masks, an additional 20,000 coarse annotated images, and 500 validation images, the definition of the classes and the color used for each class’ mask is shown in Figure 4c. We trained our models with the 5000 fine annotated images as we required high-resolution segmentation and depth estimation with fine borders. The original size of the images in the dataset was 2048 × 1024 . We resized the training and testing images to half this size, that is, to 1024 × 512 . This was still a high resolution for DTS-Net but allowed considerably faster training and inference. We then validated the model with the validation data. For DTS-Net-Lite, we resized the images to approximately one-third of the original size, that is, to 704 × 352 , for higher-speed processing while sacrificing some accuracy. The feature map sizes before the DTS layer were 32 × 16 and 22 × 11 with a depth of 1024 for DTS-Net and DTS-Net-Lite, respectively.

4.4. Training and Testing Configuration

We trained the models using a desktop PC with an Intel Core i7-8700 CPU @3.20 GHz, Nvidia RTX3090 GPU, and 64 GB RAM. We used Tensorflow-Keras to implement the proposed method and all of the proposed models were initialized with the pre-trained version of the model on the ImageNet [22] dataset to accelerate the model fitting. We used Adam’s optimizer with an initial learning rate of 0.001, beta1 of 0.9, and beta2 of 0.999. We then evaluated the models using the same hardware configurations. All models were trained for approximately 1000 epochs.

4.5. Quality Comparison between Different CNN Backbones

We performed several experiments to test the ability of different CNN architectures to perform semantic segmentation using the proposed method. We trained three other architectures as feature extractors (VGG16 [10], ResNet50 [11], and DenseNet121 [37]) in addition to Xception and MobileNetV2. All were trained on the NYUV2 segmentation dataset with an image size of 640 × 480 for 1000 epochs. We then compared the features and the constructed segmentation masks obtained using the DTS layer (a sample test image is shown in Figure 5). For VGG16 and DenseNet121, the results showed features with inaccurate blurry edges, which were also apparent in the constructed segmentation masks. Better features and constructed masks with far more detail of the object edges and boundaries were extracted in the case of ResNet50 and MobileNetV2. The highest-detail features and the best-constructed segmentation maps were obtained for the Xception architecture. Therefore, we adopted this architecture as the feature extractor of our method. We adopted MobileNetv2 as the feature extractor for DTS-Net-Lite because that architecture exhibited average performance but the fewest multiplication/addition computations (MACs) (0.3 Giga; far lower than those of the other architectures), to achieve higher-speed processing. Table 1 compares the five architectures based on the mean pixel accuracy (Pix.acc. ) of each CNN with the DTS layer. This is a measure of the percentage of the correctly classified pixels in the predicted mask. The MAC and parameter numbers are also compared. To further support our selection and the proposed architecture, we trained DTS-Net on DeepGlobe satellite images dataset for road extraction using binary segmentation. We compared the obtained segmentation results from our architectures (using Xception and MobileNetV2 backbones) with ResNet-D2S and VGG16-D2S proposed in [28] (we also used Segnet results on DeepGlobe as the authors of this paper [28] trained SegNet [6] on DeepGlobe and added the results in their paper) in Figure 6. The comparison is simply a quality comparison of samples from the DeepGlobe validation satellite images as the validation server for this challenge is closed and there is no other way to validate the results. Our results on the binary segmentation of the satellite images show better quality in extracting more details than the results produced by ResNet-50 [11] and VGG16 [10] backbones in [28], even better than SegNet [6], which proves the excellent performance of the proposed architectures using Xception and MobileNetV2 backbones.
We selected depth-wise separable convolution as being suitable for the proposed DTS module based on the following main reason. The DTS layer constructs the final image using a pixel reordering technique through the channel axis to form super-pixels in the segmentation map, which highly depend on the feature-map quality. Depth-wise convolution extracts the features from each channel separately and then projects them using point-wise convolution. This approach extracts more accurate feature maps as the depth-wise convolution focuses on the data in each channel separately. These characteristics explain the good quality of the features extracted by the Xception and MobileNetV2 feature extractors.

4.6. Joint Semantic Segmentation and Depth Estimation

We further trained the proposed method on the simultaneous task of semantic segmentation and depth estimation. NYUV2 provides dense depth data captured by a Kinect sensor that is suitable for direct use, and CITYSCAPES provides disparity data that are equivalent to depth (as the disparity has a linear relationship with the depth). Here, we trained DTS-Net on the disparity without conversion. We trained our main model, DTS-Net, on NYUV2 with an image size of 640 × 480 and on Cityscape with an image size of 1024 × 512 for the depth estimation and semantic segmentation tasks. For the depth estimation, we added a third branch, which was identical to the branch for the segmentation branch, that is, a 1 × 1 × 1024 convolutional layer followed by a DTS layer to construct the depth map. We retained the two main branches, that is, the semantic segmentation branch and the multi-label classification branch. We then trained the network for 1000 epochs under the same hardware configuration, as previously mentioned in Section 4.4. As the depth estimation and semantic segmentation have a common training procedure and the aim is a dense map, we believe that the two tasks help each other. This is confirmed in Section 5, where the accuracy and error values for the depth are reported; these values were obtained simultaneously with the performance of the semantic segmentation task.

5. Results

In this section, we report the noteworthy results obtained with the proposed method. These results indicate the high accuracy of both our semantic segmentation approach and the joint task with depth estimation. We compare our results with those of SOTA methods for semantic segmentation and show that our method outperforms the alternatives by some margin. We also compare the joint task model performance with those of SOTA joint methods.

5.1. Semantic Segmentation mIOU Results

Promising mIOU values were exhibited by the proposed method on the three datasets, particularly with the application of the NLF technique. Figure 7 shows sample results from the three datasets with the predicted and filtered (NLF Seg) segmentation results. On the PASCAL VOC2012 benchmark, DTS-Net and DTS-Net-Lite attained mIOU values of 91.1% and 83.8% on the validation set, respectively. On the NYUV2 test benchmark, DTS-Net attained high accuracy with a pixel accuracy (Pix.acc.) of 80.72%, while DTS-Net-Lite attained a pixel accuracy (Pix.acc.) of 73.7%. On the CITYSCAPES validation benchmark, DTS-Net and DTS-Net-Lite attained mIOU values of 80.72% and 61.5%, respectively. These mIOU and Pix.acc. values were superior to those of all SOTA methods for semantic segmentation, even though the proposed architecture is far simpler than the SOTA-method architectures.
The results in Figure 7 and Figure 8 show the quality of the segmentation results, with fine and sharp edges. Table 2 lists the mIOU values for both DTS-Net and DTS-Net-Lite on the three benchmarks; these results highlight the unique performance of the proposed method. Note that the predicted segmentation masks contained many mispredicted pixel values that had neighboring labels to the true label. Most of those pixels could be corrected easily by the NLF technique. The main reason for the high mIOU and Pix.acc. values obtained by our method is the NLF algorithm which improves the mIOU by 63.4% on PASCAL VOC2012, improves the mIOU by 28.8% on NYUV2, and improves the mIOU by 33.0% on CITYSCAPES. The pixel accuracy (Pix acc.) values also improved when applying NLF but not with a big margin due to the nature of the pixel accuracy metric that reports the number of correctly classified pixels regardless of the ratio of the overlap between the predicted and the ground truth masks which is used in the mIOU metric. Those high improvements in the mIOU and Pix acc. prove the important role of the NLF algorithm in getting more accurate segmentation results than those without NLF.

5.2. Semantic Segmentation Speed Results

We measured the processing speed of each frame for both models. Recall that we chose MobileNetV2 as the feature extractor for DTS-Net-Lite to achieve real-time processing. In general, depth-wise separable convolution-based architecture has considerably higher processing speed than normal convolutional architectures, as discussed in detail in Section 3.1. DTS-Net-Lite is faster than DTS-Net because inverted residual and linear bottle-necks are used; however, DTS-Net can achieve an average processing speed. In the experiments, DTS-Net-Lite attained a speed of 19.2 frames per second (fps) with NLF filtration on PASCAL VOC2012 as the image size was 256 × 256 . Without the NLF, it attained 24 fps, as the patch-by-patch filtration required more computations, which consumed more time. DTS-Net with NLF achieved only 17.5 fps on PASCAL VOC2012 and 22 fps without NLF. DTS-Net-Lite achieved speeds of 14.1 and 21.3 fps with and without NLF, respectively, on the NYUV2 dataset. On the CITYSCAPES dataset, DTS-Net-Lite achieved a low speed of 11.7 fps with NLF, as the image size was 512 × 256 , and 14.7 fps without NLF. Table 2 lists the results achieved with DTS-Net and DTS-Net-Lite, that is, the mIOU, the Pix.acc, and the speed.

5.3. Comparison with SOTA Semantic Segmentation Methods

We compared the proposed method with SOTA methods in terms of semantic segmentation for the three benchmarks. On PASCAL VOC2012, the proposed method outperformed the recent SOTA methods in terms of mIOU. In detail, the best-performing SOTA method, that is, EfficientNet-L2+NAS-FPN [21] (based on EfficientNetB7, neural architecture search, and a features pyramid network [38]) achieved an mIOU of 90.0%, whereas DTS-Net achieved 91.1%. Table 3 compares the proposed method and the SOTA methods on the PASCAL VOC2012 validation set in terms of mIOU, and indicates the backbone architecture of each method. Further, Table 4 compares the proposed method with the SOTA methods with regard to semantic segmentation on the NYUV2 benchmark in terms of Pix.acc. Our method DTS-NEt+NLF outperformed the best SOTA method on NYUV2 13 classes (BPNET [39]) by 11.2%, and DTS-NET-Lite+NLF achieves approximately the same performance as BPNET with 0.2% lower pixel accuracy. We also compared our method with the SOTA methods for semantic segmentation on the CITYSCAPES validation benchmark. As detailed in Table 5, the proposed method outperformed the best SOTA method in terms of mIOU by 0.15%. These results prove that the proposed method as a learning technique exhibits higher performance than the recent SOTA encoder-decoder architectures i.e., Deeplabv3, DeepLabv3+ which also depends on depthwise separable convolution architectures (Xception and MobileNet encoders). We also compared the speed of each model (the speeds are copied from the original papers) in Table 5 while the speed comparison is not so informative since the hardware used for testing and the input image size highly affect the speed of the model, and each method uses different hardware specifications and image size. In general, the proposed method exhibited distinctive performance primarily because of the DTS layer which constructs the initial segmentation result and, secondarily, the NLF filtration which enhances this segmentation result. Figure 9 shows a quality comparison between the semantic segmentation results obtained from the best architecture of the proposed method (DTS-Net+NLF) and those of the SOTA methods which have encoder-decoder architecture and some of them have Xception architecture encoder, that is, DeepLabv3 and DeepLabv3+, on PASCAL VOC2012 and CITYSCAPES benchmarks. In Figure 9, the comparison on PASCAL VOC2012 shows that our method can predict a precise segmentation mask even the unlabeled areas (white pixels in the groundtruth) are labeled as background, noting that the borders of our results have few mislabeled pixels. Those pixels exist due to the use of an image reconstruction method (DTS) to construct the segmentation mask, and the NLF algorithm sometimes fails to correct those pixels as their label is a true label of the segmentation output. On a CITYSCAPES quality comparison in Figure 9, our methods can predict high-quality masks with few noisy borders due to the same reasons mentioned at the previous sentence. Note that the black regions contain many mislabeled pixels and this situation is acceptable as the black pixels are unlabeled pixels that are not counted in the segmentation evaluation.

5.4. DTS-Net for Joint Depth Estimation and Semantic Segmentation

We trained DTS-Net on the simultaneous joint task of depth estimation and semantic segmentation, as discussed in Section 4.5. The proposed method yielded promising results even though the network performed both tasks simultaneously. The output semantic segmentation and depth estimation had fine details. DTS-Net extracted the important features for each task while using the same feature extractor but with different branches of 1 × 1 × 1024 convolutional layers, which were followed by the DTS layer. NLF filtration was then applied to the segmentation output. The network showed promising error values and accuracy for the depth estimation task while attaining a slightly inferior semantic segmentation performance. To evaluate the depth estimation, we evaluated the following depth estimation metrics:
  • the absolute relative error (REL): 1 n i n | y y ^ | y ;
  • the relative difference squared (Sq_REL): 1 n i n | | y y ^ | | 2 y ;
  • the root mean squared error (RMSE): 1 n i n ( y y ^ ) 2 ;
  • the disparity error (px): 1 n i n | y y ^ | ;
  • the threshold accuracy δ i of y p : m a x ( y y ^ , y ^ y ) = δ < t h r for the commonly used threshold values t h r = 1.25 , 1.25 2 , 1.25 3 ,
where n is the number of pixels in the depth map, i indicates an iterator over the pixels, y and y ^ are the ground truth and predicted pixel values, respectively.
The joint trained network yielded high-quality results, as shown in Figure 10. DTS-Net+NLF attained a REL of 0.0987 and δ 1 of 90.5% on the CITYSCAPES benchmark. Regarding the semantic segmentation in the joint task, the network attained mIOU values of 72.1% with NLF filtration which is much lower than this for the single task. Further, DTS-Net+NLF attained a REL value of 0.102 and δ 1 of 94.59% while attaining an mIOU of 49.1% on the NYUV2 benchmark. The proposed method outperformed the SOTA methods for joint semantic segmentation and depth estimation on the depth estimation by some margin, however, the semantic segmentation was not compared because the segmentation configuration for the models in the comparison is not similar to ours. Table 6 presents the comparison for the depth estimation results with the SOTA methods in terms of REL, Sq Rel, RMSE, the disparity error (px), and the delta accuracies. The proposed method had low depth errors and high delta accuracy for depth estimation surpassing all other joint semantic segmentation and depth estimation accuracies and error values while maintaining a high mIOU for semantic segmentation.

6. Future Work

The proposed DTS-Net shows promising results for dense prediction tasks, such as semantic segmentation and depth estimation. In the future, this method can be extended to more challenging dense prediction tasks, such as instance segmentation, through combination with an object detection technique. This approach will allow for discrimination between instances of the same class and can mask those instances with different masks. The feature extractor component could also be replaced with a considerably smaller and simple architecture equivalent to the opposite layer (the space-to-depth), which could then be followed by a small number of convolutional filters. This would reduce the complexity of the feature extractor CNNs. The nearest label filtration (NLF) technique can also be improved using a separate CNN trained on small patches of the training images instead of the integrated branch for classification in current implementation for whole image classification, and the local prediction may attain better filtration accuracy as the current NLF considers the global labels of the image, and not all of those labels exist in every patch of the image.

7. Conclusions

The proposed DTS-Net is a simple approach inspired by the DTS image reconstruction technique used in image and video super-resolution. The proposed DTS-Net implementation is simple and yet achieves extremely high accuracy and an average speed for dense prediction tasks such as semantic segmentation, as proven by the results obtained on the different semantic segmentation benchmarks reported herein. For the simultaneous joint task of semantic segmentation and depth estimation, DTS-Net exhibited promising accuracy for both semantic segmentation and depth estimation while the two tasks were trained simultaneously. These results prove the strength of the DTS module, which is the core of the proposed method, and demonstrate its potential for re-employment in dense prediction tasks. The proposed DTS-Net-Lite also provides the ability to perform real-time processing by using a lightweight encoder achieving relatively good segmentation accuracy.

Author Contributions

Conceptualization, H.I., A.S.; methodology, H.I., A.S.; software, H.I.; formal analysis, H.I.; investigation, H.-S.K.; resources, H.-S.K.; data curation, H.I.; writing—original draft preparation, H.I.; writing—review and editing, H.I., H.-S.K.; validation, H.I., H.-S.K.; visualization, H.-S.K.; supervision, H.-S.K.; project administration, H.-S.K.; funding acquisition, H.-S.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education under Grant 2020R1I1A3A04037680 and in part conducted as a part of the research project of “Development of IoT Infrastructure Technology for Smart Port” financially supported by the Ministry of Oceans and Fisheries.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasests used in this paper are public datasets. We also provide the test and the evaluation codes of the proposed method at: https://github.com/HatemHosam/DTS-Net which is created and accessed on 21 June 2021.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-Time Single Image and Video Super-Resolution Using an Efficient Sub-Pixel Convolutional Neural Network. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar] [CrossRef] [Green Version]
  2. Everingham, M.; Gool, L.V.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar]
  3. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor Segmentation and Support Inference from RGBD Images. In Proceedings of the European Conference on Computer Vision (ECCV) 2012, Florence, Italy, 7–13 October 2012. [Google Scholar]
  4. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The Cityscapes Dataset for Semantic Urban Scene Understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  5. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar] [CrossRef] [Green Version]
  6. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  7. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  8. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected crfs. In Proceedings of the ICLR, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  9. Chen, L.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834848. [Google Scholar] [CrossRef]
  10. Liu, S.; Deng, W. Very deep convolutional neural network based image classification using small training sample size. In Proceedings of the 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), Kuala Lumpur, Malaysia, 3–6 November 2015; pp. 730–734. [Google Scholar] [CrossRef]
  11. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  12. Chen, L.-C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  13. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  14. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1800–1807. [Google Scholar] [CrossRef] [Green Version]
  15. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6230–6239. [Google Scholar] [CrossRef] [Green Version]
  16. Zhao, H.; Qi, X.; Shen, X.; Shi, J.; Jia, J. ICNet for Real-Time Semantic Segmentation on High-Resolution Images. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 405–420. [Google Scholar]
  17. Yu, C.; Wang, J.; Peng, C.; Gao, C.; Yu, G.; Sang, N. BiSeNet: Bilateral Segmentation Network for Real-time Semantic Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 325–341. [Google Scholar]
  18. Zhang, H.; Wu, C.; Zhang, Z.; Zhu, Y.; Zhang, Z.; Lin, H.; Sun, Y.; He, T.; Muller, J.; Manmatha, R.; et al. Resnest: Splitattention networks. arXiv 2020, arXiv:2004.08955. [Google Scholar]
  19. Shi, H.; Li, H.; Meng, F.; Wu, Q.; Xu, L.; Ngan, K.N. Hierarchical Parsing Net: Semantic Scene Parsing From Global Scene to Objects. IEEE Trans. Multimedia 2018, 20, 2670–2682. [Google Scholar] [CrossRef]
  20. Chen, T.; Xie, G.; Yao, Y.; Wang, Q.; Shen, F.; Tang, Z.; Zhang, J. Semantically Meaningful Class Prototype Learning for One-Shot Image Segmentation. IEEE Trans. Multimedia 2021. [Google Scholar] [CrossRef]
  21. Zoph, B.; Ghiasi, G.; Lin, T.-Y.; Cui, Y.; Liu, H.; Cubuk, E.D.; Le, Q.V. Rethinking pre-training and self-training. arXiv 2020, arXiv:2006.06882. [Google Scholar]
  22. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef] [Green Version]
  23. Tan, M.; Le, Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 6105–6114. [Google Scholar]
  24. Rashwan, A.; Du, X.; Yin, X.; Li, J. Dilated spinenet for semantic segmentation. arXiv 2021, arXiv:2103.12270. [Google Scholar]
  25. Bai, S.; Kolter, J.Z.; Koltun, V. Deep Equilibrium Models. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Vancouver, BC, Canada, 8–14 December 2019; Volume 32. [Google Scholar]
  26. Termritthikun, C.; Jamtsho, Y.; Ieamsaard, J.; Muneesawang, P.; Lee, I. EEEA-Net: An Early Exit Evolutionary Neural Architecture Search. Eng. Appl. Artif. Intell. 2021, 104, 104397. [Google Scholar]
  27. Ding, X.; Zhang, X.; Ma, N.; Han, J.; Ding, G.; Sun, J. RepVGG: Making VGG-Style ConvNets Great Again. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13733–13742. [Google Scholar]
  28. Aich, S.; van der Kamp, W.; Stavness, I. Semantic Binary Segmentation Using Convolutional Networks without Decoders. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 182–1824. [Google Scholar] [CrossRef] [Green Version]
  29. Demir, I.; Koperski, K.; Lindenbaum, D.; Pang, G.; Huang, J.; Basu, S.; Hughes, F.; Tuia, D.; Raskar, R. Deepglobe 2018: A challenge to parse the earth through satellite images. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  30. Kang, B.; Lee, Y.; Nguyen, T.Q. Depth-Adaptive Deep Neural Network for Semantic Segmentation. IEEE Trans. Multimedia 2018, 20, 2478–2490. [Google Scholar] [CrossRef] [Green Version]
  31. Gu, Z.; Niu, L.; Zhao, H.; Zhang, L. Hard Pixel Mining for Depth Privileged Semantic Segmentation. IEEE Trans. Multimedia 2020, 23, 3738–3751. [Google Scholar] [CrossRef]
  32. Mousavian, A.; Pirsiavash, H.; Košecká, J. Joint Semantic Segmentation and Depth Estimation with Deep Convolutional Networks. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 611–619. [Google Scholar] [CrossRef] [Green Version]
  33. Zhang, Z.; Cui, Z.; Xu, C.; Jie, Z.; Li, X.; Yang, J. Joint Task-Recursive Learning for Semantic Segmentation and Depth Estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 235–251. [Google Scholar]
  34. Lin, X.; Sánchez-Escobedo, D.; Casas, J.R.; Pardàs, M. Depth Estimation and Semantic Segmentation from a Single RGB Image Using a Hybrid Convolutional Neural Network. Sensors 2019, 19, 1795. [Google Scholar] [CrossRef] [Green Version]
  35. He, L.; Lu, J.; Wang, G.; Song, S.; Zhou, J. SOSD-Net: Joint semantic object segmentation and depth estimation from monocular images. Neurocomputing 2021, 440, 251–263. [Google Scholar]
  36. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef] [Green Version]
  37. Huang, G.; Liu, Z.; Maaten, L.V.D.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef] [Green Version]
  38. Lin, T.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar] [CrossRef] [Green Version]
  39. Hu, W.; Zhao, H.; Jiang, L.; Jia, J.; Wong, T.-T. Bidirectional Projection Network for Cross Dimension Scene Understanding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14373–14382. [Google Scholar]
  40. Gao, S.-H.; Cheng, M.-M.; Zhao, K.; Zhang, X.-Y.; Yang, M.-H.; Torr, P.H. Res2Net: A New Multi-Scale Backbone Architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 652–662. [Google Scholar] [CrossRef] [Green Version]
  41. Zhang, Z.; Zhang, X.; Peng, C.; Xue, X.; Sun, J. ExFuse: Enhancing Feature Fusion for Semantic Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 269–284. [Google Scholar]
  42. Ghiasi, G.; Cui, Y.; Srinivas, A.; Qian, R.; Lin, T.-Y.; Cubuk, E.D.; Le, Q.V.; Zoph, B. Simple Copy-Paste Is a Strong Data Augmentation Method for Instance Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 2918–2928. [Google Scholar]
  43. Handa, A.; Patraucean, V.; Badrinarayanan, V.; Stent, S.; Cipolla, R. Understanding Real World Indoor Scenes with Synthetic Data. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 4077–4085. [Google Scholar]
  44. Hermans, A.; Floros, G.; Leibe, B. Dense 3D semantic mapping of indoor scenes from RGB-D images. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 2631–2638. [Google Scholar]
  45. McCormac, J.; Handa, A.; Davison, A.; Leutenegger, S. Semanticfusion: Dense 3dsemantic mapping with convolutional neural networks. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 4628–4635. [Google Scholar]
  46. Dai, A.; Chang, A.X.; Savva, M.; Halber, M.; Funkhouser, T.; Niessner, M. Scannet: Richly annotated 3D reconstructions of indoor scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  47. Lin, M.; Chen, Q.; Yan, S. Network in Network. arXiv 2014, arXiv:1312.4400. [Google Scholar]
  48. Dai, A.; Nießner, M. 3DMV: Joint 3D-multi-view prediction for 3D semantic scene segmentation. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  49. Paszke, A.; Chaurasia, A.; Kim, S.; Culurciello, E. ENet: A Deep Neural Network Architecture for Real-Time Semantic Segmentation. In Proceedings of the International Conference on Learning Representations (ICLR), Toulon, France, 24–26 April 2017. [Google Scholar]
  50. Mehta, S.; Hajishirzi, H.; Rastegari, M. DiCENet: Dimension-wise Convolutions for Efficient Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 14, 1–10. [Google Scholar] [CrossRef]
  51. Poudel, R.P.K.; Bonde, U.; Liwicki, S.; Zach, C. ContextNet: Exploring Context and Detail for Semantic Segmentation in Real-time. In Proceedings of the British Machine Vision Conference BMVC 2018, Newcastle, UK, 3–6 September 2018. [Google Scholar]
  52. Nekrasov, V.; Shen, C.; Reid, I. Template-Based Automatic Search of Compact Semantic Segmentation Architectures. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Snowmass, CO, USA, 1–5 March 2020; pp. 1980–1989. [Google Scholar]
  53. Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning Transferable Architectures for Scalable Image Recognition. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8697–8710. [Google Scholar] [CrossRef] [Green Version]
  54. Chen, W.; Gong, X.; Liu, X.; Zhang, Q.; Li, Y.; Wang, Z. FasterSeg: Searching for Faster Real-time Semantic Segmentation. In Proceedings of the ICLR, Simien Mountains, Ethiopia, 27–30 April 2020. [Google Scholar]
  55. Shaw, A.; Hunter, D.; Landola, F.; Sidhu, S. SqueezeNAS: Fast Neural Architecture Search for Faster Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27–28 October 2019. [Google Scholar]
Figure 1. Sample predicted and filtered semantic segmentation results of proposed method: (top to bottom) samples from PASCAL VOC2012 [2], NYUV2 [3], and CITYSCAPES [4] benchmarks, respectively.
Figure 1. Sample predicted and filtered semantic segmentation results of proposed method: (top to bottom) samples from PASCAL VOC2012 [2], NYUV2 [3], and CITYSCAPES [4] benchmarks, respectively.
Sensors 22 00337 g001
Figure 2. The architecture of the proposed method. With an Xception (DTS-NET) or MobileNetV2 (DTS-Net-Lite) backbone, the DTS layer (module) is added after a 1 × 1 × r 2 convolutional layer. The pixels in the h × w × r 2 low-resolution feature maps are arranged in an r h × r w segmentation map. Another branch features global average pooling and a fully connected layer is added for multi-label classification. (a) Training architecture of the proposed method and (b) inference architecture, which features one additional step, that is, the nearest label filtration (NLF). NLF is a patch-by-patch image filtration based on the global labels of the image and the patch labels. The noisy pixels with labels that are not included in the global image labels are mapped to the nearest neighbor in the intersection between the patch and global labels using a simple iterative difference operation.
Figure 2. The architecture of the proposed method. With an Xception (DTS-NET) or MobileNetV2 (DTS-Net-Lite) backbone, the DTS layer (module) is added after a 1 × 1 × r 2 convolutional layer. The pixels in the h × w × r 2 low-resolution feature maps are arranged in an r h × r w segmentation map. Another branch features global average pooling and a fully connected layer is added for multi-label classification. (a) Training architecture of the proposed method and (b) inference architecture, which features one additional step, that is, the nearest label filtration (NLF). NLF is a patch-by-patch image filtration based on the global labels of the image and the patch labels. The noisy pixels with labels that are not included in the global image labels are mapped to the nearest neighbor in the intersection between the patch and global labels using a simple iterative difference operation.
Sensors 22 00337 g002
Figure 3. (a) Xception block. There are three depth-wise separable convolutions, each having prior ReLU activation. A residual connection connects the block start and end in the entry-flow stage. A 1 × 1 convolution is implemented in the middle-flow stage. (b) MobileNetV2 block. An expansion module is followed by a Depth-wise separable convolution layer, then a projection module was added with a residual connection between the beginning and the end of the block.
Figure 3. (a) Xception block. There are three depth-wise separable convolutions, each having prior ReLU activation. A residual connection connects the block start and end in the entry-flow stage. A 1 × 1 convolution is implemented in the middle-flow stage. (b) MobileNetV2 block. An expansion module is followed by a Depth-wise separable convolution layer, then a projection module was added with a residual connection between the beginning and the end of the block.
Sensors 22 00337 g003
Figure 4. Class definitions and colormaps for the used benchmarks. (a) PASCAL VOC2012 [2] class definitions and colormap, (b) NYUV2 [3] class definitions and colormap, and (c) Cityscapes [4] class definitions and colormap.
Figure 4. Class definitions and colormaps for the used benchmarks. (a) PASCAL VOC2012 [2] class definitions and colormap, (b) NYUV2 [3] class definitions and colormap, and (c) Cityscapes [4] class definitions and colormap.
Sensors 22 00337 g004
Figure 5. Quality comparison of (a) extracted feature maps before the DTS layer and (b) constructed segmentation mask using DTS layer (without NLF) between VGG16, ResNet50, DenseNet121, MobileNetV2, and Xception architectures.
Figure 5. Quality comparison of (a) extracted feature maps before the DTS layer and (b) constructed segmentation mask using DTS layer (without NLF) between VGG16, ResNet50, DenseNet121, MobileNetV2, and Xception architectures.
Sensors 22 00337 g005
Figure 6. Quality comparison of the binary segmentation task on DeepGlobe [29] dataset between the models proposed by Aich et al. [28] (ResNet-D2S and VGG16-D2S) and SegNet [6] in the same paper and our models: DTS-Net and DTS-Net-Lite which uses Xception and MobileNetV2 as backbones, respectively.
Figure 6. Quality comparison of the binary segmentation task on DeepGlobe [29] dataset between the models proposed by Aich et al. [28] (ResNet-D2S and VGG16-D2S) and SegNet [6] in the same paper and our models: DTS-Net and DTS-Net-Lite which uses Xception and MobileNetV2 as backbones, respectively.
Sensors 22 00337 g006
Figure 7. Sample results of predicted and NLF-filtered semantic segmentation (NLF Seg) results that were obtained using DTS-Net on PASCAL VOC2012 (rows 1–3), NYUV2 (rows 4–6), and CITYSCAPES (rows 7–9), with input images and ground-truth segmentation masks (Seg).
Figure 7. Sample results of predicted and NLF-filtered semantic segmentation (NLF Seg) results that were obtained using DTS-Net on PASCAL VOC2012 (rows 1–3), NYUV2 (rows 4–6), and CITYSCAPES (rows 7–9), with input images and ground-truth segmentation masks (Seg).
Sensors 22 00337 g007
Figure 8. Sample results of predicted and NLF-filtered semantic segmentation (NLF Seg) results that were obtained using DTS-Net-Lite on PASCAL VOC2012 (rows 1 and 2), NYUV2 (rows 3 and 4), and CITYSCAPES (rows 5 and 6) with input images and ground-truth segmentation masks (Seg).
Figure 8. Sample results of predicted and NLF-filtered semantic segmentation (NLF Seg) results that were obtained using DTS-Net-Lite on PASCAL VOC2012 (rows 1 and 2), NYUV2 (rows 3 and 4), and CITYSCAPES (rows 5 and 6) with input images and ground-truth segmentation masks (Seg).
Sensors 22 00337 g008
Figure 9. Quality comparison between the semantic segmentation results obtained from the proposed method (DTS-Net+NLF), DeepLab, DeepLabv3, PSPNet, and DeepLabv3+ on PASCAL VOC2012 validation benchmark, and between DTS-Net+NLF, PSPNet, and DeepLabv3 on CITYSCAPES validation benchmarks.
Figure 9. Quality comparison between the semantic segmentation results obtained from the proposed method (DTS-Net+NLF), DeepLab, DeepLabv3, PSPNet, and DeepLabv3+ on PASCAL VOC2012 validation benchmark, and between DTS-Net+NLF, PSPNet, and DeepLabv3 on CITYSCAPES validation benchmarks.
Sensors 22 00337 g009
Figure 10. Sample results of the joint semantic segmentation and depth estimation task using DTS-Net. The upper sample is from the CITYSCAPES benchmark and the bottom sample is from the NYUV2 benchmark. To the right of each input image, the ground truth and the output depth and semantic segmentation are given.
Figure 10. Sample results of the joint semantic segmentation and depth estimation task using DTS-Net. The upper sample is from the CITYSCAPES benchmark and the bottom sample is from the NYUV2 benchmark. To the right of each input image, the ground truth and the output depth and semantic segmentation are given.
Sensors 22 00337 g010
Table 1. Comparison of VGG16, ResNet50, DenseNet121, MobileNetV2, and Xception as feature extractors for DTS-Net without filtration in terms of MACs, the number of parameters (Params), and Pix.acc. on the NYUV2 dataset.
Table 1. Comparison of VGG16, ResNet50, DenseNet121, MobileNetV2, and Xception as feature extractors for DTS-Net without filtration in terms of MACs, the number of parameters (Params), and Pix.acc. on the NYUV2 dataset.
Feature ExtractorParams (M)MACs (G)Pix.acc. %
VGG16138.3 M62.1 G42.1
DenseNet1218 M11.1 G51.3
ResNet5025.6 M15.4 G62.5
MobileNetV23.5 M1.2 G66.7
Xception22.4 M18.5 G79.4
Table 2. Results that were obtained from DTS-Net and DTS-Net-Lite with and without NLF for each dataset VOC2012, NYUV2, and CITYSCAPES with the mIOU%, Pixel accuracy(Pix.acc.)% and the speed in FPS.
Table 2. Results that were obtained from DTS-Net and DTS-Net-Lite with and without NLF for each dataset VOC2012, NYUV2, and CITYSCAPES with the mIOU%, Pixel accuracy(Pix.acc.)% and the speed in FPS.
ModelImg SizeDatasetmIOU%Pix acc.%Speed (FPS)
DTS-Net480 × 480VOC201227.6690.0722
DTS-Net640 × 480NYUV233.9279.418.1
DTS-Net1024 × 512CITYSCAPES47.785.19.5
DTS-Net-Lite256 × 256VOC201214.7287.024
DTS-Net-Lite320 × 256NYUV224.566.7621.3
DTS-Net-Lite512 × 256CITYSCAPES35.382.814.7
DTS-Net+ NLF480 × 480VOC201291.198.815.5
DTS-Net+NLF640 × 480NYUV262.7784.7312.8
DTS-Net+NLF1024 × 512CITYSCAPES80.7293.16.6
DTS-Net-Lite+NLF256 × 256VOC201283.896.119.2
DTS-Net-Lite+NLF320 × 256NYUV248.4673.314.2
DTS-Net-Lite+NLF512 × 256CITYSCAPES61.582.0611.7
Table 3. Comparison of the proposed method and SOTA methods on PASCAL VOC2012 validation benchmark.
Table 3. Comparison of the proposed method and SOTA methods on PASCAL VOC2012 validation benchmark.
MethodBackbonemIOU (%)
DeepLab-CRF [9]ResNet-101+CRF77.69
DeepLabv3 [12]ResNet-10176.5
DeepLabv3+Res2Net [40]Res2Net-10179.3
DeepLabv3+ [13]Aligned Xception84.56
PSPNet [15]RestNet-15285.4
SpineNet-S143 [24]Dilated SpineNet85.64
HPN [19]ResNet-10185.8
ExFuse [41]ResNeXt-13185.8
Eff-B7 NAS-FPN [42]EfficientNet-B7+NAS+FPN86.6
EfficientNet-L2+NAS-FPN [21]EfficientNet-B7+NAS+FPN90.0
bf DTS-Net-Lite+NLFMobileNetV2+DTS83.8
DTS-Net+NLFXception+DTS91.1
Table 4. Comparison between the proposed method and SOTA methods on the NYUV2 semantic segmentation test benchmark. Note that ScanNet was evaluated on 11 classes only.
Table 4. Comparison between the proposed method and SOTA methods on the NYUV2 semantic segmentation test benchmark. Note that ScanNet was evaluated on 11 classes only.
MethodBackbonePix.acc.%
SceneNet [43]VGG1652.5
Hermans et al. [44]RDF+CRF54.3
SemanticFusion [45]VGG1659.2
ScanNet [46]NiN [47]60.7
3DMV [48]Enet [49]71.2
BPNet [39]U-Net73.5
DTS-Net-Lite+NLFMobileNetV2+DTS73.3
DTS-Net+NLFXception+DTS84.7
Table 5. Comparison between the proposed method and SOTA methods on CITYSCAPES semantic segmentation validation benchmark in terms of mIOU and speed (FPS).
Table 5. Comparison between the proposed method and SOTA methods on CITYSCAPES semantic segmentation validation benchmark in terms of mIOU and speed (FPS).
MethodBackbonemIOU (%)Speed (FPS)
DICENet [50]ShuffleNetV263.417.2
ContextNet [51]Shallow+deep CNN65.918.3
Template-Based NAS-arch1 [52]NAS-Net [53]69.597.0
HPN [19]ResNet-10171.74.0
FasterSeg [54]NAS-Net [53]73.16.1
SqueezeNAS [55]NAS-Net [53]75.20.0065
Dilated-ResNet [11]Dilated ResNet-10175.7-
EEEA-Net-C2 [26]NAS-Net [53]76.811.6
MDEQ-large [25]MDEQ77.8-
DeepLabv3 [12]Dilated ResNet-10178.52.0
DeepLabv3+ [13]Dilated Xception-7179.64.2
PSPNet [15]Dilated ResNet-10179.71.6
MDEQ-XL [25]MDEQ80.3-
RepVGG-B2 [27]VGG-like80.574.5
DTS-Net-Lite+NLFMobileNetV2+DTS61.511.7
DTS-Net+NLFXception+DTS80.726.6
Table 6. Comparison of the proposed method (DTS-Net+NLF) and the SOTA methods on depth estimation using a joint semantic segmentation and depth estimation model in terms of depth error (REL, Sq Rel, and RMSE), depth accuracy ( δ 1 , δ 2 , and δ 3 ), and disparity error (px).
Table 6. Comparison of the proposed method (DTS-Net+NLF) and the SOTA methods on depth estimation using a joint semantic segmentation and depth estimation model in terms of depth error (REL, Sq Rel, and RMSE), depth accuracy ( δ 1 , δ 2 , and δ 3 ), and disparity error (px).
MethodDatasetRELSq RelRMSE δ 1 δ 2 δ 3 px
SemDepth+CRF [32]NYUV20.1580.1210.6410.7690.9500.988-
TRL-ResNet50 [33]NYUV20.144-0.5010.8150.9620.992-
HybridNet A2 [34]CS0.2404.2712.090.5970.8220.929-
NYUV20.2020.1860.6820.6130.8920.974-
ESOSD-Net [35]NYUV20.145-0.5140.8050.9620.992-
CS------2.41
DTS-Net+NLFNYUV20.1020.5030.3100.9460.9970.999-
CS0.0981.0937.6270.9050.9560.9711.62
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ibrahem, H.; Salem, A.; Kang, H.-S. DTS-Net: Depth-to-Space Networks for Fast and Accurate Semantic Object Segmentation. Sensors 2022, 22, 337. https://doi.org/10.3390/s22010337

AMA Style

Ibrahem H, Salem A, Kang H-S. DTS-Net: Depth-to-Space Networks for Fast and Accurate Semantic Object Segmentation. Sensors. 2022; 22(1):337. https://doi.org/10.3390/s22010337

Chicago/Turabian Style

Ibrahem, Hatem, Ahmed Salem, and Hyun-Soo Kang. 2022. "DTS-Net: Depth-to-Space Networks for Fast and Accurate Semantic Object Segmentation" Sensors 22, no. 1: 337. https://doi.org/10.3390/s22010337

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop