Next Article in Journal
Stress and Workload Assessment in Aviation—A Narrative Review
Previous Article in Journal
A Field Study Implementing New Monitoring Technology for Roof Caving and Systematic Monitoring for Gob-Side Entry Retaining via Roof Cutting in Underground Coal Mining
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Regularized Denoising Masked Visual Pretraining for Robust Embodied PointGoal Navigation

1
Institute of Microelectronics of the Chinese Academy of Sciences, Beijing 100029, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(7), 3553; https://doi.org/10.3390/s23073553
Submission received: 1 February 2023 / Revised: 21 March 2023 / Accepted: 24 March 2023 / Published: 28 March 2023
(This article belongs to the Section Navigation and Positioning)

Abstract

:
Embodied PointGoal navigation is a fundamental task for embodied agents. Recent works have shown that the performance of the embodied navigation agent degrades significantly in the presence of visual corruption, including Spatter, Speckle Noise, and Defocus Blur, showing the weak robustness of the agent. To improve the robustness of embodied navigation agents to various visual corruptions, we propose a navigation framework called Regularized Denoising Masked AutoEncoders Navigation (RDMAE-Nav). In a nutshell, RDMAE-Nav mainly consists of two modules: a visual module and a policy module. In the visual module, a self-supervised pretraining method, dubbed Regularized Denoising Masked AutoEncoders (RDMAE), is designed to enable the Vision Transformers (ViT)-based visual encoder to learn robust representations. The bidirectional Kullback–Leibler divergence is introduced in RDMAE as the regularization term for a denoising masked modeling task. Specifically, RDMAE mitigates the gap between clean and noisy image representations by minimizing the bidirectional Kullback–Leibler divergence. Then, the visual encoder is pretrained by RDMAE. In contrast to existing works, RDMAE-Nav applies denoising masked visual pretraining for PointGoal navigation to improve robustness to various visual corruptions. Finally, the pretrained visual encoder with frozen weights is applied to extract robust visual representations for policy learning in the RDMAE-Nav. Extensive experiments show that RDMAE-Nav performs competitively compared with state of the arts (SOTAs) on various visual corruptions. In detail, RDMAE-Nav performs the absolute improvement: 28.2% in SR and 23.68% in SPL under Spatter; 2.28% in SR and 6.41% in SPL under Speckle Noise; and 9.46% in SR and 9.55% in SPL under Defocus Blur.

1. Introduction

In recent years, Embodied AI [1] has received extensive attention; it requires the agent to complete a specific task by interacting with the environment. As one of the specific tasks, PointGoal navigation has made great progress with the development of deep reinforcement learning (DRL), computer vision and robotics, and the emergence of a large number of photo-realistic simulation platforms [2,3,4]. PointGoal navigation is one of the most fundamental and important tasks in Embodied AI, and it is also the basis for the embodied agent to complete other more difficult tasks. In PointGoal navigation, an agent is required to move from its current position to a given coordinate point in 3D environments by using egocentric RGB (or RGB-D) observations and GPS+Compass localization only [5].
Traditional navigation methods, such as SLAM (Simultaneous Localization and Mapping)-based navigation methods, rely on environmental prior maps and high-precision sensors. Therefore, they exhibit poor generalization for prior unknown environments [6,7]. These difficulties have motivated a flux of research into DRL-based visual navigation techniques, which provide an end-to-end map-free navigation approach. By leveraging the powerful visual representation and decision-making capabilities of DRL, visual navigation is able to understand the environment with inexpensive sensors (such as monocular RGB cameras) and make full use of previously seen environments to learn knowledge for unseen environments’ generalization.
So far, a lot of work has been performed for PointGoal visual navigation [8,9,10,11,12]. Most of these efforts focus on improving generalization to novel environments, where the agent’s egocentric RGB observations are without visual corruption on evaluation. However, the ultimate purpose of the PointGoal navigation agent is to work in real environments with large variations in visual corruptions, e.g., the cameras are occluded by water droplets.
To evaluate the robustness of embodied navigation agents in various visual corruptions, ROBUSTNAV [13] provides a variety of realistic visual corruptions for emulating corruptions in the real world, and the robustness of the agent is evaluated after learning navigation policies with those visual corruptions. Ref. [13] points out that the performance of standard navigation agents, which work effectively with clean observation, drops dramatically in the presence of visual corruption. As shown in Figure 1a, the navigation agent’s observation is clean, and it is straightforward to choose the optimal action. However, it is more difficult for the agent to choose the optimal action in the presence of visual corruption, as shown in Figure 1b. It is worth noting that the agent has to choose an action at each time step in navigation. The weak robustness to visual corruption causes the agent to struggle to reach the target.
To address this issue, we propose a novel Regularized Denoising Masked AutoEncoders Navigation framework (RDMAE-Nav), which is a robust PointGoal navigation agent for various visual corruptions. To be specific, RDMAE-Nav mainly consists of two modules: a visual module and a policy module. In the visual module, a Vision Transformer (ViT)-based [14] visual encoder is introduced to extract features of the agent’s egocentric RGB observations. Moreover, a novel pretraining method is designed to learn robust visual representations, which is called Regularized Denoising Masked AutoEncoders (RDMAE). Inspired by the denoising mechanism in DMAE [15], which shows great robustness in image classification, the proposed RDMAE feeds both masked clean and noisy images in the encoder–decoder scheme to reconstruct the original clean images in a self-supervised way, where the two latent feature representations from clean and noisy images are mitigated via a constructed regularized loss by evaluating the bidirectional Kullback–Leibler (KL) divergence. Then, the visual encoder is pretrained to obtain robust and efficient visual priors via RDMAE, which provides a denoising masked modeling task with a regularized term. After the pretraining, the encoder of the RDMAE is taken as the visual encoder of the proposed RDMAE-Nav. Noteworthily, in the subsequent navigation policy learning of RDMAE-Nav, the pretrained visual encoder is applied to extract robust representations of the agent’s RGB observations only, and its parameters are not updated. To the best of our knowledge, RDMAE-Nav is the first attempt to exploit denoising masked visual pretraining for PointGoal navigation to improve robustness to various visual corruptions. As a result, the visual module encodes the agent’s egocentric RGB observations into the visual embedding. In the policy module, the goal localization is encoded into the goal embedding. Accordingly, the visual embedding, the goal embedding, and the previous hidden states are aggregated into the navigation policy network. The policy network takes advantage of the memory-capable Gated Recurrent Unit (GRU) [16] network to make future decisions based on past information. Additonally, the Decentralized Distributed Proximal Policy Optimization (DD-PPO) [8] paradigm is adopted to learn the policy network.
By following ROBUSTNAV [13], the experiments are conducted on the AI2THOR simulation platform [2], which provides a vast navigation environment with large variations. The results of robust navigation evaluation show the efficacy of our contribution and the advantages of our method over existing methods in a variety of visual corruptions. Overall, our contributions are summarized as follows:
  • We propose a robust PointGoal navigation framework RDMAE-Nav as the first attempt to apply denoising masked visual pretraining for embodied PointGoal navigation, which is a robust navigation agent for various visual corruptions.
  • We design a novel pretraining method, dubbed RDMAE, which introduces a regularization term for a denoising masked modeling task. RDMAE mitigates the gap of representation distributions between clean images and noisy ones by minimizing the bidirectional Kullback–Leibler (KL) divergence and consequently enables the visual encoder to obtain more robust and efficient visual priors.
  • Our method can achieve competitive performance over all competitors through experiments on the ROBUSTNAV benchmark [13], demonstrating the effectiveness and efficiency of the proposed RDMAE-Nav by employing Regularized Denoising masked visual pretraining for various visual corruptions.

2. Related Work

2.1. Embodied PointGoal Navigation

With the emergence of a large number of simulation platforms, such as Habitat [17] and AI2THOR [2], PointGoal navigation tasks have a standard dataset, agent configuration, and evaluation metrics, which have greatly boosted the research progress of PointGoal navigation. A landmark work is [8], whose agent architecture consists mainly of a visual encoder and a policy network. The visual encoder uses ResNet50 to extract features of the RGB-D visual input. The policy network consists of a two-layer Long Short-Term Memory (LSTM) [18] that takes the previous action, the localization information, and the output of the visual encoder as input. By proposing the DD-PPO algorithm, the training process of distributed reinforcement learning in a resource-intensive simulation environment is greatly accelerated. After about 2.5 billion frames of training, it achieved SOTA results at the Habitat Autonomous Navigation Challenge 2019, and the results are near-perfect. However, this approach relies on a large amount of computation and requires about 6 months of GPU time for training, which is unaffordable for the average researcher.
To improve sample efficiency and save computational resources, [9] investigated the PointGoal navigation method under resource constraints (specifically about 75 million frames and 1 GPU for 1 day). It adopts the more lightweight ResNet18 as the visual encoder and proposes not to use Generalized Advantage Estimation (GAE) [19] in the training and to use a larger batch size. Experimental results indicate that the performance of this approach can even outperform the method proposed by [8] in the resource-constrained condition.
An important reason why PointGoal navigation achieves near-perfect performance is that the simulation environment has perfect localization, which is usually difficult to satisfy in the real environment. Ref. [10] used visual odometry to replace the GPS+Compass sensor and achieved surprising performance on the Habitat PointNav benchmark. Ref. [11] combined classical SLAM approaches with learning-based approaches and proposed a learning-based differentiable SLAM approach to achieve State-Of-The-Art (SOTA) performance in the Habitat 2020 PointNav challenge. Ref. [12] proposed the use of a self-supervised method for monocular depth estimation to replace depth sensors on the PointNav navigation task and achieved excellent performance. There are also some works [20,21] studying sim2real, where they train the agent in a simulation environment and then use it in a real environment.
Most of the current work focuses on improving the generalization of the agent to novel environments, and the agent’s visual observations are clean during training and validation. However, there are various visual corruptions in real environments. Although [10,22,23] considered the visual sensors and actuation noise that may exist in navigation, these noise types are too simple to cover the rich visual corruptions that exist in real environments. In order to be able to quantitatively evaluate the robustness of embodied navigation agents to visual corruptions that may exist in the real environment, [13] proposed an evaluation framework called ROBUSTNAV. ROBUSTNAV contains a variety of visual corruptions, including Spatter, Speckle Noise, Camera Crack, Lower FOV, Defocus Blur, Motion Blur, etc. These visual corruptions can simulate real-world perturbation scenarios well. Ref. [13] pointed out that the performance of some near-perfect methods is severely degraded when evaluated in the presence of visual corruption.

2.2. Pretrained Visual Encoders in Embodied Visual Navigation

In recent years, pretrained visual encoders have been widely used in visual navigation. Compared with training from scratch, pretrained visual encoders that introduce visual priors lead to higher sample efficiency and better generalization of the navigation agent [24]. In [24], using midlevel visual representations to learn navigation policies instead of learning directly from the raw agent’s visual observations is proposed. Specifically, it pretrained the visual encoder on specific visual tasks and then used its freeze weights to extract visual representations from the raw images. Extensive experiments have demonstrated that using a pretrained visual encoder can greatly improve the sample efficiency and generalization of the navigation agent. In [25], the visual encoder of the navigation agent was allowed to be pretrained on a self-supervised environment prediction task. The pretrained visual encoder was able to learn the spatial representation of the environment and could be better used for downstream navigation tasks. In [26], VTNet (Visual Transformer NetWork) was used to correlate visual representations with navigation signals for visual pretraining, which accelerated the learning on navigation policy. In [27], a CLIP (Contrastive language image pretraining) [28] encoder with frozen weights was directly used as a visual encoder without any fine-tuning and achieved surprising performance on multiple embodied visual navigation tasks.
There is also some work showing the use of a pretrained visual encoder can improve the performance of navigation agents in real environments. In [29], a pretrained DINO [30] was used as the visual encoder and fine-tuned on 70 RGB images with coarse semantic segmentation labels collected in a real environment. The results show that the robot was able to perform the visual navigation task well in a real environment. In [31], spatial autoencoders were used to pretrain the visual encoders on real environment images. The visual encoders with frozen weights were used in the learning of navigation policies during training in the simulated environment. The results show that the navigation policy had better sim2real performance on visual navigation.

2.3. Masked Autoencoders in Reinforcement Learning and Robotics

Masked Autoencoders (MAEs) [32] are self-supervised pretraining models based on an encoder–decoder structure that enable the encoder to learn visual representations by reconstructing the masked image. During pretraining, an MAE randomly masks the image patches with a fixed ratio, then the unmasked image patches are input to the encoder to extract latent representations, and finally, the decoder receives latent representations to reconstruct the original image. After the pretraining, the encoder will be used for downstream tasks. MAEs are now widely used in computer vision tasks [33] and have been shown to be a robust data augmentation method [34].
Although related work is relatively scarce, MAEs have also been used in reinforcement learning and robotics. In [35], an MAE was pretrained with real environment images, and the encoder with frozen weights was directly used as a visual encoder for a variety of downstream motor control tasks without any task-specific fine-tuning. The results show that the performance of the MAE pretrained encoder is significantly improved compared with the supervised encoder for downstream motor control tasks. In [36], an MAE with CNN-based models was compared on image-based reinforcement learning control tasks, and it was shown that the MAE was able to outperform CNN-based models on some tasks.

3. Method

3.1. Task Definition

In the PointGoal navigation task, the agent must navigate to the target position by using RGB observations and GPS+Compass localization only. Concretely, the agent is initialized with a random location and orientation at the beginning of the episode. As shown in Figure 2, at each time step t, the agent obtains RGB observation O t from the monocular camera and target localization information L t from GPS+Compass, respectively. In particular, the localization information L t consists of two parts, d t and o t , where d t is the distance of the target relative to the agent, and o t is the orientation of the target relative to the agent. Given the visual information O t , the localization information L t , and previous hidden information h t 1 , the agent is required to learn an optimal navigation policy π a t | s t for the output action a t via DRL, where the state space is s t = { O t , L t , h t 1 } , and the action space a t contains four discrete actions, namely m o v e f o r w a r d 0.25 m , t u r n l e f t 30 , t u r n r i g h t 30 , s t o p . Overall, the episode is considered a success if the agent issues a s t o p action within 0.2 m of the target within a maximum of 300 steps. Otherwise, it is considered a failure. Note that the geodesic distance is used in the evaluation.
Noteworthily, previous works focus on the generalization to novel environments [8,9,10,11,12]. Nevertheless, the robustness of embodied navigation agents to visual corruptions also plays a crucial role in the navigation policy. ROBUSTNAV [13], as the first benchmark platform for robustness evaluation of embodied navigation agents, provides a number of solutions for visual corruptions. As one step further, we emphasize that the navigation policy should pay attention to generalization as well as robustness to various visual corruptions as the optimal strategy to follow. To this end, we propose the robust embodied navigation framework, RDMAE-Nav, with the details described in the following sections.

3.2. Overall Architecture of RDMAE-Nav

The overall architecture of the proposed RDMAE-Nav is shown in Figure 2. The framework follows the DRL pipeline, with the state space formed by the agent’s egocentric RGB observation, the target localization, and the previous hidden states. AI2THOR [2] is chosen as the simulation platform, which updates the state and outputs a reward for the next training step of DRL.
In the visual module, the visual encoder is the parameter-fixed encoder from a pretrained Regularized Denoising Mask AutoEncoder (RDMAE), which is depicted as RDMAE Encoder in Figure 2. The pretraining of the visual encoder is implemented by a self-supervised learning procedure with encoding–decoding as the auxiliary task. The pretrained visual encoder is involved in the succeeding navigation policy learning. A regularization term is imposed on the denoising paradigm by tackling both clean and noisy images from the RGB observations, which benefits the visual representation extraction for robust environment perception and understanding. The visual representations extracted by the visual encoder are projected as the visual embedding through a linear layer. Additionally, the target localization is projected as the target embedding by another linear layer. The visual embedding and the target embedding are aggregated as the joint embedding before being fed into the policy module.
The policy module includes a single-layer GRU and two single-layer MLPs, which are served as the actor and the critic of the DRL algorithm, respectively. The actor maps the output of GRU as the logits, which represent the discrete distribution of the actions, and the critic maps the output of GRU into a scalar value. The GRU network enables the memory ability of the policy network to consider historical experiences for future action determination. Decentralized Distributed Proximal Policy Optimization (DD-PPO) [8] is adopted as the DRL algorithm, which is a distributed Proximal Policy Optimization (PPO) [37] algorithm to accelerate the training process of the simulation environment effectively.
For DRL, the reward function r t is important to update the policy network, where r t = { r s u c c e s s , r m o v e } . In our work, if the task is completed successfully at the time step t, the reward is set as 10, i.e., r s u c c e s s = 10 ; otherwise, the reward for moving one step r m o v e is set as GeoDist s t , a t GeoDist s t 1 , a t 1 0.01 , where −0.01 is the time penalty. Here, GeoDist denotes the geodesic distance from the agent’s current position to the target point. To sum up, the reward function r t can be written as
r t = 10 , if success , GeoDist s t 1 , a t 1 GeoDist s t , a t 0.01 , otherwise .

3.3. Regularized Denoising Masked AutoEncoders (RDMAE) for Visual Pretraining

The visual encoder in RDMAE-Nav is pretrained by RDMAE, as shown in Figure 3. Assume x is the input clean image, and η is the Gaussian additive noise with noise level σ , where η N 0 , σ 2 I , I is the identity matrix. The Gaussian noise is added onto the clean image to generate a noisy image x + η . Both the noisy image x + η and the original clean one x are divided into the nonoverlapping patches, which are masked into Mask ( x + η ) and Mask ( x ) , respectively, by a predefined masking ratio.
A self-supervised learning scheme is utilized to extract latent features for discriminative visual representations, with an auxiliary task constructed under the condition of insufficient labels. The autoencoder of the encoding–decoding mechanism is employed in the self-supervised learning scheme by following [15], with some modifications. Specifically, the encoder maps the input of both noisy and clean images into the low-dimensional feature space, and the decoder reconstructs the potential features of the noisy images back to the original clean images. Since both noisy and clean images are masked to be involved in the encode–decode process, a bidirectional Kullback–Leibler (KL) divergence loss is constructed to mitigate the representation distribution gap between them,
L K L = 1 2 D K L h | | h + D K L h | | h ,
where h and h denote the representation distributions of the masked clean images and noisy ones, respectively, and are defined as below:
h = Encoder Mask x ,
h = Encoder Mask x + η .
D K L is the KL divergence to evaluate the difference between two distributions,
D K L p x | | q x = E x p x log p x q x ,
where p ( x ) and q ( x ) stand for two probability distributions; E is the notation for expectation. The proposed bidirectional KL divergence loss is activated to mitigate the gap of representational distributions between clean and noisy images, which in turn is to achieve denoising implicitly.
The reconstruction procedure is also enabled in pretraining as the auxiliary denoising task by leveraging the reconstructed image x ^ . The reconstruction loss is implemented by calculating Mean Square Error (MSE) loss at the pixel level of all patches between the reconstructed image x ^ and the clean image x by following [15],
L r e c o n s = 1 N i = 1 N x ^ i x i 2 ,
where N is the number of pixels of each image.
The overall loss L is composed of the KL loss L K L and the reconstruction loss L r e c o n s , which is expressed as
L = L r e c o n s + α L K L ,
where α is the coefficient weight to control the KL loss.
Since image reconstruction and denoising are performed simultaneously, the visual encoder is capable of extracting discriminative semantic features as well as generalizing robustness against noises.
After pretraining, the decoder is discarded and the pretrained visual encoder is mounted in the RDMAE-Nav with the fixed parameters to make full use of its robustness in noises for the downstream navigation task with visual corruptions. Additionally, in the subsequent navigation policy learning, the masks are weeded out, and full sets of image patches are applied to the encoder.

3.4. Vision-Transformer-Based Visual Encoder

The visual encoder of the proposed RDMAE leverages the powerful representation capability of the Vision Transformer [14], which was originally designed for image classification, for a more challenging visual navigation task. The difference between the visual encoder and the ViT is that only unmasked patches are applied to the visual encoder during pretraining. The visual encoder is based on the ViT, and they share the same network architecture, as detailed in Figure 4. The Transformer [38] is equipped with powerful visual representation capabilities endowed by a self-attention-based encoder–decoder. As mentioned previously, the input agent’s RGB observation is divided into fixed-size patches, each of which is projected as patch embedding by a linear layer. An extralearnable patch embedding is prepended to the sequence of patch embedding, marked as h 0 * in Figure 4, whose final state h N * through the Transformer encoder is adopted as the image representation for the downstream navigation task. Then, position embeddings are added to the patch embeddings and fed into the Transformer encoder. The Transformer encoder constitutes N stacked Transformer blocks, and each block consists of Multiheaded Attention (MHA) and MLP modules, where LayerNorm is applied before every module and residual connections are applied after every module. MHA allows the encoder to jointly attend to information of the input embeddings from different representation subspaces at different positions [38]. Therefore, the encoder is able to integrate information across the entire image and has stronger representation learning capabilities, which is beneficial to visual navigation.

4. Experiments

4.1. Simulation Platform

The experiments are conducted on the simulation platform AI2THOR [2] with the dataset RoboTHOR [39]. RoboTHOR contains 75 indoor scenes of 8.8 m × 3.9 m, of which 60 scenes are used for training (Figure 5a) and 15 scenes are used for validation (Figure 5b). There is a total of 108,000 different navigation tasks in the training scenes and 1100 different navigation tasks in the validation scenes. Some selected samples of scenes are shown in Figure 5. The validation scenes are set differently from the training scenes to evaluate the agent’s generalization to the novel scenes. The LoCoBot [40] robot is used as the navigation agent, which is equipped with an Intel RealSense camera.

4.2. Data Preraration

For RDMAE visual pretraining, we collected 60k egocentric RGB images of the agent in a resolution of 300 × 400 from 60 training scenes of RoboTHOR, of which 1k images were collected for each scene. The data collection was accomplished by making the agent move randomly in the scene to save the egocentric images. Some of the collected images are shown in Figure 6.

4.3. Visual Corruptions Description

In addition to evaluating the agent’s generalization ability to new scenes, we also evaluated its robustness to visual corruptions. The considered visual corruptions are Spatter, Speckle Noise, Camera Crack, Lower Fov, Motion Blur, and Defocus Blur, which are supported by ROBUSTNAV [13], as shown in Figure 7. Specifically, Spatter simulates the presence of camera lens occlusion, Speckle Noise simulates the inherent granular interference in the camera, Camera Crack simulates the presence of camera lens cracking, Lower Field Of View (FOV) requires the agent to use a lower camera FOV in evaluating (39.5°) than that used in training (79°), Motion Blur simulates blurred images caused by the agent jitters, Defocus Blur simulates the camera being out of focus, and Clean is the absence of visual corruption. By following ROBUSTNAV [13], the above visual corruptions are unseen by the agent during navigation policy training.

5. Experimental Results and Discussion

5.1. Evaluation Metrics

We adopt the commonly used Success Rate (SR) and Success weighted by Path Length (SPL) [5] as the evaluation metrics. During the evaluation, the agent is asked to navigate from the current position to a given target point, and SR indicates the success rate of navigation, which is
SR = 1 N i = 1 N S i ,
where N denotes the number of evaluated tasks, and S i denotes whether the i-th task is successful, which is 1 if successful and 0 otherwise. SPL represents the ratio of the path length of the successful tasks to the shortest path length, which is
SPL = 1 N i = 1 N S i l i max p i , l i ,
where l i is the shortest distance of the i-th task from the start to the target, and p i is the actual path length. A higher SR indicates a higher effectiveness of the navigation agent, and a higher SPL indicates higher efficiency.
In addition to the two common metrics mentioned above, we adopt another two metrics [13] for analyzing the agent’s behavior: average reward (R) and Dist2Target (Dist). R is the average reward obtained by the agent on evaluation, defined as below:
R = 1 N i = 1 N r i ,
where r i is the reward obtained by the agent of the i-th task. Dist is the average distance from the target when the agent issues the s t o p command, defined as below:
Dist = 1 N i = 1 N d i ,
where d i is the distance from the target of the i-th task when the agent issues the s t o p command. In general, if the agent obtains a higher R and a lower Dist, the agent achieves more reasonable behavior.

5.2. RDMAE Pretraining Configuration

RDMAE is pretrained on the collected 60k agents’ egocentric images, all of which are resized to a fixed solution of 224 × 224. Unlike the original RDMAE, we use a smaller encoder–decoder structure based on ViT [14], which is ViT-S with an input patch size of 16 × 16. The ViT-S encoder consists of 12 Transformer blocks with embedding dimensions of 384, 6 attention heads, and an MLP multiplier of 4. ViT-S has about 22M parameters. The decoder uses 8 Transformer blocks with embedding dimensions of 256 and 16 attention heads. This asymmetric encoder–decoder structure ensures the encoder learns rich semantic features and reduces the pretraining time significantly [32].
This paper follows the mask ratio of 0.75, as in [32]. The learning rate is set as 1.5 × 10 4 , the weight decay is 0.05, the batch size is set as 256, the epoch is 1000, and the noise level σ is set to 0.5. Empirically, the coefficient weight of KL Loss α is set as 0.1. The AdamW [41] optimizer with β 1 = 0.9 and β 2 = 0.95 is adopted. To prevent overfitting, we use simple data augmentation, including RandomResizedCrop with a crop ratio of (0.2, 1.0) and RandomHorizontalFlip. Two NVIDIA GeForce RTX 3090 GPUs are used for training.
After training, the RDMAE visualization of the reconstructed images is shown in Figure 8a–d from four different scenes in the RoboTHOR validation set. For each image from left to right are the original image, the Gaussian noise image (from top to bottom noise level σ = 0.5, 0.25, 0.1), the masked Gaussian noise image, and the reconstructed image. It is worth noting that the reconstructed images we show here are from the RoboTHOR validation set, which has not been seen by the agent during RDMAE training. Compared with the original clean image, RDMAE is able to reconstruct the masked noisy image well, indicating that RDMAE is able to extract robust features.

5.3. RDMAE-Nav Training Configuration

The procedure of RDMAE-Nav training is as follows. The agent’s egocentric RGB observation with a fixed resolution of 300 × 400 is resized to 224 × 224 before being fed into the visual module. The RGB observation is encoded into a 384 dim embedding by the pretrained visual encoder with frozen weights and then projected into a 512 dim visual embedding by a linear layer. The target localization is projected into a 32 dim target embedding by another linear layer. Next, the visual embedding and the target embedding are combined into a 544 dim joint embedding. The joint embedding is projected as a 1568-dimensional vector by a linear layer and subsequently fed into the GRU with 512 hidden units, along with the previous hidden state. The GRU outputs a 512 dim vector and the next hidden state. Finally, the 512 dim vector is received by two separate MLPs, which output 4 dim action logits and a scalar value, respectively.
Referring to [13], the learning rate is set as 3 × 10 4 and declines linearly. The discount factor is set to 0.99, and the GAE parameter is set to 0.95. The length of the rollout is 128. The optimizer uses Adam [42]. The DD-PPO clip parameter is set to 0.1. The epochs of DD-PPO is set to 4. For a fair comparison with [13], the navigation policy was trained for a total of 75M frames, in line with itself. Training was performed on a single NVIDIA GeForce RTX2060 GPU.

5.4. Performance on Visual Corruptions

We compare our method with the following approaches proposed in ROBUSTNAV.
(1)
ROBUSTNAV is the standard approach to evaluate the benchmarks by using ResNet18 to extract visual features.
(2)
ROBUSTNAV+AP is based on ROBUSTNAV by introducing an additional action prediction self-supervised task to resist visual corruptions.
(3)
ROBUSTNAV+AP+SS-Adapt is based on ROBUSTNAV+AP by introducing self-supervised adaptation on specific corruptions (Spatter, Camera Crack, Lower-FOV, and Defocus Blur).
(4)
ROBUSTNAV+RP is based on ROBUSTNAV by introducing an additional rotation prediction task to resist visual corruptions.
(5)
ROBUSTNAV+RP+SS-Adapt is based on ROBUSTNAV+RP by introducing the self-supervised adaptation on specific corruptions (Spatter, Camera Crack, Lower-FOV, and Defocus Blur).
(6)
ROBUSTNAV+Data Aug introduces various data augmentation methods during training. For more information about the above approaches, please refer to [13].
The comparisons on visual corruptions range from Clean, Spatter, Speckle Noise, Camera Crack, Lower Fov, Defocus Blur, and Motion Blur. Among the above corruptions, Spatter, Speckle Noise, Defocus Blur, and Motion Blur are classified into five severity levels, where level 5 indicates the most severe corruptions. For a fair comparison, the adjustable corruptions are set as level 5 by following ROBUSTNAV [13].
Table 1 shows the performance comparison of the navigation methods under clean and visual corruptions. For Spatter, compared with the suboptimal methods ROBUSTNAV+RP+SS-Adapt, which are 61.06% in SR and 47.16% in SPL, the proposed RDMAE-Nav improves the performance in both metrics by large margins, which are 89.26% in SR and 70.84% in SPL. The absolute improvement is 28.2% in SR and 23.68% in SPL. Similar results are observed in Speckle Noise and Defocus Blur. For Speckle Noise, compared with the suboptimal methods ROBUSTNAV+RP (under the SR metric) and ROBUSTNAV+Data Aug (under the SPL metric), the RDMAE-Nav improves the metric SR by 2.28% and the metric SPL by 6.41%. For Defocus Blur, compared with the suboptimal methods ROBUSTNAV+AP (under the SR metric) and ROBUSTNAV+RP+SS-Adapt (under the SPL metric), the RDMAE-Nav improves the metric SR by 9.46% and the metric SPL by 9.55%. For Cam Crack, Lower-FOV, and Motion Blur, the proposed RDMAE-Nav achieves comparable performance as well. Moreover, a remarkable observation is that for Defocus Blur and Motion Blur, mere performance degradation is observed compared with that in a clean environment. To sum up, the proposed RDMAE-Nav shows robustness against various visual corruptions.
To further analyze the agent behavior of the proposed RDMAE-Nav, we conducted the experiments under the metrics of R and Dist, and the results are shown in Table 2. Exemplarily, for Spatter, RDMAE-Nav exceeds the suboptimal method ROBUSTNAV+RP+SS-Adapt by 3.224 with 8.238 compared with 5.014, while several methods even pose a negative reward. In addition, RDMAE-Nav exhibits the lowest Dist of 0.5502. Similar conclusions can be drawn for Speckle Noise and Defocus Blur. For Cam Crack, Lower-FOV, and Motion Blur, RDMAE-Nav also achieves comparable performance. In the presence of visual corruption, RDMAE-Nav obtains higher rewards and stops closer to the target, which indicates that it makes more reasonable decisions.
Additionally, examples of the evaluated navigation trajectory visualization of RDMAE-Nav, ROBUSTNAV, and ROBUSTNAV+RP+SS-Adapt under Spatter are shown in Figure 9. The blue cardinal represents the agent’s initial position and orientation, the orange square represents the agent’s stop position, and the pentagram represents the target. Compared with ROBUSTNAV, RDMAE-Nav shows superior navigation performance in the presence of Spatter. Compared with ROBUSTNAV+RP+SS-Adapt, even though it reaches the target (middle part of Figure 9c), its actual path is longer compared with RDMAE-Nav (middle part of Figure 9a).

5.5. Ablations

5.5.1. Contributions of KL Loss

The proposed RDMAE-Nav introduces a KL-divergence between the representation distribution of clean and noisy images. In this study, we specifically analyzed the contributions of KL Loss. This ablation removes the L K L loss of the visual encoder and only keeps the L r e c o n s loss for pretraining. The results are shown in Figure 10 and Table 3.
It is known from Figure 10 that introducing the KL Loss into the pretraining for the visual encoder improves the performance of RDMAE-Nav both in clean and various visual corruptions. In detail, for Speckle Noise, the performance improvement is the most significant by introducing KL Loss to RDMAE-Nav, with absolute improvements of 23.48% in SR and 21.63% in SPL, as in Table 3. For other corruptions, there are also varying degrees of performance improvement. From Table 4, RDMAE-Nav obtains higher R and lower Dist compared with RDMAE-Nav without KL Loss, which indicates that the introduction of KL Loss makes the navigation agent have more reliable and robust decisions in the presence of visual corruption.

5.5.2. Contributions of Regularized Denoising

Note that the visual encoder of RDMAE-Nav is pretrained by a Regularized Denoising reconstruction task. In this study, we specifically analyzed the contributions of Regularized Denoising. This ablation removes Regularized Denoising during pretraining of the visual encoder. To conform with the configuration, we do not add Gaussian noise to the clean image and reconstruct the masked clean image accordingly by the encoder–decoder. The results are shown in Figure 11 and Table 5.
Basically, RDMAE-Nav without Regularized Denoising is on par with the complete version of RDMAE-Nav in a clean environment. However, in the presence of visual corruption, the performance of RDMAE-Nav without Regularized Denoising degrades dramatically. For Spatter, the SR and SPL of RDMAE-Nav without Regularized Denoising are both extremely low, only 8.553% and 5.923%, respectively, while the proposed Regularized Denoising mechanism makes great contributions to boosting performance. Similar results are observed in the other visual corruptions. From Table 6, RDMAE-Nav obtains higher R and lower Dist compared with RDMAE-Nav whiteout Regularized Denoising, which indicates that the introduction of Regularized Denoising makes the agent navigation have more reasonable decisions in the presence of visual corruption. It is worth noting that the Regularized Denoising mechanism has a significant improvement for Spatter and Speckle Noise, probably because these two corruptions are close to Gaussian noise, and RDMAE-Nav obtains stronger resistance after pretraining.

6. Conclusions

In this work, we proposed a robust PointGoal navigation framework, called RDMAE-Nav, which is robust to various visual corruptions. We designed a self-supervised pretraining method to enable the visual encoder of RDMAE-Nav to learn robust representations, which is called Regularized Denoising Masked AutoEncoders (RDMAE). Thanks to a denoising reconstruction task introduced in the RDMAE, RDMAE-Nav achieves robust representation ability. To the best of our knowledge, RDMAE-Nav is the first attempt to apply denoising masked visual pretraining in the PointGoal navigation task. Furthermore, RDMAE constructs a regularization loss by calculating the bidirectional Kullback–Leibler divergence between clean and noisy image representations, which mitigates the gap between them and thus enables the visual encoder of RDMAE-Nav to learn more robust representations. Extensive experiments on ROBUSTNAV, the benchmark platform for robustness evaluation for embodied visual navigation, demonstrate that RDMAE-Nav exhibits competitive robust performance compared with the state of the art.
Future works include the following aspects. Although RDMAE-Nav shows great advances in various visual corruptions, the performance still declines conspicuously compared with that in a clean environment. Therefore, more efforts are needed to alleviate the performance drop. Moreover, the structure of the visual encoder can be further improved to maximize the performance of RDMAE. Furthermore, evaluating the robustness of RDMAE-Nav in real-world scenarios is challenging but important work. Finally, the reason why Regularized Denoising masked visual pretraining shows different performance improvements for different visual corruptions deserves deeper research.

Author Contributions

Conceptualization, J.P. and J.L.; methodology, J.P. and Y.X.; software, J.P.; validation, J.P.; formal analysis, J.P.; investigation, J.P. and Y.X.; resources, J.P. and J.L.; data curation, J.P. and Y.X.; writing—original draft preparation, J.P.; writing—review and editing, J.P., Y.X., L.L., H.L., K.L. and J.L.; visualization, J.P.; supervision, J.L., L.L. and H.L.; project administration, J.L.; funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Key Research and Development Program of China (No.2021YFB2501403) and the Science and Technology Service Network Initiative Program of The Chinese Academy of Sciences (STS20201600200122).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data in this study are available from the first author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Duan, J.; Yu, S.; Tan, H.L.; Zhu, H.; Tan, C. A survey of embodied ai: From simulators to research tasks. IEEE Trans. Emerg. Top. Comput. Intell. 2022, 6, 230–244. [Google Scholar] [CrossRef]
  2. Kolve, E.; Mottaghi, R.; Han, W.; VanderBilt, E.; Weihs, L.; Herrasti, A.; Gordon, D.; Zhu, Y.; Gupta, A.; Farhadi, A. Ai2-thor: An interactive 3d environment for visual ai. arXiv 2017, arXiv:1712.05474. [Google Scholar]
  3. Li, C.; Xia, F.; Martín-Martín, R.; Lingelbach, M.; Srivastava, S.; Shen, B.; Vainio, K.E.; Gokmen, C.; Dharan, G.; Jain, T.; et al. iGibson 2.0: Object-Centric Simulation for Robot Learning of Everyday Household Tasks. In Proceedings of the Conference on Robot Learning, London, UK, 8–11 November 2022; pp. 455–465. [Google Scholar]
  4. Szot, A.; Clegg, A.; Undersander, E.; Wijmans, E.; Zhao, Y.; Turner, J.; Maestre, N.; Mukadam, M.; Chaplot, D.S.; Maksymets, O.; et al. Habitat 2.0: Training home assistants to rearrange their habitat. Adv. Neural Inf. Process. Syst. 2021, 34, 251–266. [Google Scholar]
  5. Anderson, P.; Chang, A.; Chaplot, D.S.; Dosovitskiy, A.; Gupta, S.; Koltun, V.; Kosecka, J.; Malik, J.; Mottaghi, R.; Savva, M.; et al. On evaluation of embodied navigation agents. arXiv 2018, arXiv:1807.06757. [Google Scholar]
  6. Tai, L.; Paolo, G.; Liu, M. Virtual-to-real deep reinforcement learning: Continuous control of mobile robots for mapless navigation. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 31–36. [Google Scholar]
  7. Bansal, S.; Tolani, V.; Gupta, S.; Malik, J.; Tomlin, C. Combining optimal control and learning for visual navigation in novel environments. In Proceedings of the Conference on Robot Learning, Virtual, 16–18 November 2020; pp. 420–429. [Google Scholar]
  8. Wijmans, E.; Kadian, A.; Morcos, A.; Lee, S.; Essa, I.; Parikh, D.; Savva, M.; Batra, D. Dd-ppo: Learning near-perfect pointgoal navigators from 2.5 billion frames. arXiv 2019, arXiv:1911.00357. [Google Scholar]
  9. Wijmans, E.; Essa, I.; Batra, D. How to Train PointGoal Navigation Agents on a (Sample and Compute) Budget. In Proceedings of the 21st International Conference on Autonomous Agents and Multiagent Systems, Virtual, 9–13 May 2022; pp. 1762–1764. [Google Scholar]
  10. Zhao, X.; Agrawal, H.; Batra, D.; Schwing, A.G. The surprising effectiveness of visual odometry techniques for embodied pointgoal navigation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 16127–16136. [Google Scholar]
  11. Karkus, P.; Cai, S.; Hsu, D. Differentiable slam-net: Learning particle slam for visual navigation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Montreal, BC, Canada, 11–17 October 2021; pp. 2815–2825. [Google Scholar]
  12. Tang, T.; Du, H.; Yu, X.; Yang, Y. Monocular Camera-based Point-goal Navigation by Learning Depth Channel and Cross-modality Pyramid Fusion. Proc. AAAI Conf. Artif. Intell. 2022, 36, 5422–5430. [Google Scholar] [CrossRef]
  13. Chattopadhyay, P.; Hoffman, J.; Mottaghi, R.; Kembhavi, A. Robustnav: Towards benchmarking robustness in embodied navigation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 15691–15700. [Google Scholar]
  14. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  15. Wu, Q.; Ye, H.; Gu, Y.; Zhang, H.; Wang, L.; He, D. Denoising Masked AutoEncoders are Certifiable Robust Vision Learners. arXiv 2022, arXiv:2210.06983. [Google Scholar]
  16. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  17. Savva, M.; Kadian, A.; Maksymets, O.; Zhao, Y.; Wijmans, E.; Jain, B.; Straub, J.; Liu, J.; Koltun, V.; Malik, J.; et al. Habitat: A platform for embodied ai research. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Rrepublic of Korea, 27 October–2 November 2019; pp. 9339–9347. [Google Scholar]
  18. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  19. Schulman, J.; Moritz, P.; Levine, S.; Jordan, M.; Abbeel, P. High-dimensional continuous control using generalized advantage estimation. arXiv 2015, arXiv:1506.02438. [Google Scholar]
  20. Truong, J.; Chernova, S.; Batra, D. Bi-directional domain adaptation for sim2real transfer of embodied navigation agents. IEEE Robot. Autom. Lett. 2021, 6, 2634–2641. [Google Scholar] [CrossRef]
  21. Sadek, A.; Bono, G.; Chidlovskii, B.; Wolf, C. An in-depth experimental study of sensor usage and visual reasoning of robots navigating in real environments. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 9425–9431. [Google Scholar]
  22. Partsey, R.; Wijmans, E.; Yokoyama, N.; Dobosevych, O.; Batra, D.; Maksymets, O. Is Mapping Necessary for Realistic PointGoal Navigation? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17232–17241. [Google Scholar]
  23. Lee, E.S.; Kim, J.; Kim, Y.M. Self-Supervised Domain Adaptation for Visual Navigation with Global Map Consistency. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, New Orleans, LA, USA, 18–24 June 2022; pp. 1707–1716. [Google Scholar]
  24. Sax, A.; Emi, B.; Zamir, A.R.; Guibas, L.; Savarese, S.; Malik, J. Mid-level visual representations improve generalization and sample efficiency for learning visuomotor policies. arXiv 2018, arXiv:1812.11971. [Google Scholar]
  25. Ramakrishnan, S.K.; Nagarajan, T.; Al-Halah, Z.; Grauman, K. Environment predictive coding for embodied agents. arXiv 2021, arXiv:2102.02337. [Google Scholar]
  26. Du, H.; Yu, X.; Zheng, L. VTNet: Visual transformer network for object goal navigation. arXiv 2021, arXiv:2105.09447. [Google Scholar]
  27. Khandelwal, A.; Weihs, L.; Mottaghi, R.; Kembhavi, A. Simple but effective: Clip embeddings for embodied ai. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 14829–14838. [Google Scholar]
  28. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Baltimore, MD, USA, 6–14 December 2021; pp. 8748–8763. [Google Scholar]
  29. Saavedra-Ruiz, M.; Morin, S.; Paull, L. Monocular Robot Navigation with Self-Supervised Pretrained Vision Transformers. arXiv 2022, arXiv:2203.03682. [Google Scholar]
  30. Caron, M.; Touvron, H.; Misra, I.; Jégou, H.; Mairal, J.; Bojanowski, P.; Joulin, A. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 9650–9660. [Google Scholar]
  31. Morad, S.D.; Mecca, R.; Poudel, R.P.; Liwicki, S.; Cipolla, R. Embodied visual navigation with automatic curriculum learning in real environments. IEEE Robot. Autom. Lett. 2021, 6, 683–690. [Google Scholar] [CrossRef]
  32. He, K.; Chen, X.; Xie, S.; Li, Y.; Dollár, P.; Girshick, R. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16000–16009. [Google Scholar]
  33. Zhang, C.; Zhang, C.; Song, J.; Yi, J.S.K.; Zhang, K.; Kweon, I.S. A survey on masked autoencoder for self-supervised learning in vision and beyond. arXiv 2022, arXiv:2208.00173. [Google Scholar]
  34. Xu, H.; Ding, S.; Zhang, X.; Xiong, H.; Tian, Q. Masked autoencoders are robust data augmentors. arXiv 2022, arXiv:2206.04846. [Google Scholar]
  35. Xiao, T.; Radosavovic, I.; Darrell, T.; Malik, J. Masked visual pre-training for motor control. arXiv 2022, arXiv:2203.06173. [Google Scholar]
  36. Tao, T.; Reda, D.; van de Panne, M. Evaluating Vision Transformer Methods for Deep Reinforcement Learning from Pixels. arXiv 2022, arXiv:2204.04905. [Google Scholar]
  37. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal policy optimization algorithms. arXiv 2017, arXiv:1707.06347. [Google Scholar]
  38. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
  39. Deitke, M.; Han, W.; Herrasti, A.; Kembhavi, A.; Kolve, E.; Mottaghi, R.; Salvador, J.; Schwenk, D.; VanderBilt, E.; Wallingford, M.; et al. Robothor: An open simulation-to-real embodied ai platform. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3164–3174. [Google Scholar]
  40. Murali, A.; Chen, T.; Alwala, K.V.; Gandhi, D.; Pinto, L.; Gupta, S.; Gupta, A. Pyrobot: An open-source robotics framework for research and benchmarking. arXiv 2019, arXiv:1906.08236. [Google Scholar]
  41. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  42. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. Illustration of navigation under clean/corrupted observation. (a) Optimal action decision under clean observation. (b) Non-optimal action decision under corrupted observation.
Figure 1. Illustration of navigation under clean/corrupted observation. (a) Optimal action decision under clean observation. (b) Non-optimal action decision under corrupted observation.
Sensors 23 03553 g001
Figure 2. Overview architecture of RDMAE-Nav framework.
Figure 2. Overview architecture of RDMAE-Nav framework.
Sensors 23 03553 g002
Figure 3. Illustration of RDMAE pretraining.
Figure 3. Illustration of RDMAE pretraining.
Sensors 23 03553 g003
Figure 4. Network architecture of the ViT-based visual encoder.
Figure 4. Network architecture of the ViT-based visual encoder.
Sensors 23 03553 g004
Figure 5. Top view of the selected scenes in the RoboTHOR dataset.
Figure 5. Top view of the selected scenes in the RoboTHOR dataset.
Sensors 23 03553 g005
Figure 6. Agent egocentric RGB images for RDMAE visual pretraining.
Figure 6. Agent egocentric RGB images for RDMAE visual pretraining.
Sensors 23 03553 g006
Figure 7. (a) The agent’s egocentric clean RGB observation. (b) The agent’s egocentric RGB observation in the presence of visual corruption.
Figure 7. (a) The agent’s egocentric clean RGB observation. (b) The agent’s egocentric RGB observation in the presence of visual corruption.
Sensors 23 03553 g007
Figure 8. RDMAE reconstruction visualization. (ad) are from four different scenes in the RoboTHOR validation set. For each image from left to right are: the original image, the Gaussian noise image (from top to bottom noise level σ = 0.5, 0.25, 0.1), the masked Gaussian noise image, and the reconstructed image.
Figure 8. RDMAE reconstruction visualization. (ad) are from four different scenes in the RoboTHOR validation set. For each image from left to right are: the original image, the Gaussian noise image (from top to bottom noise level σ = 0.5, 0.25, 0.1), the masked Gaussian noise image, and the reconstructed image.
Sensors 23 03553 g008
Figure 9. Examples of evaluated navigation trajectory visualization of RDMAE-Nav, ROBUSTNAV, and ROBUSTNAV+RP+SS-Adapt under Spatter.
Figure 9. Examples of evaluated navigation trajectory visualization of RDMAE-Nav, ROBUSTNAV, and ROBUSTNAV+RP+SS-Adapt under Spatter.
Sensors 23 03553 g009
Figure 10. Histogram of contributions of KL Loss on the navigation performance of RDMAE-Nav.
Figure 10. Histogram of contributions of KL Loss on the navigation performance of RDMAE-Nav.
Sensors 23 03553 g010
Figure 11. Histogram of contributions of Regularized Denoising on the navigation performance of RDMAE-Nav.
Figure 11. Histogram of contributions of Regularized Denoising on the navigation performance of RDMAE-Nav.
Sensors 23 03553 g011
Table 1. Navigation performance comparison under clean and visual corruption. The comparative performance metrics here are SR and SPL, with higher SR and SPL reflecting the higher effectiveness and efficiency of the navigation method.
Table 1. Navigation performance comparison under clean and visual corruption. The comparative performance metrics here are SR and SPL, with higher SR and SPL reflecting the higher effectiveness and efficiency of the navigation method.
Approach Visual Corruption
CleanSpatterSpeckle NoiseCamera CrackLower-FOVDefocus BlurMotion Blur
ROBUSTNAVSR98.8233.5867.4282.0742.4975.8995.72
SPL83.1324.7248.5763.8331.7353.5573.37
ROBUSTNAV+APSR98.4520.3865.6172.7045.6883.3594.81
SPL83.2815.7047.0356.8235.1461.5174.3
ROBUSTNAV+AP+SS-AdaptSR37.3114.19\57.8732.9440.95\
SPL31.0310.29\46.7226.0933.35\
ROBUSTNAV+RPSR98.7323.4878.9867.0644.9532.2191.63
SPL82.5318.6355.9253.7032.7422.4765.27
ROBUSTNAV+RP+SS-AdaptSR94.6361.06\60.4250.5979.16\
SPL77.2547.16\49.3736.1062.74\
ROBUSTNAV+Data AugSR98.4523.9377.2588.4471.7081.2696.91
SPL81.0818.4157.9571.5754.5461.3275.97
RDMAE-NavSR94.0989.2681.2683.0854.4192.8193.45
SPL74.6470.8464.3665.2740.5872.2973.22
Table 2. Agent behavior analysis of the proposed RDMAE-Nav by comparing R and Dist metric. In general, if the agent obtains a higher R and a lower Dist, the agent achieves more reasonable behavior.
Table 2. Agent behavior analysis of the proposed RDMAE-Nav by comparing R and Dist metric. In general, if the agent obtains a higher R and a lower Dist, the agent achieves more reasonable behavior.
Approach Visual Corruption
CleanSpatterSpeckle NoiseCamera CrackLower-FOVDefocus BlurMotion Blur
ROBUSTNAVR9.5130.4583.5775.0432.3215.4258.017
Dist0.13933.3062.21.5542.1151.4680.6826
ROBUSTNAV+APR9.51−0.084345.2176.0473.0537.3498.937
Dist0.13163.6731.6521.2892.1150.87850.349
ROBSUTNAV+AP+SS-AdaptR1.478−1.193\4.1751.1246.834\
Dist2.5913.397\1.9713.0821.126\
ROBUSTNAV+RPR9.6020.15466.8165.292.8431.1658.411
Dist0.098484.0521.1641.652.312.9610.4539
ROBUSTNAV+RP+SS-AdaptR9.0465.014\4.5063.5246.741\
Dist0.32481.838\1.881.9651.136\
ROBUSTNAV+Data AugR9.4650.085646.6628.116.0066.7229.183
Dist0.15313.8111.0770.58681.2451.0150.1957
RDMAE-NavR8.8958.2387.2537.4593.798.6878.778
Dist0.34580.55020.97840.90051.9820.39050.3561
Table 3. Detail of contributions of KL Loss on the navigation performance of RDMAE-Nav.
Table 3. Detail of contributions of KL Loss on the navigation performance of RDMAE-Nav.
Approach Visual Corruption
CleanSpatterSpeckle NoiseCamera CrackLower-FOVDefocus BlurMotion Blur
RDMAE-Nav
w/o KL Loss
SR89.0878.1657.7867.1545.1385.8186.81
SPL69.366042.7350.8932.6265.6566.91
RDMAE-NavSR94.0989.2681.2683.0854.4192.8193.45
SPL74.6470.8464.3665.2740.5872.2973.22
Table 4. Contributions of KL Loss on the navigation behavior of RDMAE-Nav by comparing R and Dist metric.
Table 4. Contributions of KL Loss on the navigation behavior of RDMAE-Nav by comparing R and Dist metric.
Approach Visual Corruption
CleanSpatterSpeckle NoiseCamera CrackLower-FOVDefocus BlurMotion Blur
RDMAE-Nav
w/o KL Loss
R8.1986.8094.2295.4192.6167.7867.906
Dist0.55811.0222.0131.5312.3660.76010.7233
RDMAE-NavR8.8958.2387.2537.4593.798.6878.778
Dist0.34580.55020.97840.90051.9820.39050.3561
Table 5. Detail of contributions of Regularized Denoising on navigation performance of RDMAE-Nav.
Table 5. Detail of contributions of Regularized Denoising on navigation performance of RDMAE-Nav.
Approach Visual Corruption
CleanSpatterSpeckle NoiseCamera CrackLower-FOVDefocus BlurMotion Blur
RDMAE-Nav
w/o Regularized Denoising
SR92.368.55310.148.9531.5767.4279.34
SPL74.945.9237.65637.6723.9951.9560.14
RDMAE-NavSR94.0989.2681.2683.0854.4192.8193.45
SPL74.6470.8464.3665.2740.5872.2973.22
Table 6. Contributions of Regularized Denoising on the navigation behavior of RDMAE-Nav by comparing R and Dist metric.
Table 6. Contributions of Regularized Denoising on the navigation behavior of RDMAE-Nav by comparing R and Dist metric.
Approach Visual Corruption
CleanSpatterSpeckle NoiseCamera CrackLower-FOVDefocus BlurMotion Blur
RDMAE-Nav
w/o Regularized Denoising
R8.678−1.909−1.7153.1660.96925.4656.941
Dist0.39383.6713.442.4282.8531.5831.096
RDMAE-NavR8.8958.2387.2537.4593.798.6878.778
Dist0.34580.55020.97840.90051.9820.39050.3561
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peng, J.; Xu, Y.; Luo, L.; Liu, H.; Lu, K.; Liu, J. Regularized Denoising Masked Visual Pretraining for Robust Embodied PointGoal Navigation. Sensors 2023, 23, 3553. https://doi.org/10.3390/s23073553

AMA Style

Peng J, Xu Y, Luo L, Liu H, Lu K, Liu J. Regularized Denoising Masked Visual Pretraining for Robust Embodied PointGoal Navigation. Sensors. 2023; 23(7):3553. https://doi.org/10.3390/s23073553

Chicago/Turabian Style

Peng, Jie, Yangbin Xu, Luqing Luo, Haiyang Liu, Kaiqiang Lu, and Jian Liu. 2023. "Regularized Denoising Masked Visual Pretraining for Robust Embodied PointGoal Navigation" Sensors 23, no. 7: 3553. https://doi.org/10.3390/s23073553

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop