Next Article in Journal
A Noisy Fractional Brownian Motion Model for Multiscale Correlation Analysis of High-Frequency Prices
Previous Article in Journal
The Mass Formula for Self-Orthogonal and Self-Dual Codes over a Non-Unitary Non-Commutative Ring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Modulated Memory Network for Video Object Segmentation

Faculty of Computing, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(6), 863; https://doi.org/10.3390/math12060863
Submission received: 16 February 2024 / Revised: 2 March 2024 / Accepted: 11 March 2024 / Published: 15 March 2024

Abstract

:
Existing video object segmentation (VOS) methods based on matching techniques commonly employ a reference set comprising historical segmented frames, referred to as ‘memory frames’, to facilitate the segmentation process. However, these methods suffer from the following limitations: (i) Inherent segmentation errors in memory frames can propagate and accumulate errors when utilized as templates for subsequent segmentation. (ii) The non-local matching technique employed in top-leading solutions often fails to incorporate positional information, potentially leading to incorrect matching. In this paper, we introduce the Modulated Memory Network (MMN) for VOS. Our MMN enhances matching-based VOS methods in the following ways: (i) Introducing an Importance Modulator, which adjusts memory frames using adaptive weight maps generated based on the segmentation confidence associated with each frame. (ii) Incorporating a Position Modulator that encodes spatial and temporal positional information for both memory frames and the current frame. The proposed modulator improves matching accuracy by embedding positional information. Meanwhile, the Importance Modulator mitigates error propagation and accumulation by incorporating confidence-based modulation. Through extensive experimentation, we demonstrate the effectiveness of our proposed MMN, which also achieves promising performance on VOS benchmarks.

1. Introduction

Video object segmentation (VOS) is an important research area within the domain of video analysis, with the potential to yield significant advantages across various computer vision applications, including but not limited to video editing and autonomous driving. The primary focus of this research is one-shot VOS, a task aimed at the continuous segmentation of specific object(s) throughout an entire video sequence, with pixel-wise mask annotations for the object(s) of the first frame provided.
Recent developments in VOS techniques have explored template matching methods, using a reference set of historical frames and their corresponding object masks. This reference set serves as the foundation for propagating labels to subsequent frames through pixel-wise matching. Notably, by introducing memory networks [1,2] within the Spatiotemporal Memory Network (STM) framework [3], a spatiotemporal memory is initially constructed for segmentation, allowing frames with intermediate predictions to be periodically updated to memory. The majority of current top-performing one-shot VOS methods [4,5,6,7,8,9,10] are based on the STM architecture, exploiting its ability to utilize comprehensive reference information from distant temporal frames. However, these matching-based methods still suffer from two significant challenges: (i) Inherent errors present in intermediate predictions may result in error propagation and accumulation when utilizing segmented frames as references. (ii) The widely employed non-local matching technique, prevalent in the aforementioned approaches, lacks the incorporation of positional information, potentially leading to incorrect matching.
In this study, we introduce the Modulated Memory Network (MMN), integrating two key components, namely, the Importance Modulator (IM) and the Position Modulator (PM), which are designed to tackle the aforementioned challenges. As illustrated in Figure 1, the IM module performs pixel-level modulation on the features of historical frames within the memory, utilizing adaptive weights that operate across both spatial and channel dimensions. These modulation weights are generated based on the retrieval from memory and the segmentation confidence associated with historical frames, serving as crucial conditions. Through adaptive feature modulation, incorrectly segmented pixels within the memory exert less influence on subsequent segmentation, effectively mitigating the propagation and accumulation of errors. Simultaneously, the PM module incorporates positional embedding into the features of both query and memory frames, encoding spatiotemporal coordinates. This positional embedding enhances the robustness of memory retrieval in the presence of distractions. Furthermore, among all historical frames, the initial frame and the last frame hold greater significance. The initial frame carries ground-truth annotations, while the last frame provides substantial spatial guidance. A toy experiment, as displayed in Table 1, demonstrates a more pronounced performance drop when removing these two frames from memory, as opposed to removing other historical frames. To optimize the utilization of the initial and the most recent frames, their temporal positions are given a special encoding.
The primary contributions of this study can be summarized as follows:
  • We introduce the Modulated Memory Network (MMN) for video object segmentation (VOS), incorporating both an Importance Modulator (IM) and a Position Modulator (PM). The MMN, as proposed, conducts adaptive memory modulation that takes into account segmentation confidence and spatiotemporal positional information. This enhancement results in improved object segmentation accuracy.
  • The IM module, designed to execute pixel-level modulation, utilizes adaptive weights generated based on segmentation confidence as conditions. Simultaneously, the PM module is implemented to embed positional information into both query and memory frames, thereby enhancing matching accuracy.
  • Our proposed Modulated Memory Network (MMN) consistently achieves performance comparable to state-of-the-art methods for established VOS benchmarks. Furthermore, ablation studies provide additional evidence of the effectiveness of the IM and PM modules.

2. Related Work

According to task specifications, the video object segmentation (VOS) task can be categorized into zero-shot VOS, one-shot VOS and interactive VOS. Zero-shot VOS [12,13,14,15,16,17,18,19,20,21,22,23,24,25,26] involves fully automated segmentation of salient objects in a video without any manual initialization. One-shot VOS [3,4,5,6,7,9,10,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41] focuses on segmenting a specified object, typically provided in the first frame. Interactive VOS [42,43,44,45,46,47] incorporates manual interaction steps. This paper addresses the relatively foundational one-shot VOS task.
Deep learning-based methods for one-shot VOS can be categorized into three distinct groups, based on their utilization of the first annotated frame during inference, as outlined by Wang et al. [48]. These categories include online fine-tuning-based methods, motion-based methods and matching-based methods. Online fine-tuning-based methods, exemplified by works such as OSVOS [29], OSVOS-S [30] and OnAVOS [38], based on the idea of image segmentation [49,50,51], involve the fine-tuning of general segmentation networks on the initial annotated frame to target specific objects. In order to address the evolving appearance of objects over time, these methods frequently conduct online updates of their segmentation models. However, this online training incurs a notable increase in testing time. In contrast, motion-based approaches, represented by works like MaskRNN [52], MaskTrack [53] and PReMVOS [39], leverage temporal continuity. These methods estimate pixel-wise motion to transform object masks from the first annotated frame to subsequent frames. Although motion-based techniques offer computational efficiency, they are susceptible to object drifts caused by occlusions and rapid object movements.
Matching-based one-shot VOS: In recent times, matching-based methods have gained prominence in video tasks [54] due to their resilience in handling occlusions and rapid object motion. In the one-shot VOS task, this category of approaches formulates the task as a template matching problem, initializing a reference set with historical frames and their respective masks. Subsequently, labels are propagated from this reference set to the current frame (the query frame) through pixel-wise matching and retrieval.
Early efforts [37,55] incorporated siamese structure networks to extract features from both the reference and query frames. Notably, VideoMatch [37] introduced the practice of updating the reference set with segmented frames. Meanwhile, OSMN [28] introduced modulators to adapt a generic segmentation network to specific objects without resorting to online fine-tuning. Under the assumption of temporal continuity, FEELVOS [35] narrowed the matching area using the last frame as a local range. Building on FEELVOS [35], CFBI [33] further enhanced the approach by establishing separate reference sets for the foreground and background.
Leveraging memory networks [1,2,56], STM [3] constructed an extensive memory encompassing pixel-level features from the reference set. The success of STM [3] significantly influenced the development of matching-based solutions. EGMN [6] introduced a graph-structured memory network, performing memory reading and writing on each node sequentially. To enhance inference efficiency, dynamic memory update strategies were explored. For instance, GCM [4] maintained a fixed memory size, updating each computed result to memory with equal weight. SwiftNet [8] triggered memory updates based on cumulative differences across frames. In order to remove obsolete features, AFB-URR [7] maintained a least-frequently used index for each feature. However, the importance of each reference feature needed to be distinguished before use, as they originated from pixels with varying levels of segmentation confidence. In our work, we propose the Modulated Memory Network, where an Importance Modulator is incorporated for reference feature modulation.
To mitigate the ambiguity of similar objects, techniques were devised to impose constraints on the retrieval of reference information. KMN [5], for instance, enhances memory reading operations with Gaussian kernels to reduce non-localization issues. RMNet [10] restricts matching to specific regions by employing bounding boxes around objects. Meanwhile, LCM [9] reinforces potential object regions with an additional position guidance module. However, these approaches do not fully harness spatial and temporal positional information nor do they explore their interconnection.
Position encoding: To compensate for the limitations of permutation-invariant architectures, Transformer [57] introduced 1D positional information encoding through positional embedding. Further advancements can be observed in approaches like DETR [58] and VisTR [57], which apply positional embedding to multi-dimensional data encompassing images and videos. In these methods, positional information for each dimension is encoded separately and then concatenated. Despite these explorations, it is believed that the positional encoding technique can be refined further by capitalizing on the inherent characteristics of a one-shot VOS task.

3. Method

3.1. Overview

As shown in Figure 2, given a video sequence S = { I i } i = 1 T , where the initial frame I 1 is provided with an associated mask P 1 , the objective of one-shot VOS is to segment the designated object(s) from the background by predicting corresponding masks P = { P i } i = 2 T . Following the paradigm of matching-based VOS models, MMN processes each frame in a sequential manner.
The segmentation process can be divided into two fundamental steps, namely, memorization and segmentation, which are carried out iteratively. In the memorization step, our Modulated Memory Network (MMN) constructs a modulated memory. Within this memory, features from historical frames are extracted and subsequently modulated by the Importance Modulator (IM). In the segmentation step, the current frames retrieve supportive features from the modulated memory through pixel-wise matching. This process embeds spatiotemporal information into the features of both the current frames and the frames stored in memory, facilitated by the Position Modulator (PM). Ultimately, a decoder is employed to transform the memory retrieval into an object mask. In the subsequent subsections, we delve into the specifics of these two steps and elucidate the architecture of the proposed modules.

3.2. Memorization

In memorization, a modulated memory is constructed for the subsequent segmentation using historical frames and their corresponding masks. The historical frames are encoded and stored in the memory sequentially. As shown in Figure 3, for historical frame I i and its object mask P i , a memory encoder E n M is first applied, which takes input of both I i and P i and output key feature K M i R H W × C k and value feature V M i R H W × C v . C k and C v are the channel dimensions of the key and value features, which are 128 and 512 in our MMN. Spatial dimensions are H = s H i m and W = s W i m , where H i m and W i m correspond to the spatial dimension of the input frames. s corresponds to the stride of the backbone in our MMN, which is 16 since we use ResNet-50 [59] as the backbone following STM [3]. Previous solutions [3,5,10] store K M i and V M i to memory directly. However, for the inherent errors in the intermediate mask predictions, segmenting with K M i and V M i may lead to error propagation and accumulation. To address this issue, we construct a modulated memory with the proposed IM module, which modulates each historical frame, respectively, considering their segmentation reliability and context importance. In particular, the IM module takes as input the value feature V M i , a readout feature R M i and the object mask P i , and it outputs a weight map A i . To obtain R M i , the features of { I j } j = 1 i 1 , i.e., key features K M i 1 = { K M j } j = 1 i 1 and value features V M i 1 = { V M j } j = 1 i 1 , are taken as a reference. An affinity matrix A i is first calculated between the key features. For each spatial location p in K M i and q in K M i 1 = { K M j } j = 1 i 1 ,
A ( p , q ) = exp ( S p , q ) q exp ( S p , q )
Function S calculates the similarity, which is calculated as follows:
S p , q = C ( K M i ( p ) , K M i 1 ( q ) )
C is the dot production function as in STM [3]. Then, R M i is aggregated through the weighted sum of value features V M i 1 = { V M j } j = 1 i 1 ,
R M i ( p ) = q ( A ( p , q ) · V M i 1 ( q ) )
The IM module can learn to estimate the reliability of frame I i by comparing R M i with V M i . Afterward, we adopt E n P to encode P i into feature P E i , of which the number of channels is 32. Then, we concatenate P E i 1 , R M i and V M i together along the channel dimension, and we input it into an estimator E s t and output the modulation map A i . Finally, modulation for frame I i is conducted as follows, for each position p in V M i :
V ^ M i ( p ) = A ( p ) · V M i ( p )
The modulation is only conducted on value features, which mainly carry the label information of the target object. We provide experimental results in Section 4.3, comparing different modulation modes. After the modulation, the modulated features of the historical frames { I j } j = 1 i are stacked together along the temporal dimension, constructing a modulated memory M i = { K M R N H W × C k , V ^ M R N H W × C v } .

3.3. Segmentation

In segmentation, MMN segments the current frame I t using the modulated memory M t 1 , where support features are retrieved through matching. To improve matching accuracy, the Position Modulator (PM) is deployed to encode spatiotemporal positional information for the key features of both memory frames and the current frame, which also helps to better exploit the two vital frames (i.e., the first and the last frames). Here, we adapt a positional encoding in VisTR [57] for the VOS task. In particular, coordinates in each dimension (i.e., height, width and temporal dimensions) are first normalized and encoded into embeddings using sine and cosine functions with different frequencies:
P E ( p , c ) = sin ( Norm ( p ) · ω k ) , for c = 2 k cos ( Norm ( p ) · ω k ) , for c = 2 k + 1
where p is the coordinate in the corresponding dimension, ω k = 1/ 10,000 2 k / d 3 and c is the channel index. Norm is a normalization function. The spatial coordinates (i.e., height and width dimensions) are sequentially numbered, and their normalization function is defined as
Norm R ( p ) = p max p P ( ( p ) )
As for the temporal dimension, to better exploit the two vital frames, we specially define temporal coordinates as P t = { 2 , 3 , 4 , , t 1 , 1 , 1 } for each video sequence S = { I 1 , I 2 , I 3 , , I t 2 , I t 1 , I t } . The normalization function for temporal coordinates is defined as
Norm R ( p ) = 1 p , p P t
The customized encoding scheme fixes the relative positional relationship between the current frame and the two vital frames, which makes it possible to approach important reference information according to their temporal positions. Positional embeddings for each dimension are concatenated together along the channel dimension and added to the key features K M and K Q t element-wise, obtaining modulated key features K ^ M and K ^ Q t . K Q t R H W × C k is the key feature of the current frame I t , encoded by E n Q .
After obtaining modulated key features, the affinity matrix can be calculated the same as Equation (1). Then, support features are retrieved from the modulated memory through weighted aggregation as in Equation (4). Finally, the mask prediction of frame I t is generated through decoding the retrieved support features by the decoder Dec .

3.4. Network Architecture and Loss Function

Same as the baseline work [3], both encoders EnM and EnQ use ResNet-50 [59] as backbones, and the output of res-4 is used as a feature map. The parameters of EnM and EnQ are not shared. Two convolution layers are adopted to transform the output of layer res-4 into key and value features. The decoder Dec shares the same architecture as that of STM [3], which is composed of several residual blocks and bilinear interpolation layers for upsampling.
The IM module has two learnable sub-modules, i.e., encoder EnP and estimator Est . The encoder EnP consists of 3 convolution layers. The first two are 4 × 4 convolution layers with stride 4, and the number of channels is 64. The last is a 1 × 1 convolution layer with stride 1, and the number of channels is 32. The output resolution is 1 16 of the input mask resolution. The Est is constructed with three 1 × 1 convolution layers with stride 1. The channel numbers of these three convolution layers are 1024, 512 and 1, respectively. Both EnP and Est activate the first two layers with ReLU [60].
We train our MMN with cross-entropy loss and mask IoU loss [61], where the mask IoU loss is defined as
L m I o U ( P i , G i ) = 1 p Ω min ( P i p , G i p ) p Ω max ( P i p , G i p )
where P and G are the predicted mask and ground-truth mask of object i and Ω represents all pixels in masks P and G.

3.5. Algorithm

The workflow of the proposed MMN is illustrated in Algorithm 1. The algorithm takes as input a sequence of video frames { I 1 , I 2 , , I T } and a given ground-truth mask P 1 . The algorithm initializes memory using the first frame I 1 and the given mask P 1 and then alternates between the segmentation and memorization processes until it obtains segmentation results { P 2 , , P T } for frames { I 2 , , I T } .
Algorithm 1 Pseudo-code for Modulated Memory Network
1: Input: Frames { I 1 , I 2 , , I T } , ground-truth mask P 1 .
2: Output: Predicted mask { P 2 , , P T } .
  
3: ▹ Memory initialization
4: K 1 M , V 1 M = EnK ( ( I 1 , P 1 ) )
5: K M = { K 1 M } , V M = { V 1 M }
6: for  i = 1 : T  do
7: ▹ Segmentation
8:      K Q , V Q = EnQ ( I i )
  
9: ▹ Position modulation
10:      P E s p a t i a l = PosEncoding ( H K M , W K M ) # shape= ( H , W , 1 )
11:      P E t e m p o r a l = 1 range ( length ( M ) + 1 ) # shape= ( 1 , 1 , i + 1 )
12:      P E s p a t i a l = expand ( P E s p a t i a l , ( H , W , i + 1 ) )
13:      P E t e m p o r a l = expand ( P E t e m p o r a l , ( H , W , i + 1 ) )
14:      P E = Concat ( [ P E s p a t i a l , P E t e m p o r a l ] )
15:      K M = K M + P E [ : 1 ]
16:      K Q = K Q + P E [ 1 ]
  
17: ▹ Memory retrieval
18:      R Q = ( K Q K M ) V M
   
19: ▹ Decode
20:      P i = Dec ( R Q , V Q )
  
21: ▹ Memorization
22: ▹ Weight map generation
23:      K i M , V i M = EnK ( ( I i , P i ) )
24:      P i E = EnP ( P i )
25:      R = ( K i M , K M ) V M
26:      A i = Est ( [ V i M , P i E , R ] )
  
27: ▹ Memory modulation
28:      V i M = V i M · A i
  
29: ▹ Memory update
30:      K M K M { K i M }
31:      V M V M { V i M }
32: end for

4. Experiments

4.1. Datasets and Evaluation

Consistent with common practices, we utilize the DAVIS 2016 [62], DAVIS 2017 [11], YouTube-VOS 2018 [63] and MOSE [64] datasets for training and evaluating the proposed MMN. The DAVIS datasets [11,62] provide high-quality object annotations. The 2016 version [62] comprises 30 videos for training and 20 for validation. Each video features a single annotated object instance in its initial frame. On the other hand, the 2017 version [11] serves as a multi-object extension, incorporating a larger number of videos in both training and validation splits. Additionally, the YouTube-VOS 2018 [63] dataset represents a formidable and expansive challenge in video object segmentation. The training set contains 3471 videos, while the validation set includes 474 videos. Among the videos in the validation set, 65 semantic categories overlap with those in the training set, forming the seen subset. In contrast, the remaining 26 semantic categories in the validation set differ from those in the training set, constituting the unseen subset. The MOSE [64] dataset is designed for complex scenes, which includes 1507 videos for training and 311 videos for validation. MOSE covers various challenging scenarios, such as interactions between multiple objects, occlusions and background interference.
Our evaluation of the MMN is conducted on the validation splits of all datasets. For DAVIS 2016 and 2017, we employ the standard evaluation toolkit provided by DAVIS 2017 [11]. This toolkit computes the mean of region similarity J , contour accuracy F and their average J & F for comparative analysis. In the case of YouTube-VOS 2018, we report J and F metrics separately for both the seen and unseen semantic categories, as well as the overall averaged score G . Metrics for YouTube-VOS and MOSE are all calculated through the competition server on Codalab.

4.2. Training and Inference

Following SwiftNet [8], we commence by pre-training our MMN on synthetic data [65], followed by fine-tuning using video data. During the pre-training stage, random affine transformations with varied parameters are applied to static images in order to generate synthetic video clips. We also employ a Cut and Paste strategy for augmenting images containing fewer than three target objects. This strategy involves cutting object instances from source images and pasting them into random positions within target images. The pre-training phase consists of a total of 6 × 10 5 iterations, employing a constant learning rate of 1 × 10 5 . The backbones of E n Q and E n M are initialized with weights pre-trained on ImageNet [66].
Upon completing the pre-training stage, the MMN undergoes fine-tuning on video datasets, with each training batch containing three frames randomly sampled from a video sequence. The sampling interval progressively increases from 1 to 25 during the first 2.5 × 10 4 iterations and subsequently decreases to 5 in the final 2 × 10 5 iterations. Random affine transformations, with distinct parameters for each frame, are applied throughout this process. The input patch size is set to 384 × 384 , and a batch size of 4 is achieved through gradient accumulation. The fine-tuning stage comprises a total of 8 × 10 5 iterations and employs a poly learning rate strategy. We use Adam [67] optimization with standard momentum and no weight decay for both the pre-training and fine-tuning stages. The MMN is trained on a Tesla V100 GPU and evaluated on a GTX 2080Ti GPU. GPUs are both manufactured by NVIDIA, located in Santa Clara, CA, USA.

4.3. Ablation Study

To demonstrate the effectiveness of each component of the proposed MMN, we conduct ablation experiments on the DAVIS 2017 validation set. For fast verification, models for ablation experiments are trained directly with the DAVIS and YouTube-VOS train set, without the pre-training stage.
Importance Modulator: The IM module is devised to modulate memory frames while considering their contextual importance and segmentation confidence. To assess the effectiveness of the proposed IM module, we compare models with and without the IM module. Furthermore, we explore various modulation approaches by comparing models where modulation is applied to the key features, value features and both key and value features. As depicted in Table 2, applying modulation to the value features yields the most favorable performance. This is attributed to the value feature primarily encoding information related to predicted masks, where modulation serves to prevent the propagation and accumulation of inherent errors. In contrast, the key feature primarily encodes appearance information used for similarity calculations. Modulating the key feature could potentially lead to incorrect object matching, thereby diminishing segmentation accuracy.
We also present qualitative results in Figure 4. It is evident that, in the absence of the IM module, segmentation errors within the memory frame (highlighted in the red box) propagate to subsequent frames (21 and 22). However, with the utilization of the IM module, adaptive modulation is applied to the inaccurately segmented region, effectively halting the propagation of errors.
Positional Modulator: The PM module is devised to incorporate spatiotemporal positional information into the segmentation process. As depicted in Table 3, we initially explore the effectiveness of integrating spatiotemporal information within the VOS task. This is demonstrated by comparing models # 1 , # 2 and # 5 . Among the models that incorporate positional information (i.e., # 2 and # 5 ), model # 5 , which integrates both spatial and temporal information, outperforms model # 2 , which solely utilizes spatial positional information. Additionally, all models incorporating positional information ( # 2 and # 5 ) achieve higher accuracy than models lacking positional information ( # 1 ).
Subsequently, we delve into several design choices for the PM module. In our MMN, special consideration is given to the design of temporal coordinates and the corresponding normalization function. For comparison, we train model # 3 , where temporal coordinates are sequentially numbered, similar to the approach in VisTR [57]. Model # 4 replaces the normalization function of our MMN with the commonly used Norm M , as defined in Equation (6). As evidenced by Table 3, replacing either the temporal coordinates numbering scheme or the normalization function results in a degradation of segmentation accuracy. We conjecture that the combination of the coordinate numbering scheme and normalization using Equation (7) enables better utilization of information contained in two critical frames, namely, the first frame and the last frame.

4.4. Comparison with State of the Art

DAVIS: The quantitative results of DAVIS 2017 and 2016 [11,62] are shown in Table 4 and Table 5, respectively. On DAVIS 2016, the proposed MMN achieves a J & F score of 91.2% on the DAVIS 2016 validation set, which is comparable to other semi-supervised VOS solutions. On DAVIS 2017, which is multi-object-annotated, our MMN achieves a J & F score of 85.8%, which is comparable to other competitive state-of-the-art methods. From the table, it can be observed that the proposed MMN outperforms the baseline STM [3] by 3.6 in terms of region similarity score J , 4.5 in terms of contour accuracy score F and 4.0 points in terms of average score J & F . This is attributed to the MMN’s adoption of the memory modulation mechanism, resulting in improvements in both pixel-level accuracy and boundary precision. Since the contour accuracy score F emphasizes boundary precision, the MMN demonstrates more significant improvements in this aspect. We also present qualitative results in Figure 5. Due to the modulated memory, the MMN manages to segment the challenging tail area of the motorcycle, while the baseline method STM [3] accumulates the segmentation error throughout the video.
YouTube-VOS: Due to the large amount of videos, the YouTube-VOS benchmark poses a huge challenge to VOS solutions. The quantitative results of the YouTube-VOS 2018 validation set are presented in Table 6, where the proposed MMN performs favorably against the state-of-the-art methods. From the table, it can be observed that the proposed MMN outperforms the baseline method STM [3] in terms of J score by 1.5 and F score by 2.3 for the seen category. For the unseen category, the MMN achieves an improvement of 4.8 in J score and 5.3 in F score. It benefits from the Position Modulator in the MMN, which enhances the positional information, thereby improving its performance particularly when encountering unseen object categories during training.
A visual comparison of segmentation results on YouTube-VOS video 37dc952545 is also provided in Figure 5. STM [3] loses track when the target head reappears at frame 90, while STCN [27] is disturbed by similar objects from frame 40. In contrast, with the proposed PM module, our MMN succeeds in relocating the target head at the frame, while avoiding interference by the background.
MOSE: The MOSE [64] dataset presents a unique challenge due to its more intricate scenes and extended video lengths, which demand robust segmentation algorithms capable of maintaining accuracy over prolonged durations. While our MMN surpasses the baseline method STM [3] on MOSE, it lags behind DEVA’s [70] results (Table 7). We attribute this discrepancy to the integration of an LSTM module designed explicitly for handling longer video sequences in DEVA. To address this, future improvements could involve augmenting the MMN architecture with a dedicated LSTM module tailored specifically for handling long video sequences.

5. Conclusions

This study presented the Modulated Memory Network (MMN) for the challenging task of one-shot video object segmentation. Our approach introduces two key components: the Importance Modulator (IM) and the Position Modulator (PM). The IM module plays a pivotal role in constructing an informative memory. It modulates memory embeddings while taking into account their contextual importance and segmentation confidence. By doing so, it enhances the quality of the memory and significantly reduces the propagation of errors during the segmentation process. Moreover, the PM module enhanced the way we incorporate spatiotemporal information into the segmentation process. Through a tailored encoding scheme, it empowers both memory and query embeddings with vital spatiotemporal information. This enhancement reinforces the accuracy of information retrieval, resulting in more precise and robust object segmentation. The experimental results prove that the proposed MMN performs favorably against the state-of-the-art methods on YouTube-VOS while achieving top-rank performance on the DAVIS validation split.

Author Contributions

Conceptualization, H.L. and W.Z.; methodology, H.L., Z.G. and W.Z.; software, H.L.; validation, H.L. and Z.G.; formal analysis, H.L., Z.G. and W.Z.; investigation, H.L.; resources, H.L.; data curation, H.L.; writing—original draft preparation, H.L.; writing—review and editing, Z.G. and W.Z.; visualization, H.L. and Z.G.; supervision, W.Z.; project administration, W.Z.; funding acquisition, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Key Research and Development Program of China under grant no. 2021ZD0112100.

Data Availability Statement

COCO [65] is available at https://cocodataset.org/, YouTube-VOS [63] is available at https://youtube-vos.org/, and DAVIS [11] is available at https://davischallenge.org/ (accessed on 9 September 2018).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Weston, J.; Chopra, S.; Bordes, A. Memory networks. arXiv 2014, arXiv:1410.3916. [Google Scholar]
  2. Graves, A.; Wayne, G.; Danihelka, I. Neural turing machines. arXiv 2014, arXiv:1410.5401. [Google Scholar]
  3. Oh, S.W.; Lee, J.Y.; Xu, N.; Kim, S.J. Video object segmentation using space-time memory networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9226–9235. [Google Scholar]
  4. Li, Y.; Shen, Z.; Shan, Y. Fast Video Object Segmentation using the Global Context Module. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 735–750. [Google Scholar]
  5. Seong, H.; Hyun, J.; Kim, E. Kernelized Memory Network for Video Object Segmentation. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 629–645. [Google Scholar]
  6. Lu, X.; Wang, W.; Danelljan, M.; Zhou, T.; Shen, J.; Van Gool, L. Video Object Segmentation with Episodic Graph Memory Networks. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; Volume 12348, pp. 661–679. [Google Scholar]
  7. Liang, Y.; Li, X.; Jafari, N.; Chen, J. Video Object Segmentation with Adaptive Feature Bank and Uncertain-Region Refinement. In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 6–12 December 2020; Curran Associates, Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 3430–3441. [Google Scholar]
  8. Wang, H.; Jiang, X.; Ren, H.; Hu, Y.; Bai, S. SwiftNet: Real-time Video Object Segmentation. arXiv 2021, arXiv:2102.04604. [Google Scholar]
  9. Hu, L.; Zhang, P.; Zhang, B.; Pan, P.; Xu, Y.; Jin, R. Learning Position and Target Consistency for Memory-based Video Object Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 4144–4154. [Google Scholar]
  10. Xie, H.; Yao, H.; Zhou, S.; Zhang, S.; Sun, W. Efficient Regional Memory Network for Video Object Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  11. Pont-Tuset, J.; Perazzi, F.; Caelles, S.; Arbelaez, P.; Sorkine-Hornung, A.; Gool, L.V. The 2017 DAVIS Challenge on Video Object Segmentation. arXiv 2017, arXiv:1704.00675. [Google Scholar]
  12. Wang, W.; Shen, J.; Shao, L. Video salient object detection via fully convolutional networks. IEEE Trans. Image Process. 2017, 27, 38–49. [Google Scholar] [CrossRef] [PubMed]
  13. Yang, C.; Lamdouar, H.; Lu, E.; Zisserman, A.; Xie, W. Self-supervised video object segmentation by motion grouping. In Proceedings of the the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 7177–7188. [Google Scholar]
  14. Zhang, K.; Zhao, Z.; Liu, D.; Liu, Q.; Liu, B. Deep transport network for unsupervised video object segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 8781–8790. [Google Scholar]
  15. Fragkiadaki, K.; Arbelaez, P.; Felsen, P.; Malik, J. Learning to segment moving objects in videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4083–4090. [Google Scholar]
  16. Lu, X.; Wang, W.; Ma, C.; Shen, J.; Shao, L.; Porikli, F. See more, know more: Unsupervised video object segmentation with co-attention siamese networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3623–3632. [Google Scholar]
  17. Lee, M.; Cho, S.; Lee, S.; Park, C.; Lee, S. Unsupervised video object segmentation via prototype memory network. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 5924–5934. [Google Scholar]
  18. Cho, S.; Lee, M.; Lee, S.; Park, C.; Kim, D.; Lee, S. Treating motion as option to reduce motion dependency in unsupervised video object segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 5140–5149. [Google Scholar]
  19. Ponimatkin, G.; Samet, N.; Xiao, Y.; Du, Y.; Marlet, R.; Lepetit, V. A simple and powerful global optimization for unsupervised video object segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 5892–5903. [Google Scholar]
  20. Garg, S.; Goel, V. Mask selection and propagation for unsupervised video object segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2021; pp. 1680–1690. [Google Scholar]
  21. Ren, S.; Liu, W.; Liu, Y.; Chen, H.; Han, G.; He, S. Reciprocal transformations for unsupervised video object segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15455–15464. [Google Scholar]
  22. Lu, X.; Wang, W.; Shen, J.; Crandall, D.; Luo, J. Zero-shot video object segmentation with co-attention siamese networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 2228–2242. [Google Scholar] [CrossRef] [PubMed]
  23. Wang, W.; Lu, X.; Shen, J.; Crandall, D.J.; Shao, L. Zero-shot video object segmentation via attentive graph neural networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9236–9245. [Google Scholar]
  24. Yang, S.; Zhang, L.; Qi, J.; Lu, H.; Wang, S.; Zhang, X. Learning motion-appearance co-attention for zero-shot video object segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1564–1573. [Google Scholar]
  25. Zhen, M.; Li, S.; Zhou, L.; Shang, J.; Feng, H.; Fang, T.; Quan, L. Learning discriminative feature with crf for unsupervised video object segmentation. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXVII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 445–462. [Google Scholar]
  26. Gu, Y.; Wang, L.; Wang, Z.; Liu, Y.; Cheng, M.M.; Lu, S.P. Pyramid constrained self-attention network for fast video salient object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 10869–10876. [Google Scholar]
  27. Cheng, H.K.; Tai, Y.W.; Tang, C.K. Rethinking Space-Time Networks with Improved Memory Coverage for Efficient Video Object Segmentation. Adv. Neural Inf. Process. Syst. 2021, 34, 11781–11794. [Google Scholar]
  28. Yang, L.; Wang, Y.; Xiong, X.; Yang, J.; Katsaggelos, A.K. Efficient Video Object Segmentation via Network Modulation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  29. Caelles, S.; Maninis, K.K.; Pont-Tuset, J.; Leal-Taixe, L.; Cremers, D.; Van Gool, L. One-Shot Video Object Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  30. Maninis, K.K.; Caelles, S.; Chen, Y.; Pont-Tuset, J.; Leal-Taixé, L.; Cremers, D.; Van Gool, L. Video Object Segmentation without Temporal Information. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 1515–1530. [Google Scholar] [CrossRef] [PubMed]
  31. Cho, S.; Lee, H.; Kim, M.; Jang, S.; Lee, S. Pixel-level bijective matching for video object segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 129–138. [Google Scholar]
  32. Johnander, J.; Danelljan, M.; Brissman, E.; Khan, F.S.; Felsberg, M. A generative appearance model for end-to-end video object segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 8953–8962. [Google Scholar]
  33. Yang, Z.; Wei, Y.; Yang, Y. Collaborative video object segmentation by foreground-background integration. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020. [Google Scholar]
  34. Bao, L.; Wu, B.; Liu, W. CNN in MRF: Video Object Segmentation via Inference in a CNN-Based Higher-Order Spatio-Temporal MRF. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5977–5986. [Google Scholar] [CrossRef]
  35. Voigtlaender, P.; Chai, Y.; Schroff, F.; Adam, H.; Leibe, B.; Chen, L.C. FEELVOS: Fast End-To-End Embedding Learning for Video Object Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  36. Khoreva, A.; Benenson, R.; Ilg, E.; Brox, T.; Schiele, B. Lucid Data Dreaming for Object Tracking. In Proceedings of the 2017 DAVIS Challenge on Video Object Segmentation—CVPR Workshops, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  37. Hu, Y.T.; Huang, J.B.; Schwing, A.G. Videomatch: Matching based video object segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 54–70. [Google Scholar]
  38. Voigtlaender, P.; Leibe, B. Online adaptation of convolutional neural networks for the 2017 davis challenge on video object segmentation. In Proceedings of the 2017 DAVIS Challenge on Video Object Segmentation—CVPR Workshops, Honolulu, HI, USA, 21–26 July 2017; Volume 5. [Google Scholar]
  39. Luiten, J.; Voigtlaender, P.; Leibe, B. PReMVOS: Proposal-generation, Refinement and Merging for Video Object Segmentation. In Proceedings of the Asian Conference on Computer Vision (ACCV), Perth, Australia, 2–6 December 2018. [Google Scholar]
  40. Li, M.; Hu, L.; Xiong, Z.; Zhang, B.; Pan, P.; Liu, D. Recurrent dynamic embedding for video object segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1332–1341. [Google Scholar]
  41. Oh, S.W.; Lee, J.Y.; Sunkavalli, K.; Kim, S.J. Fast Video Object Segmentation by Reference-Guided Mask Propagation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  42. Xu, N.; Price, B.; Cohen, S.; Yang, J.; Huang, T.S. Deep interactive object selection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 373–381. [Google Scholar]
  43. Cheng, H.K.; Tai, Y.W.; Tang, C.K. Modular interactive video object segmentation: Interaction-to-mask, propagation and difference-aware fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 5559–5568. [Google Scholar]
  44. Heo, Y.; Koh, Y.J.; Kim, C.S. Guided interactive video object segmentation using reliability-based attention maps. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7322–7330. [Google Scholar]
  45. Oh, S.W.; Lee, J.Y.; Xu, N.; Kim, S.J. Fast user-guided video object segmentation by interaction-and-propagation networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5247–5256. [Google Scholar]
  46. Heo, Y.; Jun Koh, Y.; Kim, C.S. Interactive video object segmentation using global and local transfer modules. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XVII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 297–313. [Google Scholar]
  47. Miao, J.; Wei, Y.; Yang, Y. Memory aggregation networks for efficient interactive video object segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10366–10375. [Google Scholar]
  48. Wang, W.; Zhou, T.; Porikli, F.; Crandall, D.; Van Gool, L. A Survey on Deep Learning Technique for Video Segmentation. arXiv 2021, arXiv:2107.01153. [Google Scholar]
  49. Li, Z.; Jiang, J.; Chen, X.; Laganière, R.; Li, Q.; Liu, M.; Qi, H.; Wang, Y.; Zhang, M. Dense-scale dynamic network with filter-varying atrous convolution for semantic segmentation. Appl. Intell. 2023, 53, 26810–26826. [Google Scholar] [CrossRef]
  50. Ou, X.; Wang, H.; Zhang, G.; Li, W.; Yu, S. Semantic segmentation based on double pyramid network with improved global attention mechanism. Appl. Intell. 2023, 53, 18898–18909. [Google Scholar] [CrossRef]
  51. Yu, Z.; Lee, F.; Chen, Q. HCT-net: Hybrid CNN-transformer model based on a neural architecture search network for medical image segmentation. Appl. Intell. 2023, 53, 19990–20006. [Google Scholar] [CrossRef]
  52. Hu, Y.T.; Huang, J.B.; Schwing, A. MaskRNN: Instance Level Video Object Segmentation. In Proceedings of the Advances in Neural Information Processing Systems; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  53. Perazzi, F.; Khoreva, A.; Benenson, R.; Schiele, B.; Sorkine-Hornung, A. Learning Video Object Segmentation from Static Images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3491–3500. [Google Scholar] [CrossRef]
  54. Yang, Y.; Gu, X. Learning rich feature representation and aggregation for accurate visual tracking. Appl. Intell. 2023, 53, 28114–28132. [Google Scholar] [CrossRef]
  55. Shin Yoon, J.; Rameau, F.; Kim, J.; Lee, S.; Shin, S.; So Kweon, I. Pixel-level matching for video object segmentation using convolutional neural networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2167–2176. [Google Scholar]
  56. Santoro, A.; Bartunov, S.; Botvinick, M.; Wierstra, D.; Lillicrap, T. Meta-learning with memory-augmented neural networks. In Proceedings of the International Conference on Machine Learning (ICML), PMLR, New York, NY, USA, 19–24 June 2016; pp. 1842–1850. [Google Scholar]
  57. Wang, Y.; Xu, Z.; Wang, X.; Shen, C.; Cheng, B.; Shen, H.; Xia, H. End-to-End Video Instance Segmentation with Transformers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  58. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 213–229. [Google Scholar]
  59. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  60. Glorot, X.; Bordes, A.; Bengio, Y. Deep Sparse Rectifier Neural Networks. In Proceedings of the 14th International Conference on Artificial Intelligence and Statistics (AISTATS), Fort Lauderdale, FL, USA, 11–13 April 2011; pp. 315–323. [Google Scholar]
  61. Milletari, F.; Navab, N.; Ahmadi, S.A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar] [CrossRef]
  62. Perazzi, F.; Pont-Tuset, J.; McWilliams, B.; Van Gool, L.; Gross, M.; Sorkine-Hornung, A. A Benchmark Dataset and Evaluation Methodology for Video Object Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 724–732. [Google Scholar] [CrossRef]
  63. Xu, N.; Yang, L.; Fan, Y.; Yang, J.; Yue, D.; Liang, Y.; Price, B.; Cohen, S.; Huang, T. Youtube-vos: Sequence-to-sequence video object segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 585–601. [Google Scholar]
  64. Ding, H.; Liu, C.; He, S.; Jiang, X.; Torr, P.H.; Bai, S. MOSE: A New Dataset for Video Object Segmentation in Complex Scenes. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023. [Google Scholar]
  65. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  66. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 248–255. [Google Scholar]
  67. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  68. Cheng, J.; Tsai, Y.H.; Hung, W.C.; Wang, S.; Yang, M.H. Fast and Accurate Online Video Object Segmentation via Tracking Parts. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  69. Duke, B.; Ahmed, A.; Wolf, C.; Aarabi, P.; Taylor, G.W. SSTVOS: Sparse Spatiotemporal Transformers for Video Object Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 5912–5921. [Google Scholar]
  70. Cheng, H.K.; Oh, S.W.; Price, B.; Schwing, A.; Lee, J.Y. Tracking Anything with Decoupled Video Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023. [Google Scholar]
Figure 1. A comparison of segmentation results when utilizing previously computed frames as a reference. The dashed bounding box highlights inaccurately segmented pixels in the reference frame, which introduces cumulative errors in subsequent frames. Conversely, our IM module effectively mitigates error propagation by modulating the inaccurately segmented pixels with reduced values (depicted in blue).
Figure 1. A comparison of segmentation results when utilizing previously computed frames as a reference. The dashed bounding box highlights inaccurately segmented pixels in the reference frame, which introduces cumulative errors in subsequent frames. Conversely, our IM module effectively mitigates error propagation by modulating the inaccurately segmented pixels with reduced values (depicted in blue).
Mathematics 12 00863 g001
Figure 2. An overview of our proposed MMN method. The specially designed Position Modulator (PM) enhances both query and memory embeddings with spatiotemporal positional information, enabling accurate reference retrieval for segmentation. Concurrently, the Importance Modulator (IM) takes segmentation quality and contextual importance into account when performing modulation on the memory, resulting in an informative memory.
Figure 2. An overview of our proposed MMN method. The specially designed Position Modulator (PM) enhances both query and memory embeddings with spatiotemporal positional information, enabling accurate reference retrieval for segmentation. Concurrently, the Importance Modulator (IM) takes segmentation quality and contextual importance into account when performing modulation on the memory, resulting in an informative memory.
Mathematics 12 00863 g002
Figure 3. The architecture of the proposed Importance Modulator (IM). The IM module takes as input the concatenation of query embedding V t , retrieved reference information R and the encoded embedding of prediction P t E . It then produces a modulation factor A t , which signifies the importance of each pixel.
Figure 3. The architecture of the proposed Importance Modulator (IM). The IM module takes as input the concatenation of query embedding V t , retrieved reference information R and the encoded embedding of prediction P t E . It then produces a modulation factor A t , which signifies the importance of each pixel.
Mathematics 12 00863 g003
Figure 4. Segmentation results comparing whether to use IM module. The red bounding box in the memory frame highlights the wrongly segmented area. In the estimated A, lower values are represented in blue, while higher values are represented in red.
Figure 4. Segmentation results comparing whether to use IM module. The red bounding box in the memory frame highlights the wrongly segmented area. In the estimated A, lower values are represented in blue, while higher values are represented in red.
Mathematics 12 00863 g004
Figure 5. Visual comparison with the state-of-the-art methods.
Figure 5. Visual comparison with the state-of-the-art methods.
Mathematics 12 00863 g005
Table 1. Quantitative results comparing the importance of frames at different temporal positions to matching-based VOS methods. Two state-of-the-art matching-based solutions STM [3] and AFB-URR [7] are chosen for comparison. The ‘Sample Rate’ is the proportion of intermediate frames collected to memory. For these two methods, without the first or last frame, performance degrades more significantly than that without multiple intermediate frames. The J & F score is the mean of region similarity J and contour accuracy F , from DAVIS [11].
Table 1. Quantitative results comparing the importance of frames at different temporal positions to matching-based VOS methods. Two state-of-the-art matching-based solutions STM [3] and AFB-URR [7] are chosen for comparison. The ‘Sample Rate’ is the proportion of intermediate frames collected to memory. For these two methods, without the first or last frame, performance degrades more significantly than that without multiple intermediate frames. The J & F score is the mean of region similarity J and contour accuracy F , from DAVIS [11].
First/LastSample Rate J & F
STM [3]AFB-URR [7]
-1/581.774.6
w/o first1/580.1 (−1.6)73.5 (−1.1)
w/o last1/580.1 (−1.6)74.0 (−0.6)
-1/1581.5 (−0.2)74.6 (−0.0)
Table 2. The results of ablation studies. We conduct an ablation study on the proposed Importance Modulator (IM). ✓ and ✗ indicate whether the IM module performs modulation on the key or value embedding. Bold indicates the best performance.
Table 2. The results of ablation studies. We conduct an ablation study on the proposed Importance Modulator (IM). ✓ and ✗ indicate whether the IM module performs modulation on the key or value embedding. Bold indicates the best performance.
KeyValue J & F J F
83.581.385.9
78.175.780.8
85.882.888.9
84.782.487.1
Table 3. The results of ablation studies. We conduct an ablation study on the proposed Position Modulator (PM). ‘Sequential’ means temporal coordinates are numbered sequentially, while ‘Customized’ represents that temporal coordinates are numbered as proposed in the Position Modulator. Bold indicates the best performance.
Table 3. The results of ablation studies. We conduct an ablation study on the proposed Position Modulator (PM). ‘Sequential’ means temporal coordinates are numbered sequentially, while ‘Customized’ represents that temporal coordinates are numbered as proposed in the Position Modulator. Bold indicates the best performance.
IMPEN J & F J F
#182.079.085.1
#22D Norm M 83.480.386.5
#33D-Sequential Norm M 83.580.486.6
#43D-Customized Norm M 80.677.583.7
#53D-Customized Norm R 84.080.987.2
#684.380.588.1
#73D-Customized Norm R 85.882.888.9
Table 4. Quantitative evaluation on the DAVIS 2017 validation set. Bold and underline indicate the best and the second-best performance, respectively.
Table 4. Quantitative evaluation on the DAVIS 2017 validation set. Bold and underline indicate the best and the second-best performance, respectively.
Methods J & F J F
OnAVOS [38]67.964.571.3
OSMN [28]54.852.557.1
OSVOS [29]59.256.661.8
RGMP [41]63.264.868.6
FAVOS [68]58.254.661.8
CINN [34]70.767.274.2
VideoMatch [37]62.456.568.2
PReMVOS [39]77.873.981.7
A-GAME [32]70.067.272.7
FEELVOS [35]71.669.174.0
STM [3]81.879.284.3
KMN [5]82.880.085.6
EGMN [6]82.980.085.9
CFBI [33]81.979.184.6
AFB-URR [7]74.673.076.1
RMNet [10]83.581.086.0
SST [69]82.579.985.1
LCM [9]83.580.586.5
STCN [27]85.482.288.6
BMVOS [31]72.770.774.7
RDE-VOS [40]84.280.887.5
DEVA [70]86.883.690.0
Ours85.882.888.9
Table 5. Quantitative evaluation on the DAVIS 2016 validation set. Bold and underline indicate the best and the second-best performance, respectively.
Table 5. Quantitative evaluation on the DAVIS 2016 validation set. Bold and underline indicate the best and the second-best performance, respectively.
Methods J & F J F FPS
OnAVOS85.586.184.90.1
OSVOS80.279.880.60.1
MaskRNN80.880.780.9-
CINN84.283.485.0<0.1
LSE81.682.980.3-
VideoMatch81.981.080.83.1
PReMVOS86.884.988.6<0.1
A-GAME82.182.282.014.3
FEELVOS82.281.788.12.2
STM89.388.789.96.3
KMN90.589.591.58.3
CFBI89.488.390.55.6
RMNet88.888.988.711.1
RDE-VOS91.189.792.535.0
Ours91.490.692.26.0
Table 6. Quantitative evaluation on the YouTube-VOS 2018 validation set. Bold and underline indicate the best and the second-best performance, respectively.
Table 6. Quantitative evaluation on the YouTube-VOS 2018 validation set. Bold and underline indicate the best and the second-best performance, respectively.
Methods G J S F S J U F U
OnAVOS55.260.162.746.651.4
OSMN51.260.060.140.644.0
OSVOS58.859.860.554.260.7
RGMP53.859.5-45.2-
BoLTVOS71.171.6-64.3-
PReMVOS66.971.475.956.563.7
A-GAME66.167.8-60.8-
STM79.479.784.272.880.9
KMN81.481.485.675.383.3
EGMN80.280.785.174.080.9
CFBI81.481.185.875.383.4
AFB-URR79.678.874.183.182.6
RMNet81.582.185.775.782.4
LCM82.082.286.775.783.4
SST81.781.2-76.0-
STCN83.081.986.577.985.7
BMVOS73.973.568.577.476.0
Ours83.181.286.577.686.2
Table 7. Quantitative evaluation on the MOSE validation set. Bold and underline indicate the best and the second-best performance, respectively.
Table 7. Quantitative evaluation on the MOSE validation set. Bold and underline indicate the best and the second-best performance, respectively.
Methods J & F J F
STM [3]48.644.252.9
RDE-VOS [40]48.844.652.9
STCN [27]50.846.655.0
DEVA [70]60.055.864.3
Ours52.146.957.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lu, H.; Guo, Z.; Zuo, W. Modulated Memory Network for Video Object Segmentation. Mathematics 2024, 12, 863. https://doi.org/10.3390/math12060863

AMA Style

Lu H, Guo Z, Zuo W. Modulated Memory Network for Video Object Segmentation. Mathematics. 2024; 12(6):863. https://doi.org/10.3390/math12060863

Chicago/Turabian Style

Lu, Hannan, Zixian Guo, and Wangmeng Zuo. 2024. "Modulated Memory Network for Video Object Segmentation" Mathematics 12, no. 6: 863. https://doi.org/10.3390/math12060863

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop