Next Article in Journal
Predicting the Early-Age Time-Dependent Behaviors of a Prestressed Concrete Beam by Using Physics-Informed Neural Network
Next Article in Special Issue
Evaluation of the Use of the 12 Bands vs. NDVI from Sentinel-2 Images for Crop Identification
Previous Article in Journal
Global Path Planning of Unmanned Surface Vehicle Based on Improved A-Star Algorithm
Previous Article in Special Issue
PCRMLP: A Two-Stage Network for Point Cloud Registration in Urban Scenes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Graph Neural Network-Based Method of Spatiotemporal Land Cover Mapping Using Satellite Imagery

Faculty of Electrical Engineering and Computer Science, University of Maribor, Koroška Cesta 46, 2000 Maribor, Slovenia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(14), 6648; https://doi.org/10.3390/s23146648
Submission received: 2 June 2023 / Revised: 14 July 2023 / Accepted: 19 July 2023 / Published: 24 July 2023
(This article belongs to the Special Issue Deep Learning for Environmental Remote Sensing)

Abstract

:
Multispectral satellite imagery offers a new perspective for spatial modelling, change detection and land cover classification. The increased demand for accurate classification of geographically diverse regions led to advances in object-based methods. A novel spatiotemporal method is presented for object-based land cover classification of satellite imagery using a Graph Neural Network. This paper introduces innovative representation of sequential satellite images as a directed graph by connecting segmented land region through time. The method’s novel modular node classification pipeline utilises the Convolutional Neural Network as a multispectral image feature extraction network, and the Graph Neural Network as a node classification model. To evaluate the performance of the proposed method, we utilised EfficientNetV2-S for feature extraction and the GraphSAGE algorithm with Long Short-Term Memory aggregation for node classification. This innovative application on Sentinel-2 L2A imagery produced complete 4-year intermonthly land cover classification maps for two regions: Graz in Austria, and the region of Portorož, Izola and Koper in Slovenia. The regions were classified with Corine Land Cover classes. In the level 2 classification of the Graz region, the method outperformed the state-of-the-art UNet model, achieving an average F1-score of 0.841 and an accuracy of 0.831, as opposed to UNet’s 0.824 and 0.818, respectively. Similarly, the method demonstrated superior performance over UNet in both regions under the level 1 classification, which contains fewer classes. Individual classes have been classified with accuracies up to 99.17%.

1. Introduction

Land cover mapping represents spatial information about the different types of the Earth’s surface coverage. Historically, one of the most influential causes of land cover changes is urbanisation [1,2]. Temporal observation of land cover changes is necessary for advanced crop [3] and disaster management (e.g., forest fires [4] and landslides [5]). Different land cover types have a heterogeneous regional impact on climate change [6]. This increases the demand for automatic land cover classification of geographically diverse regions further.
Breakthroughs in satellite control systems [7] and the development of techniques for well-organised formation flight above the Earth [8] have expanded the potential applications of satellite technology significantly. Satellite imagery, as a form of Big Data obtained with multispectral scanners, offers new perspectives in using remote sensing for land cover mapping and change detection [9,10,11]. Non-commercial satellites (e.g., Landsat, Sentinel) are a good alternative to commercial satellites for obtaining high spatial resolution imagery [12,13,14].
Several satellite imagery land cover datasets have been created. The Reunion Island dataset [15] consists of a time series of 21 Sentinel-2 images acquired in 2017. A respective pixel-wise ground truth map was built from various sources and completed by field experts [16]. Given the increasing desertification, rapid deforestation, abandonment of poor farmland, wetland drying, and continuous coastal urban development, there is an urgent need for comprehensive land cover inventories [17]. The Corine Land Cover (CLC) dataset [18], which provides pixel-wise ground truth maps of Europe and consists of 44 Land Use/Land Cover (LULC) classes, serves as an effective tool to manage these widespread environmental changes resulting from human activities [17]. The classification within the CLC is organised into three class levels: level 1 has 5 classes, level 2 contains 15 classes, and the most detailed, level 3, has 44 classes. The CLC inventory was initiated in 1985, and had the first release for the reference year of 1990 [19]. The inventory has been updated several times since, specifically in 2000 [20], 2006 [21], 2012 [22] and 2018 [23]. Throughout these years there have been significant improvements in the spatial resolution of the CLC dataset, from 50 m in the CLC1990 to 10 m in CLC2018, as well as the quality of satellite data used, transitioning from Landsat-5 images in 1990 to the use of Sentinel-2 and Landsat-8 images in 2018. The Land Use and Coverage Area frame Survey (LUCAS) [24], managed by Eurostat, has conducted in-depth surveys across Europe every three years since 2006, accumulating over 1 million data points from over 650,000 surveyed locations. These surveys provide rich data on the LULC, and agro-environmental information for supporting spatial, territorial and agricultural analyses. To address the dataset’s limitations and to make it more accessible, an open-source system was developed to integrate the LUCAS dataset into workflows [25]. A recently released dataset named Dynamic World NRT offers a global 10 m resolution LULC map with 9 classes [26].
Satellite imagery has been fused with derived data from other types of remote sensing, such as Light Detection and Ranging (LiDAR). The obtained data can be used further for creating digital 3-D models (e.g., terrain and cities). A city’s 3-D model and multispectral satellite data were used for spatial modelling of the land surface temperature [27]. Different combinations of imagery and Digital Elevation Models (DEMs) were used for landform classification using a deep learning-based approach [28].
New data sources and computational power have improved LULC mapping significantly using data from satellite imagery, land surveys and online portals. Different modelling techniques are used to predict LULC changes, including statistical models, cellular automata models (simulations that apply transition rules to study spatial dynamics/interactions), economic models, agent-based models (simulations of interactions of autonomous entities, e.g., landowners and policymakers), hybrid models that combine the strengths of individual models and time series modelling [29]. Machine Learning (ML) based methods are emerging as a powerful tool for LULC mapping by performing a pixel-based, a sub-pixel based or an object-based classification of satellite imagery [12]. The object-based classification is preferred, due to the high spatial resolution of satellite images. Several studies reported that the object-based approach produced better classification results compared to the pixel-based methods [30,31,32]. A specialised schema for enhancing the land cover identification [33] iteratively performs the assessment of the classification errors and lowers the error rate by creating new subclasses. Neural Networks (NNs) are the driving force in achieving state-of-the-art performance for land cover mapping, with Convolutional Neural Networks (CNNs) being the most frequently used architectures [34,35,36,37,38,39]. Recurrent Neural Networks (RNNs), especially Long Short-Term Memory (LSTM) networks, have shown a strong potential in utilising temporal information. This includes land cover classification [40,41], crop classification [42], predicting flood disasters [43] and performing natural disaster analysis [44]. Graph theory has achieved significant research advances [45,46,47] that have found applications across various industries and other research fields. Lately, researchers have started to utilise Graph Neural Networks (GNNs) for object-based land cover classification. GNN has been used in a method for 3-D building reconstruction, which used street view images and block meshes as inputs [48]. Researchers have proven the potential of Graph Attention Networks (GATs) [16] and fusion of aggregators in GNNs [49] to perform state-of-the-art land cover mapping as a node classification task.
In this paper, we present a novel spatiotemporal method for object-based LULC classification of satellite imagery using a GNN. The method starts with a superpixel segmentation, followed by segments being connected through time. Each individual segment is represented as a node in the graph. A target node (i.e., a selected segment) is classified by passing the subgraph with the target node and its neighbourhood through the classification pipeline. In the first step, the pipeline performs the image feature extraction with CNN. Importantly, the selection of the CNN architecture is not constrained to any specific requirements. This provides freedom to use any of the various state-of-the-art architectures, such as Vision Transformer (ViT) [50], EfficientNetV2 [51], EfficientNet [52], ResNet [53], ResNeXt [54], InceptionV3 [55], or DenseNet [56]. Following the extraction of features, the pipeline proceeds to the classification step, utilising a GNN to classify the target node. The choice of a GNN is as versatile as the CNN selection, offering options such as the Graph Attention Network (GAT) [57] or the SAmple and aggreGatE (GraphSAGE) framework [58]. This highlights the flexibility and modularity of the proposed method.
The main contributions of this paper, listed below, introduce novelties that have not been considered by the prior existing state-of-the-art ([16]):
  • A new representation of the sequential satellite images as a directed graph by connecting segmented land region through time, based on sequential spatial segment overlaps.
  • A new land cover mapping method as node classification in the derived directed graph using the GNN. The proposed method allows selection of a target node’s neighbourhood, which contains historical temporal context information of connected land region segments. The size of the neighbourhood determines the volume of input spatial and temporal information that the selected GNN uses for node classification.
  • A modular target node classification pipeline, which offers flexible selection of a CNN for image feature extraction and a GNN for node classification.
  • The first application of using EfficientNetV2 as a feature extractor for GraphSAGE classification models to perform intermonthly land cover classification.
  • Complete intermonthly land cover classification maps for the given regions by using Sentinel-2 imagery, as shown in the Section 5.
The rest of this paper is organised as follows. Section 2 reviews related works. Section 3 presents the proposed land cover classification method. Dataset creation to evaluate the proposed method to the state-of-the-art is described in Section 4. The classification results and discussion are given in Section 5. Section 6 concludes this paper.

2. Related Work

In this section, we review related work in two categories: object-based land cover classification of satellite imagery and GNNs for land cover classification.

2.1. Object-Based Land Cover Classification of Satellite Imagery

A classification in hierarchical geospatial databases with a two-step deep learning framework has enabled simultaneous prediction of land use in all hierarchical levels [34]. This method achieved overall accuracies in the order of 90% for two datasets [34]. The main dataset contained satellite images of the settlements of Hameln and Schleswig in Germany with 8 land cover classes. State-of-the-art comparison was performed using the dataset from the ISPRS 2D semantic labelling challenge with satellite images of the settlements of Vaihingen and city of Potsdam in Germany with 6 land cover classes [59].
A multi-level context-guided method with the object-based CNN offers learning of high-level features from spectral patterns, geometric characteristics and object-level contextual information, based on a segmented context patch [32]. This method achieved remarkable accuracy (>80%) compared to traditional methods [32].
A semi-automated object-based land cover classification to observe changes in the Kurdistan region of Iraq using Landsat imagery was proposed [60]. After segmentation of the region, the spatial and spectral information were combined and followed by the object feature extraction. The classification was performed with a nearest neighbour classifier, achieving accuracies in the range of 88% to 89%.
A multi-view object-based classification with a CNN has proved to achieve a higher overall accuracy of 82% compared to application of a CNN on an orthoimage, which achieved an accuracy of 65% [61]. The results highlight that the window-based version of the proposed method can achieve better accuracy than the full object version.
Images from a multispectral camera, LISAT, onboard the third-generation satellite by LAPAN, named LAPAN-A3, were used in a study to perform object-based land cover classification of the Rote Island, using a tree method algorithm [62]. The results achieved overall classification accuracy of 86% for different classes, including open land, settlements, water bodies and diverse vegetation.
The study examined the potential of high-resolution PlanetScope data, made accessible on Google Earth Engine, to enhance land cover classification of Central Brazil, utilising both pixel-based and object-based methods [63]. The results revealed that the exclusive use of PlanetScope data yielded a 67% overall accuracy for pixel-based techniques and 82% for object-based methods. The fusion of PlanetScope data with Sentinel imagery data raised the overall accuracy to 82% for pixel-based and an impressive 91% for object-based methods.
A proposed method for paddy field mapping in Iran used rice phenology, yearly land surface temperature data and multi-temporal Landsat-8 imagery [64]. Enriched by a digital elevation model and spectral indices, the object-based classification outperformed the pixel-based approach with overall accuracy of 94% vs. 92%.

2.2. GNNs for Land Cover Classification

A Graph Convolutional Network (GCN) is a variant of CNNs which operates on graphs [65]. The layers in the model learn to encode both the local graph structure and the node features efficiently by performing approximation of the spectral convolution. Temporal GCN (T-GCN) is a combination of GCN and a Gated Recurrent Unit (GRU) to capture spatial and temporal dependencies simultaneously. T-GCN was designed originally for predicting traffic [66].
The Graph Attention Network (GAT) leverages masked self-attentional layers to help in extracting information from features of nodes [57]. The Symmetric Relation based Heterogeneous Graph Attention Network, denoted as SR-HGAT, is an extension of the original GAT network [67]. SR-HGAT takes the high-order relations of heterogeneous graphs into account [67]. A Robust Representation Learning method for multilabel images driven by GATs (RRL-GAT) reduces noisy and false connections between the connected image objects [68].
GraphSAGE (SAmple and aggreGatE) is a framework for generating inductive embeddings by sampling a neighbourhood of a node and then using a learnable aggregator function to aggregate features [58]. The data-efficient algorithm PinSAGE combines highly efficient random walks (improved upon GraphSAGE) and the graph convolutions for processing of graphs with billions of nodes [69].
A U-shaped object graph neural network (U-OGNN) for accurate land cover mapping using high-resolution remote sensing images was proposed [70]. The U-OGNN is composed of a self-adaptive graph construction that employs similarity measures to generate a contextual-aware graph structure. It also features a graph encoder and decoder, which fuse information across various scales, capturing hierarchical features of adjacent objects in images. The method demonstrated the overall accuracy of 88%.
A method for land cover classification, named AF2GNN, was proposed, using hyperspectral images as input to a GNN with adaptive filters and aggregator fusion [49]. Degree-scalers are defined to combine the multiple filters and present the graph structure. This method addresses the common problems of land cover discrimination, noise impact and spatial feature learning. The method achieved overall accuracies above 94% for various datasets.
An Attentive Spatial Temporal Graph CNN exploits the spatial and temporal information in satellite image time series in the context of land cover mapping [16]. After performing the segmentation with the SLIC algorithm [71], the graph is formed by connecting the adjacent segments. Each georeferenced segment or node contains a time series of segments. Target and neighbourhood segments are passed to separate inputs of the classification NN, which utilises the attention mechanism. The NN outputs the classification of the target segment. To the best of our knowledge, this is the first and only spatiotemporal application of GNNs for remote sensing data.
The input graphs into the Attentive Spatial Temporal Graph CNN represent connected segments based on an adjacency in the space without using any kind of temporal information [16]. This method obtained the graphs by segmenting a hand-picked time series image that experts considered suitable. The proposed method in this paper takes shifting and growing of segments into account additionally. The thresholded temporal connections between segments are formed, based on sequential spatial segment overlaps in segmented time series images. The GNN in the proposed method can be replaced by any chosen GNN algorithm, making the proposed method modular, as opposed to the Attentive Spatial Temporal Graph CNN [16], which uses the attention mechanism from GATs exclusively.

3. Methodology

This paper proposes a novel spatiotemporal method for the LULC classification of satellite imagery using a GNN. The workflow of the proposed method is shown in Figure 1. The proposed method starts by performing superpixel segmentation (step 1 in Figure 1) on each multispectral image in the time series of satellite images T S i m a g e of length T. This results in a time series of segmentation masks T S m a s k . The next step is the graph construction (step 2 in Figure 1), based on temporal segment overlaps, where each segment in each satellite image is represented as a node v. An edge between two nodes is made if two segments in two sequential satellite images overlap sufficiently. The constructed graph G contains the same number of nodes as the number of all superpixels in T S m a s k . Each v contains the segment representation as a resized multispectral image bounding box, b b o x . Next, a subgraph G s u b is created for each target node v t a r g e t based on the input edges towards it (step 3 in Figure 1). The v t a r g e t represents the selected segment s s e l e c t e d to classify. The G s u b is then passed into the target node classification pipeline (step 4 in Figure 1). It performs feature extraction from b b o x with any chosen CNN (step 4.1 in Figure 1). This is followed by v t a r g e t classification with a selected GNN (step 4.2 in Figure 1), which outputs the class probabilities. The land cover class of s s e l e c t e d is determined based on the highest probability. The final result is the time series of the land cover mappings T S c o v e r .
The following subsections provide detailed descriptions of all four main steps of the proposed method.

3.1. Superpixel Segmentation

Superpixels are connected groups of similar pixels that form a visually meaningful entity while reducing the number of primitives for subsequent processing. This makes many algorithms computationally feasible on high dimensional images. Pixels are grouped by colour, proximity and other low-level properties [72].
In the proposed method, a multispectral image with C layers is segmented to create meaningful land segments, which surround objects, e.g., buildings, field crops, forests, rivers and roads. The superpixels are calculated with Felzenszwalb’s efficient graph based image segmentation [73]. The algorithm has three parameters [73]:
  • σ —the Standard Deviation of the Gaussian kernel to smooth the image in the preprocessing stage;
  • k—a scale of observations for the threshold function, which controls the degree of required difference between two adjacent superpixels (a larger k causes a preference for larger superpixels);
  • m i n _ s i z e —the minimum number of pixels inside the superpixel to control the merging of neighbouring superpixels in the postprocessing stage.
The result of segmenting a single satellite image, captured at time t, is a segmentation mask m a s k t , which is the t-th mask in T S m a s k .
The application of Felzenszwalb’s image segmentation algorithm is demonstrated in Figure 2. The segmentation masks with k = 70 reveal that large superpixels (i.e., segments) enclose too many meaningful land objects. The results from a small k value contain too many small segments, which leads directly to higher computational cost in the later steps of the method. The value of m i n _ s i z e should be adjusted accordingly, to avoid the presence of the remaining small segments in heterogeneous areas of the image.

3.2. Graph Construction

The T S m a s k is the fundamental for constructing a weighted directed graph G, connecting segments through time. Each i-th segment s from the m a s k t , labelled as s i , t , is added as a node v i , t in G. The created G contains the same amount of nodes as there are segments in T S m a s k . An individual v i , t forms a directed edge to v j , t + 1 , if s i , t overlaps at least τ (the value range is the ( 0 , 1.0 ] ) portion of the s j , t + 1 . The overlap portion serves as the weight w i , j of the formed edge. An example of graph construction is shown in Figure 3.

3.3. Segment Representation and Subgraph Construction

An individual v G holds the corresponding segment’s multispectral image bounding box, b b o x . It can be extended in width and height to capture additional spatial contextual information around the segment. Due to segments being of different shapes and, consequently, the bounding boxes being of various sizes, the resizing of each b b o x must be performed to the width b b o x w and height b b o x h . The resizing keeps the original aspect ratio of the multispectral image inside the b b o x by zero padding. This resizing is needed to ensure that all data inside the nodes are of equal size. It is worth noting that b b o x contains a b b o x C number of layers.
The target node v t a r g e t (i.e., the selected segment s s e l e c t e d ) and it’s neighbourhood of nodes in G form a subgraph G s u b . The parameter T l o o k b a c k controls the historical temporal context information in G s u b by determining the maximum distance of the nodes, based on the input edges towards the v t a r g e t . By setting the value of T l o o k b a c k = 0, only the v t a r g e t at time t will be present in the constructed G s u b . By increasing the value of T l o o k b a c k to, e.g., 2, the G s u b will feature the v t a r g e t at time t, all nodes at time t 1 with directed edges towards the v t a r g e t , and all nodes at time t 2 with edges directed towards all previously included nodes in time t 1 .
The largest subgraphs can be constructed for the leaf nodes of G, which represent segments obtained for the final image in T S s a t e l l i t e _ i m a g e at time t = T. Even though T l o o k b a c k can be of any number greater than or equal to 0, each G s u b can contain nodes and edges only up to the first image at t = 0. An important fact is that subgraphs of all target nodes in t = 0 actually contain only the v t a r g e t itself, because the nodes at time t = 0 do not contain any input edges. The G s u b construction is demonstrated in Figure 4. All the visualised subgraphs have bounding boxes drawn around the included nodes.

3.4. Node Classification Pipeline

The G s u b serves as the input to the target node classification pipeline, which outputs the classification probabilities of the s s e l e c t e d , represented by the v t a r g e t . The node classification pipeline is visualised in Figure 5. The first step is the feature extraction, which is performed v G s u b (step 1 in Figure 5). Due to the b b o x inside each v containing a b b o x C number of layers and the majority of state-of-the-art CNNs requiring a 3 channel input, the b b o x is first passed through the 2D convolution with 3 kernels of size 1 × 1 and a stride of 1. This converts the depth of the input from b b o x C to 3. The 3 feature maps are then passed to the CNN. It extracts F features in the form of a feature vector, which is then stored inside the v. The value of F depends on the output size of the last feature extraction layer of the CNN (the softmax classification layer at the end of NN is removed). For majority of established CNNs, the number of output features from last feature extraction layer cannot be adjusted without modifying previous layers of the NN. After feature extraction, the G s u b is passed as the input to the GNN (step 2 in Figure 5). It contains the same number of layers as the value of T l o o k b a c k (the only exception being the case of T l o o k b a c k = 0 with GNN containing a single layer). The GNN outputs N class probabilities of v t a r g e t . The class of s s e l e c t e d is determined as the class with the highest probability. After the classification of v t a r g e t G , the time series is created of land cover mappings T S c o v e r .

4. Dataset Preparation

Datasets with LULC classification of satellite imagery on a daily/weekly/monthly basis for a multi-year period of time are not available. This section describes creation of an intermonthly dataset to evaluate and compare the performance of the proposed method to the state-of-the-art. The first Subsection provides details about acquisition of Sentinel-2 images of the selected regions. The second Subsection describes the process of creating land cover ground truth.

4.1. Intermonthly Satellite Imagery Acquisition

To create the land cover dataset, the first step was to acquire atmospherically corrected Sentinel-2 L2A images, with the earliest globally available being from 2017. From 1992 to 2015 the conversion of cropland to artificial surfaces was notably high in Austria and Slovenia, with rates above 4% and 3% respectively, exceeding the average rate of above 2% observed across the European Union (EU28). In the same period, Austria and Slovenia also underwent a conversion of (semi-)natural vegetated areas to cropland at rates above 2% and nearly 6% respectively, with the EU28 average just under 3% (https://www.oecd.org/environment/indicators-modelling-outlooks/monitoring-land-cover-change.htm (accessed on 12 July 2023)). Given these trends, a region of Graz in Austria, and the region of the cities—Portorož, Izola and Koper—in Slovenia were selected, due to their spatial heterogeneity.
Images were obtained for the two selected regions: Graz, Austria, encapsulated within the WGS84 bounding box [15.390816°, 46.942176°, 15.515785°, 47.015961°], and the region of Portorož, Izola and Koper in Slovenia, defined by the bounding box [13.590260°, 45.506948°, 13.744411°, 45.554449°]. These values denote pairs of minimum and maximum longitudes and latitudes, respectively. The created T S i m a g e for Graz contained 40 intermonthly Sentinel-2 multispectral images with minimal cloud coverage, spanning from January 2017 to July 2021. Similarly, the T S i m a g e for the region of Portorož, Izola and Koper contained 41 intermonthly Sentinel-2 multispectral images, also characterised by minimal cloud coverage, captured from January 2017 through June 2021. Months with too high cloud coverage were skipped.
Each image stores data as unsigned 16-bit integers. The images for the Graz region have the size of 946 × 825 pixels, corresponding to an area of 78 km 2 at a 10 m spatial resolution, while those for the region of Portorož, Izola and Koper are 1212 × 508 pixels in size, representing an area of 61 km 2 at the same spatial resolution. All images contain 13 basic spectral layers (B01–B12) and 4 calculated layers: Normalised Difference Vegetation Index (NDVI), Normalised Difference Moisture Index (NDMI), Normalised Difference Water Index (NDWI) and Normalised-Difference Snow Index (NDSI). The spatial resolution of layers ranged from 10 m to 60 m. The described T S i m a g e for both regions are visualised in Figure 6.

4.2. Land Cover Ground Truth Creation

The next step of dataset creation was the superpixel segmentation. Segmentation masks serve as the basis for creation of the time series of ground truth land cover mappings T S g t _ c o v e r . An individual image was segmented, based on the normalised values of 13 basic spectral layers. The Felzenszwalb’s segmentation algorithm used the parameters σ = 0.5, k = 12 and m i n _ s i z e = 15 pixels. These values were selected empirically, to ensure a balanced tradeoff between segments being small enough to capture complex land objects and the total number of segments not exceeding the storage limitations of hardware that we used for deep learning.
The created T S m a s k for the region of Graz contains 224,665 segments (avg. 5616 segments/ m a s k ). Similarly, the T S m a s k for the region of Portorož, Izola and Koper has 122,685 segments (avg. 2992 segments/ m a s k ). Examples of the segmentation results are visible in Figure 7.
The last step of dataset creation was the ground truth land cover labelling. For each dataset region, an individual s of m a s k T S m a s k was labelled with a single class among 15 LULC classes in CLC2018 level 2 [18]. To speed up the labelling process, we utilised a pretrained UNet model by Esri (https://www.arcgis.com/home/item.html?id=afd124844ba84da69c2c533d4af10a58 (accessed on 12 July 2023)), which achieved an overall pixel-based accuracy of 84.0% for CLC level 2 classification. A multispectral image with 13 basic spectral layers served as the input to the pretrained model. Before passing the image to the UNet model, each layer had to be normalised and standardised, based on provided pretrained model statistics. The last step of preprocessing was scaling of the input image to 1280 × 1280 pixels (nonsquare images were zero padded). The input image was then passed into the UNet model, which output a tensor of size 1280 × 1280 × 15 (15 CLC level 2 classes). The output tensor was reshaped into a size of 1280 × 1280 × 1 by performing the a r g m a x function across the 3rd axis. The classification tensor was then reshaped to match the size of the original image by downscaling and removal of the padding. The classification tensor and respective m a s k were then used to create a ground truth image. This was performed automatically by selecting each individual s from m a s k and assigning it the majority class of associated pixels from the classification tensor. The automatically created ground truth image was finalised with manual label corrections of segments. This was necessary due to rare misclassifications from the UNet model. By performing the described process on all images in individual T S i m a g e , the time series was created of the ground truth land cover mappings T S g t _ c o v e r .
The T S g t _ c o v e r for the region of Graz contains 12 of the 15 CLC level 2 classes, excluding Inland Wetlands, Maritime wetlands, and Marine waters; meanwhile, the T S g t _ c o v e r for the region of Portorož, Izola and Koper includes 13 out of the 15 classes, omitting Mine, dump and construction sites, and Inland wetlands from the selected region. Examples of classification output, obtained with the UNet model, and manually corrected time series of the ground truth T S g t _ c o v e r , are visualised in Figure 8.
For each region, the T S i m a g e was split into T S t r a i n _ i m a g e and T S t e s t _ i m a g e . Specifically, the Graz region’s T S i m a g e was split into a T S t r a i n _ i m a g e that containes 29 consecutive images (accounting for 72.5% of all images) from January 2017 to April 2020, and a T S t e s t _ i m a g e with 11 consecutive images (making up 27.5% of all images) from May 2020 to July 2021. Similarly, the T S i m a g e for the region of Portorož, Izola and Koper was split into a T S t r a i n _ i m a g e , containing 29 consecutive images (70.7% of all images) from January 2017 to October 2019, and a T S t e s t _ i m a g e , containing 12 consecutive images (29.3% of all images) from January 2020 to June 2021. The split ratio was decided upon by the fact that each T S t e s t _ i m a g e should contain images across more than a year. The T S g t _ c o v e r and T S m a s k of each region were split in the same manner into T S t r a i n _ g t _ c o v e r and T S t e s t _ g t _ c o v e r , as well as T S t r a i n _ m a s k and T S t e s t _ m a s k . The distribution of ground truth land cover labels of pixels in individual T S t r a i n _ g t _ c o v e r and T S t e s t _ g t _ c o v e r are shown in Figure 9. Note that T S t e s t _ g t _ c o v e r for the region of Graz lacks pixels with land cover classification of Permanent crops and Open spaces with little or no vegetation.

5. Results and Discussion

The proposed method was evaluated on the created dataset, with its performance compared against that of the current state-of-the-art. The first Subsection describes the application of the proposed method on the created dataset. The second Subsection describes the analysis of GNN usage in the proposed method, while the classification performance and discussion are presented in the third Subsection.

5.1. Application of the Proposed Method and Experimental Parameters

From the dataset creation process, the T S t r a i n _ m a s k and T S t e s t _ m a s k of both regions were carried over. For each individual region in the dataset, the T S t r a i n _ i m a g e , T S t r a i n _ m a s k and T S t r a i n _ g t _ c o v e r were used for constructing the training graph G t r a i n , with T S t e s t _ i m a g e , T S t e s t _ m a s k and T S t e s t _ g t _ c o v e r being used to construct the test graph G t e s t . The construction of both graphs for individual region was performed using the empirically set parameter τ = 0.2. This value ensured that the number of inbound edges per node was limited to 5.
For the region of Graz, the constructed G t r a i n contained 163,414 nodes, and the G t e s t had 61,251 nodes. Similarly, for the region of Portorož, Izola and Koper, the G t r a i n had 86,462 nodes and the G t e s t 36,223 nodes. The distribution of ground truth land cover classification of nodes in G t r a i n and G t e s t for both dataset regions is shown in Figure 10.
For the classification of individual region within the created dataset, we used the following set of experimental parameters:
  • Individual v contained the b b o x , which was extended by 20 pixels in both width and height. The b b o x was resized to the size of b b o x w = 48 and b b o x h = 48. It also contained b b o x C = 7 layers, specifically, B04, B03, B02, NDVI, NDMI, NDWI and NDSI.
  • The feature extraction part of the target node classification pipeline starts by passing the segment’s b b o x through the trainable 2D convolution with 3 kernels. This outputs a tensor with 3 feature maps. These are then passed into the CNN - the state-of-the-art EfficientNetV2-S [51] was selected. It had the fully-connected output layer removed and all the layers trainable. Before passing the 3 feature maps into the EfficientNetV2-S, they were passed through the EfficientNetV2-S’s preprocessing transforms. The final output of the EfficientNetV2-S was F = 1280 extracted features.
  • Each but the last layer in the GNN had a hidden dimension of 256, with a dropout of 0.5. The output dimension of the final layer in the GNN was set to match the number of classification classes N. Specifically, it was set to 12 for the Graz region and 13 for the region of Portorož, Izola and Koper. All the GNN layers were trainable.
  • Before training, the EfficientNetV2-S model for feature extraction was initialised with ImageNet pretrained weights, because the preliminary experiments showed that this led to lower training loss. The classification pipeline forms a single model and it was, same as in [58], trained in a single training process using the Adam optimiser [74]. The initial learning rate was set to 0.001 and the model was trained for 10 epochs with a random shuffle of the training samples (i.e., subgraphs) inbetween. The batch size was 30. The loss function was categorical cross-entropy, which used class weights, due to unbalanced ground truth land cover labels of the nodes in G t r a i n . The edge weights in G s u b were ignored during the message passing through the GNN, because initial empirical experiments had shown that that led to lower training loss.
Regarding the experimental protocol, for each region in the dataset the experiments were performed for multiple values of T l o o k b a c k . A total of 10 runs were performed per experiment.
Overall accuracy reflects correct classification for all classes combined, while the class-specific accuracy assesses the correctness for each individual class independently [75]. The equations for the i-th class accuracy and overall accuracy are defined with (1) [76]. The equations contain T P i , T N i , F P i and F N i , to denote True Positives (the number of samples identified correctly as positive), True Negatives (the number of samples identified correctly as negative), False Positives (the number of samples identified incorrectly as positive) and False Negatives (the number of samples identified incorrectly as negative) for the i-th class, respectively. Recall (or sensitivity) quantifies how many actual positive instances were identified correctly [76]. Precision measures the proportion of true positive predictions among all the positive predictions [76]. The F1-score is the harmonic mean of precision and recall, providing a balance between precision and recall, making it especially useful for imbalanced datasets [76]. All three class-specific metrics are defined with (2) [75]. The metrics can be micro-averaged (all samples contribute equally), macro-averaged (all classes constribute equally) and weighted-averaged (the contribution of each class is weighted by the proportion of its instances in the total data, labelled as w i ). The equations for micro- and macro-averaged recall, precision and F1-score are defined with (3), (4) and (5), respectively [75]. The weighted-averaged variants of the metrics are defined with (6) and (7) [75]. The choice between micro-, macro-, and weighted-averaging depends on the balance of the dataset and the problem that is being solved. Micro-averaging is the preferred approach if every sample in the dataset holds equal importance. On the other hand, macro-averaging is used when the significance is distributed uniformly among all the classes. Finally, weighted-averaging should be used with imbalanced datasets where the relevance of classes is defined by their size [75]. Given the significant imbalance in the created dataset, as previously shown in Figure 10, the weighted F1-score alongside accuracy were used as the primary evaluation metrics.
A c c u r a c y i = T P i + T N i T P i + F P i + F N i + T N i , A c c u r a c y = i T P i + T N i i ( T P i + F P i + F N i + T N i )
R e c a l l i = T P i T P i + F N i , P r e c i s i o n i = T P i T P i + F P i , F 1 i = 2 · P r e c i s i o n i · R e c a l l i P r e c i s i o n i + R e c a l l i
M i c r o R e c a l l = i T P i i ( T P i + F N i ) , M a c r o R e c a l l = 1 N i R e c a l l i
M i c r o P r e c i s i o n = i T P i i ( T P i + F P i ) , M a c r o P r e c i s i o n = 1 N i P r e c i s i o n i
M i c r o F 1 = 2 · M i c r o P r e c i s i o n · M i c r o R e c a l l M i c r o P r e c i s i o n + M i c r o R e c a l l , M a c r o F 1 = 1 N i F 1 i
W e i g h t e d R e c a l l = i w i R e c a l l i , W e i g h t e d P r e c i s i o n = i w i P r e c i s i o n i
W e i g h t e d F 1 = i w i F 1 i

5.2. Analysis of GNN Usage in the Proposed Method

In order to choose the best performing GNN for the proposed method’s target node classification pipeline, the experiments were performed on a subset of the created dataset. More precisely, only nodes in G t r a i n from the first seven satellite images (from January 2017 to August 2017) were used to train classification models, to classify the G t e s t of the region of Portorož, Izola and Koper. The experimental values for T l o o k b a c k were 0 , 1 , 2 , 3 . The average accuracy and weighted F1-score were calculated for the CLC classification at class level 2 considering 10 runs. Two GNNs were considered for this evaluation:
  • The GraphSAGE algorithm [58], which applied LSTM aggregation in each of its layers. The output of the hidden layers was passed through the rectified linear unit (ReLU) activation function, as it was done in [58].
  • GAT [57] with outputs of hidden layers being passed through the exponential linear unit (ELU) activation function.
In addition, we also examined the effect of eliminating the GNN from the target node classification pipeline. In this case, global average pooling was performed for each individual extracted feature across all nodes in the subgraph G s u b . More specifically, feature vectors from all nodes in each G s u b were stacked in a tensor with 1280 columns. Each individual feature then underwent global average pooling, which resulted in a tensor of size 1 × 1280. Following this process, an output fully connected layer was utilised, receiving the tensor as input and yielding the output classification tensor with size equal to the number of classes N. Importantly, when T l o o k b a c k = 0 and the GNN is not used, the whole classification process essentially becomes an image classification task using EfficientNetV2-S. The experiments of excluding a GNN from the target node classification pipeline were conducted, to determine the necessity of using a GNN in the proposed method to achieve high overall classification performance.
The results displayed in Figure 11 demonstrate how both the choice of GNN and the value of T l o o k b a c k affect the performance of the proposed method for the CLC level 2 classification. Based on the obtained results, utilising GraphSAGE as the GNN in the target node classification pipeline yielded the highest performance for T l o o k b a c k values of 0, 2, and 3. Specifically, these values resulted in average accuracies of 0.625, 0.628 and 0.615, and average weighted F1-scores of 0.630, 0.634 and 0.622, respectively. However, an exception was observed at a T l o o k b a c k value of 1, where GraphSAGE did not exhibit the best performance, resulting in a decrease in average accuracy to 0.594 and weighted F1-score to 0.612. At that specific value of T l o o k b a c k , the GAT outperformed GraphSAGE, achieving an average accuracy of 0.635 and a weighted F1-score of 0.643. Notably, there was a more significant decline in performance for GAT compared to GraphSAGE, particularly at a T l o o k b a c k value of 3. This observation suggests that GAT is more sensitive to higher values of T l o o k b a c k , resulting in a larger drop in classification performance.
Removal of the GNN entirely from the proposed method led to the worst performance across all tested T l o o k b a c k values. However, as the value of T l o o k b a c k increased, there was a noticeable improvement in the results. The average accuracy improved from 0.334 to 0.534, while the average weighted F1-score increased from 0.360 to 0.542. These results demonstrate the necessity of utilising a GNN to achieve high classification performance with the proposed method. GraphSAGE demonstrated its superiority as the best-performing GNN.

5.3. Classification Performance and Discussion

Based on the results from the previous subsection, the GraphSAGE algorithm, utilising LSTM aggregation, showcased the highest performance among all the evaluated GNNs. Hence, it was chosen as the GNN within the target node classification pipeline for the complete performance evaluation of the proposed method, using both regions in the dataset. Unlike with training in the prior subsection, all nodes in each G t r a i n were utilised to train the classification models. The experiments were performed for T l o o k b a c k = 0 , 1 , , 7 for each region in the dataset. The average and Standard Deviation of accuracy, precision, recall and F1-score were calculated for the CLC classification at both class levels 1 and 2 considering all runs. The level 1 classification results were derived by using the level 2 results and combining them according to the predefined class structure of level 1, as defined in the CLC dataset [18].
The training for both regions was undertaken on two distinct platforms. For the region of Graz, a dedicated Deep Learning server was used, equipped with two Intel Xeon E5-2683 v4 processors, 98GB of RAM, and four NVIDIA GeForce GTX TITAN X graphics cards. Here, the average epoch training time varied from 18 min for T l o o k b a c k = 0 to 64 min for T l o o k b a c k = 7. The training times for 10 runs at T l o o k b a c k = 0 , 1 , , 7 were 30, 40, 41, 44, 56, 70, 88 and 107 h, respectively. The total training time was 476 h.
In contrast, the region of Portorož, Izola and Koper was trained on a personal computer powered by an AMD Ryzen 9 7900X processor, 64GB of RAM and an NVIDIA GeForce RTX 4090 graphics card. The average epoch training time was shorter, from 2 min at T l o o k b a c k = 0, extending to 12 min for T l o o k b a c k = 7. The accumulated training times for 10 runs at T l o o k b a c k = 0 , 1 , , 7 were reduced considerably, specifically to 3, 4, 6, 8, 10, 13, 17 and 21 h, respectively. This sums up to a total of 82 h of training time.
The pixel-based CLC level 2 classification results of the T S t e s t _ i m a g e , for both dataset regions, are depicted using various metrics in Table 1 and Table 2. For the Graz region, the Esri’s UNet model achieved an accuracy of 0.818, weighted precision of 0.841, weighted recall of 0.818 and weighted F1-score of 0.824. The proposed GNN-based method achieved better performance across all metrics, but at various sizes of subgraphs. By classifying the subgraphs formed with T l o o k b a c k = 2, an accuracy of 0.831 was achieved. The highest weighted precision of 0.865 was achieved by classifying the subgraphs formed with T l o o k b a c k = 1. Classifying the subgraphs formed with T l o o k b a c k = 2 led to the highest weighted recall of 0.831 and weighted F1-score of 0.841. The highest macro- and micro-averaged values for precision (0.442 and 0.831) and F1-score (0.468 and 0.831) were all achived with subgraphs formed with T l o o k b a c k = 2. The same applies for the micro recall (0.831), however, the highest macro recall value of 0.552 was achieved at T l o o k b a c k = 0. Increasing the value of T l o o k b a c k above 2 led to gradual worsening of the results across all the observed metrics, both in terms of lower averages and higher Standard Deviations.
As for the region of Portorož, Izola and Koper, Esri’s UNet model performed better than the proposed GNN-based method based on every metric. The UNet model’s performance statistics were: accuracy—0.792, weighted precision—0.827, macro precision—0.578, and micro precision—0.792. Equally, it demonstrated a consistent weighted and micro recall of 0.792, with a lower macro recall of 0.677. The F1-scores were 0.801 (weighted), 0.589 (macro), and 0.792 (micro). The proposed method achieved the best results when classification was performed of subgraphs formed with T l o o k b a c k = 0. This resulted in an average overall accuracy and all micro measures (precision, recall, and F1-score) of 0.741. The weighted metrics were as follow: precision—0.770, recall—0.741, and F1-score—0.746. In contrast, the macro measures were quite lower, with precision at 0.520, recall at 0.577, and F1-score at 0.529. Similar to the Graz region, an overall decline in performance across all metrics was observed, beginning immediately after the T l o o k b a c k value exceeded 0.
Table 3 and Table 4 present the CLC level 1 pixel-based classification results for the T S t e s t _ i m a g e , for both dataset regions, evaluated via multiple metrics. In the Graz region the GNN-based method, using T l o o k b a c k = 1, outperformed Esri’s UNet model across all evaluation metrics. The comparative averaged scores for the GNN-based method vs. Esri’s UNet model were as follow: overall accuracy (0.888 vs. 0.857), weighted precision (0.890 vs. 0.863), micro precision, weighted recall, and micro recall (all 0.888 vs. 0.857). The results highlight a significant divergence, especially in macro precision, with a substantial difference of 0.309 (0.877 vs. 0.568), and in macro recall with an equally striking difference of 0.337 (0.907 vs. 0.570). Similarly, the F1-scores—weighted, macro, and micro (0.888, 0.889, 0.888 vs. 0.857, 0.567, 0.857) - showed a substantial contrast, proving the superior performance of the proposed GNN-based method. The performance of the proposed method outperformed Esri’s UNet model across the majority of evaluation metrics for all the tested T l o o k b a c k values up to and including 5. However, increasing the T l o o k b a c k value beyond 5 resulted in a degradation of performance.
For the region of Portorož, Izola and Koper, the GNN-based method using T l o o k b a c k = 1 demonstrated superior performance across the majority of the evaluation metrics when compared to Esri’s UNet model. The performance comparison for each averaged metric is as follows: overall accuracy (0.871 vs. 0.864), weighted precision (0.876 vs. 0.869), micro precision (0.871 vs. 0.864), weighted recall (0.871 vs. 0.864), and micro recall (0.871 vs. 0.864). A significant difference of 0.096 was observed in macro recall (0.890 vs. 0.794). The F1-score comparisons were also in favour of the proposed method with weighted, macro, and micro F1-scores being 0.872 vs. 0.862, 0.862 vs. 0.821, and 0.871 vs. 0.864, respectively. The only exception was macro precision, where the UNet model scored higher (0.862) than the proposed method using subgraphs formed with T l o o k b a c k = 0 (0.847). Increasing the value of T l o o k b a c k beyond 1 led to a decline in the performance of the proposed method, compared to Esri’s UNet model, for the majority of the evaluation metrics.
For the CLC level 2 classification of the region of Graz, the models of the proposed method, formed with T l o o k b a c k = 2 (using spatiotemporal information from up to two previous multispectral images), secured the highest average weighted F1-score of 0.841. The model with the best F1-score of 0.846 was selected to create the confusion matrix for the classification of CLC class level 2, as depicted in Figure 12a. For the CLC level 1 classification, the highest average weighted F1-score of 0.888 was achieved by models utilising T l o o k b a c k = 1, employing information from up to one prior multispectral image. The model boasting the top F1-score of 0.900 was used to generate the confusion matrix for CLC class level 1, showcased in Figure 12b. These confusion matrices provide insights into per-class accuracy and misclassifications exhibited by the proposed GNN-based method.
The results in the confusion matrix in Figure 12a highlight the connection between the accuracy and the number of training samples per class. High classification accuracies were achieved for classes with a large number of training samples, namely, Urban fabric (87.99% with 47,341 subgraphs), Industrial, commercial and transport units (85.98% with 24,658 subgraphs), Arable land (78.53% with 55,852 subgraphs), Forests (90.03% with 25,037 subgraphs) and Inland waters (98.66% with 4834 subgraphs). The only exception to this pattern was Artificial, non-agricultural vegetated areas with an accuracy of 83.15% and 821 training subgraphs, which is considerably less compared to other classes with high classification accuracy. The classes with a smaller number of training samples were classified poorly. Namely, Scrub and/or herbaceous vegetation associations achieved a classification accuracy of 0% by being wrongly mistaken in 68.80% of pixels for a Forest, and also to a lesser extent for other similar land cover classes. The poor accuracy results were also achieved for Mine, dump and construction sites (44.17%), Pastures (8.11%) and Heterogeneous agricultural areas (41.65%).
As for the confusion matrix for CLC level 1 classification in Figure 12b, the best performing model of the proposed GNN-based method achieved accuracy of 91.69% for Artificial surfaces, 87.28% for Agricultural areas, 90.84% for Forest and seminatural areas and 96.90% for Water bodies. The most common misclassifications occurred between Artificial surfaces and Agricultural areas.
For the CLC level 2 classification of the Portorož, Izola and Koper region, the proposed method’s models, using T l o o k b a c k = 0 (i.e., without using any spatiotemporal information from previous multispectral images), achieved the highest average weighted F1-score of 0.764. The top-performing model, boasting an F1-score of 0.764, was utilised to construct the confusion matrix for CLC level 2 classification, as illustrated in Figure 13a. As for the CLC level 1 classification, the models with T l o o k b a c k = 1 (which used data from one preceding multispectral image) obtained the highest average weighted F1-score of 0.872. The leading model, with an F1-score of 0.889, was employed to develop the confusion matrix for CLC level 1 classification, presented in Figure 13b.
The confusion matrix for CLC level 2 classification in Figure 13a illustrates the good performance of the proposed GNN-based method for achieving accuracy above 90% in classifying Industrial, commercial, and transport units (90.80%), Maritime wetlands (96.82%), Inland waters (98.55%), and Marine waters (99.17%). The model also performed well for Urban fabric (84.69%) and Forests (79.83%). Just as in the case of the region of Graz, the results for classes with fewer training samples (subgraphs) were suboptimal. For example, the Open spaces with little or no vegetation class, having 22 subgraphs, attained an overall classification accuracy of 6.63%. Similarly, the Pastures class, with a mere 48 training subgraphs, achieved a classification accuracy of just 1.97%.
The top-performing model of the proposed GNN-based method, as depicted in the CLC level 1 classification confusion matrix in Figure 13b, attained accuracies of 89.66% for Artificial surfaces, 87.04% for Agricultural areas, 76.02% for Forest and seminatural areas, 97.88% for Wetlands and 99.15% for Water bodies. The most common misclassifications were observed between Agricultural areas and Forest and seminatural areas, a trend observed previously in the more detailed CLC level 2 classification confusion matrix in Figure 13a.
For both regions of the dataset, the best proposed method’s models were used to calculate individual weighted F1-scores for CLC level 2 classification of all images in the T S t e s t _ i m a g e , as illustrated in Figure 14. For the region of Graz these scores ranged from 0.813 to 0.889, averaging at 0.860. However, the range was wider for the region of Portorož, Izola and Koper, with scores as low as 0.65 to as high as 0.801, and an average score of 0.739. The variance in these scores for the region of Portorož, Izola and Koper was nearly double (0.0015) that of the Graz region (0.0008).
The pixel-based heatmaps, depicted in Figure 15, illustrate the spatial distribution of cumulative inaccuracies in the CLC level 2 classification for both dataset regions. A subset of these land cover predictions is depicted in Figure 16. Notably, for the Graz region, the urban and industrial areas experienced the highest frequency of misclassifications, followed closely by agricultural zones. In the region of Portorož, Izola and Koper, the seminatural and agricultural landscapes were the most common sites of classification errors.
The findings demonstrate that the proposed GNN-based method outperformed Esri’s UNet model consistently across all datasets for CLC level 1 classification. However, when it comes to CLC level 2 classification, the GNN-based method’s performance was on par with that of Esri’s UNet model. While the proposed method achieved better results across all evaluation metrics in the Graz region, it underperformed in the region of Portorož, Izola and Koper.
Various factors contribute to misclassifications and performance degradation as the value of T l o o k b a c k increases. This is evident in Figure 17, which displays statistics of subgraphs for both dataset regions. For the region of Graz, a decline in average accuracy and weighted F1-score begins when T l o o k b a c k equals 3. This decline aligns with the average cumulative spatial coverage of segments in both training and testing subgraphs exceeding ≈0.07 km 2 . The region of Portorož, Izola and Koper displays a different pattern. The average cumulative spatial coverage of segments in the subgraphs are significantly larger than in Graz for all the tested T l o o k b a c k values. The achieved metric scores start lower in this region in comparison to Graz. The degradation begins immediately as we increase T l o o k b a c k above zero. In the region of Portorož, Izola and Koper, it is evident that, as the number of nodes in G s u b increases, there is a growing disparity in the average cumulative spatial coverage of segments between the training and testing subgraphs. While this disparity could potentially cause a decline in performance, it’s worth noting that such a trend was not observed in the Graz region.
The number of nodes in a G s u b doesn’t influence classification performance directly. However, it does control spatial coverage, which affects classification performance significantly. Hence, the primary cause of poor classification performance are overly large spatial regions captured by larger subgraphs through time. Such expansive regions can contain conflicting segment information, resulting in misclassifications. This issue is mainly a result of segmentation parameters, although an increase in the value of T l o o k b a c k also plays a contributing role. The described issue with conflicting segment information is particularly evident for classes like Industrial, commercial and transport units, Mine, dump and construction sites, as well as Arable land, Pastures, and Heterogeneous agricultural areas, as visualised by the confusion matrices in Figure 12a and Figure 13a.
Dataset imbalance impacts the class-specific accuracies for both dataset regions, as demonstrated in Figure 18. The trend lines suggest higher accuracies are associated with classes that occupy a larger proportion of node labels within the G t r a i n . However, several classes have emerged as exceptions to this rule. For instance, in the Graz region, only 2.96% of nodes in the G t r a i n are labelled as Inland waters, but this still results in an accuracy of 98.66%. To achieve better performance, complex classes such as Pastures and Open spaces with little or no vegetation require a greater quantity of nodes with corresponding labels.
The proposed method faces certain limitations. Firstly, the segmentation parameters should be selected carefully, because small segments result in G containing a large set of nodes. This leads to the increase of the processing time required to classify all target nodes within the G. Secondly, a smaller value of the τ parameter during graph construction leads to more edges being established between nodes, consequently enlarging the G s u b of each individual v t a r g e t . Larger G s u b can potentially capture a too broad spatial region through time, and may introduce conflicting information from the segments, thereby undermining the classification results. Additionally, larger subgraphs result in slower training and extended inference times, which limits the efficiency of the proposed GNN-based method.

6. Conclusions

This paper presented a novel spatiotemporal method for an object-based land cover classification of satellite imagery using a GNN. A 4-year intermonthly land cover ground truth dataset of Sentinel-2 imagery was created for two regions: Graz in Austria, and the region of Portorož, Izola and Koper in Slovenia. This newly created dataset, classified in accordance with the CLC level 2 nomenclature, was utilised to evaluate the performance of the proposed method. For testing purposes, the method’s classification pipeline used EfficientNetV2-S for image feature extraction, and the GraphSAGE algorithm with LSTM aggregation for target node classification.
For the CLC level 2 classification of the region of Graz, the proposed method outperformed the state-of-the-art UNet model across all evaluation metrics. The proposed method yielded the best average weighted F1-score of 0.841 and an accuracy of 0.831, which surpassed the UNet model’s results of 0.824 and 0.818, respectively. For the classification of the region of Portorož, Izola and Koper, the UNet model achieved better results compared to the proposed method. The proposed method achieved a peak average weighted F1-score of 0.746 and accuracy of 0.741, underperforming the UNet model’s results of 0.801 and 0.792, respectively.
However, by combining the CLC level 2 into the predefined CLC level 1 classification, the proposed method demonstrated far superior performance for both regions. The proposed method yielded the best average weighted F1-score and accuracy of 0.888 for the Graz region, outperforming the UNet model’s performance, which achieved a score of 0.857 for both metrics. Similarly, for the region of Portorož, Izola and Koper, the proposed method achieved the highest average weighted F1-score of 0.872 and accuracy of 0.871, beating the UNet model, which scored 0.862 and 0.864, respectively.
The findings indicate that the classification performance of the proposed GNN-method starts to decline when an excessive amount of spatiotemporal information from numerous preceding multispectral images is included in the subgraphs. The proposed method exhibits limitations related to the choice of segmentation and graph construction parameters.
In the future, the proposed method’s parameter of maximum distance of nodes based on input edges towards the target node will become adaptable using the multispectral data. The proposed method will also be extended with additional types of input data, and modified to perform object-based regression tasks, e.g., particulate matter trends.

Author Contributions

Conceptualization, D.K.; methodology, D.K.; software, D.K.; validation, D.M., B.Ž. and N.L.; formal analysis, D.K.; investigation, D.K.; resources, D.K.; data curation, D.K.; writing—original draft preparation, D.K.; writing—review and editing, D.K.; visualization, D.K.; supervision, N.L.; project administration, N.L.; funding acquisition, D.M. and B.Ž. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Slovenian Research Agency grant number P2-0041 and L7-2633.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this paper are available on request from the corresponding author. The data are not publicly available due to ongoing analysis and publication commitments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bhandari, A.; Joshi, R.; Thapa, M.S.; Sharma, R.P.; Rauniyar, S.K. Land Cover Change and Its Impact in Crop Yield: A Case Study from Western Nepal. Sci. World J. 2022, 2022, 5129423. [Google Scholar] [CrossRef] [PubMed]
  2. Hussain, S.; Mubeen, M.; Ahmad, A.; Majeed, H.; Qaisrani, S.; Hammad, H.; Amjad, M.; Ahmad, I.; Fahad, S.; Ahmad, N.; et al. Assessment of land use/land cover changes and its effect on land surface temperature using remote sensing techniques in Southern Punjab, Pakistan. Environ. Sci. Pollut. Res. 2022. [Google Scholar] [CrossRef] [PubMed]
  3. Som-ard, J.; Immitzer, M.; Vuolo, F.; Ninsawat, S.; Atzberger, C. Mapping of crop types in 1989, 1999, 2009 and 2019 to assess major land cover trends of the Udon Thani Province, Thailand. Comput. Electron. Agric. 2022, 198, 107083. [Google Scholar] [CrossRef]
  4. Koetz, B.; Morsdorf, F.; van der Linden, S.; Curt, T.; Allgöwer, B. Multi-source land cover classification for forest fire management based on imaging spectrometry and LiDAR data. For. Ecol. Manag. 2008, 256, 263–271. [Google Scholar] [CrossRef]
  5. Hao, L.; van Westen, C.; Rajaneesh, A.; Sajinkumar, K.; Martha, T.R.; Jaiswal, P. Evaluating the relation between land use changes and the 2018 landslide disaster in Kerala, India. CATENA 2022, 216, 106363. [Google Scholar] [CrossRef]
  6. Shuaishuai, J.; Yang, C.; Wang, M.; Failler, P. Heterogeneous Impact of Land-Use on Climate Change: Study From a Spatial Perspective. Front. Environ. Sci. 2022, 10, 1–17. [Google Scholar]
  7. Aslam, S.; Chak, Y.C.; Hussain Jaffery, M.; Varatharajoo, R.; Ahmad Ansari, E. Model predictive control for Takagi–Sugeno fuzzy model-based Spacecraft combined energy and attitude control system. Adv. Space Res. 2023, 71, 4155–4172. [Google Scholar] [CrossRef]
  8. Yahya, N.; Varatharajoo, R.; Mohd Harithuddin, A.S. Satellite Formation Flying Relative Geodesic and Latitudinal Error Measures. J. Aeronaut. Astronaut. Aviat. Ser. A 2020, 52, 83–94. [Google Scholar]
  9. Li, J.; Chen, B. Global Revisit Interval Analysis of Landsat-8 -9 and Sentinel-2A -2B Data for Terrestrial Monitoring. Sensors 2020, 20, 6631. [Google Scholar] [CrossRef]
  10. Yin, J.; Dong, J.; Hamm, N.A.; Li, Z.; Wang, J.; Xing, H.; Fu, P. Integrating remote sensing and geospatial big data for urban land use mapping: A review. Int. J. Appl. Earth Obs. Geoinf. 2021, 103, 102514. [Google Scholar] [CrossRef]
  11. Hansen, M.C.; Loveland, T.R. A review of large area monitoring of land cover change using Landsat data. Remote Sens. Environ. 2012, 122, 66–74. [Google Scholar] [CrossRef]
  12. Phiri, D.; Simwanda, M.; Salekin, S.; Nyirenda, V.R.; Murayama, Y.; Ranagalage, M. Sentinel-2 Data for Land Cover/Use Mapping: A Review. Remote Sens. 2020, 12, 2291. [Google Scholar] [CrossRef]
  13. Zhu, Z.; Wang, S.; Woodcock, C.E. Improvement and expansion of the Fmask algorithm: Cloud, cloud shadow, and snow detection for Landsats 4–7, 8, and Sentinel 2 images. Remote Sens. Environ. 2015, 159, 269–277. [Google Scholar] [CrossRef]
  14. Frantz, D.; Haß, E.; Uhl, A.; Stoffels, J.; Hill, J. Improvement of the Fmask algorithm for Sentinel-2 images: Separating clouds from bright surfaces based on parallax effects. Remote Sens. Environ. 2018, 215, 471–481. [Google Scholar] [CrossRef]
  15. Dupuy, S.; Gaetano, R. Reunion Island-2019, Land Cover Map (Spot6/7)-1.5 m. 2020. Available online: https://dataverse.cirad.fr/dataset.xhtml?persistentId=doi:10.18167/DVN1/YZJQ7Q (accessed on 12 July 2023).
  16. Censi, A.M.; Ienco, D.; Gbodjo, Y.J.E.; Pensa, R.G.; Interdonato, R.; Gaetano, R. Attentive Spatial Temporal Graph CNN for Land Cover Mapping From Multi Temporal Remote Sensing Data. IEEE Access 2021, 9, 23070–23082. [Google Scholar] [CrossRef]
  17. Heymann, Y.; Steenmans, C.; Croisille, G.; Bossard, M.; Lenco, M.; Wyatt, B.; Jean-Louis, W.; O’Brian, C.; Cornaert, M.-H.; Nicolas, S. Corine Land Cover Technical Guide, Part I; Commission of the European Communities: Mestreech, The Netherlands, 1994. [Google Scholar]
  18. Copernicus Land Monitoring Service 2018. 2018. Available online: https://land.copernicus.eu/pan-european/corine-land-cover/clc2018 (accessed on 12 July 2023).
  19. Soukup, T.; Feranec, J.; Hazeu, G.; Jaffrain, G.; Jindrova, M.; Kopecky, M.; Orlitova, E. Chapter 10 CORINE Land Cover 1990 (CLC1990): Analysis and Assessment: CORINE Land Cover Data. In European Landscape Dynamics; Feranec, J., Soukup, T., Hazeu, G., Jaffrain, G., Eds.; Taylor & Francis Group: London, UK, 2016; pp. 69–78. [Google Scholar]
  20. Buttner, G.; Feranec, J.; Jaffrain, G.; Mari, L.; Maucha, G.; Soukup, T. The CORINE land cover 2000 project. EARSeL Eproceedings 2004, 3, 331–346. [Google Scholar]
  21. Soukup, T.; Feranec, J.; Hazeu, G.; Jaffrain, G.; Jindrova, M.; Kopecky, M.; Orlitova, E. Chapter 12 CORINE Land Cover 2006 (CLC2006): Analysis and Assessment: CORINE Land Cover Data. In European Landscape Dynamics; Feranec, J., Soukup, T., Hazeu, G., Jaffrain, G., Eds.; Taylor & Francis Group: London, UK, 2016; pp. 87–92. [Google Scholar]
  22. Soukup, T.; Büttner, G.; Feranec, J.; Hazeu, G.; Jaffrain, G.; Jindrova, M.; Kopecky, M.; Orlitova, E. Chapter 13 CORINE Land Cover 2012 (CLC2012): Analysis and Assessment: CORINE Land Cover Data. In European Landscape Dynamics; Feranec, J., Soukup, T., Hazeu, G., Jaffrain, G., Eds.; Taylor & Francis Group: London, UK, 2016; pp. 93–98. [Google Scholar]
  23. Aune-Lundberg, L.; Strand, G.H. The content and accuracy of the CORINE Land Cover dataset for Norway. Int. J. Appl. Earth Obs. Geoinf. 2020, 96, 102266. [Google Scholar] [CrossRef]
  24. Eurostat. LUCAS—EU Land Use and Cover Area Survey—2021 Edition; EU: Mestreech, The Netherlands, 2021. [Google Scholar]
  25. Landa, M.; Brodský, L.; Halounová, L.; Bouček, T.; Pešek, O. Open Geospatial System for LUCAS In Situ Data Harmonization and Distribution. ISPRS Int. J. Geo-Inf. 2022, 11, 361. [Google Scholar] [CrossRef]
  26. Brown, C.; Brumby, S.; Guzder-Williams, B.; Birch, T.; Hyde, S.; Mazzariello, J.; Czerwinski, W.; Pasquarella, V.; Haertel, R.; Ilyushchenko, S.; et al. Dynamic World, Near real-time global 10 m land use land cover mapping. Sci. Data 2022, 9, 251. [Google Scholar] [CrossRef]
  27. Hofierka, J.; Gallay, M.; Onačillová, K.; Hofierka, J. Physically-based land surface temperature modeling in urban areas using a 3-D city model and multispectral satellite data. Urban Clim. 2020, 31, 100566. [Google Scholar] [CrossRef]
  28. Li, S.; Xiong, L.; Tang, G.; Strobl, J. Deep learning-based approach for landform classification from integrated data sources of digital elevation model and imagery. Geomorphology 2020, 354, 107045. [Google Scholar] [CrossRef]
  29. Gaur, S.; Singh, R. A Comprehensive Review on Land Use/Land Cover (LULC) Change Modeling for Urban Development: Current Status and Future Prospects. Sustainability 2023, 15, 903. [Google Scholar] [CrossRef]
  30. Gašparović, M.; Jogun, T. The Effect of Fusing Sentinel-2 Bands on Land-Cover Classification. Int. J. Remote Sens. 2018, 39, 822–841. [Google Scholar] [CrossRef]
  31. Gómez, C.; White, J.C.; Wulder, M.A. Optical remotely sensed time series data for land cover classification: A review. ISPRS J. Photogramm. Remote Sens. 2016, 116, 55–72. [Google Scholar] [CrossRef] [Green Version]
  32. Zhang, C.; Yue, P.; Tapete, D.; Shangguan, B.; Wang, M.; Wu, Z. A multi-level context-guided classification method with object-based convolutional neural network for land cover classification using very high resolution remote sensing images. Int. J. Appl. Earth Obs. Geoinf. 2020, 88, 102086. [Google Scholar] [CrossRef]
  33. Mongus, D.; Žalik, B. Segmentation schema for enhancing land cover identification: A case study using Sentinel 2 data. Int. J. Appl. Earth Obs. Geoinf. 2018, 66, 56–68. [Google Scholar] [CrossRef]
  34. Yang, C.; Rottensteiner, F.; Heipke, C. A hierarchical deep learning framework for the consistent classification of land use objects in geospatial databases. ISPRS J. Photogramm. Remote Sens. 2021, 177, 38–56. [Google Scholar] [CrossRef]
  35. Fitton, D.; Laurens, E.; Hongkarnjanakul, N.; Schwob, C.; Mezeix, L. Land cover classification through Convolutional Neural Network model assembly: A case study of a local rural area in Thailand. Remote Sens. Appl. Soc. Environ. 2022, 26, 100740. [Google Scholar]
  36. Fu, J.; Yi, X.; Wang, G.; Mo, L.; Wu, P.; Kapula, K.E. Research on Ground Object Classification Method of High Resolution Remote-Sensing Images Based on Improved DeeplabV3+. Sensors 2022, 22, 7477. [Google Scholar] [CrossRef]
  37. Li, M.; Lu, Y.; Cao, S.; Wang, X.; Xie, S. A Hyperspectral Image Classification Method Based on the Nonlocal Attention Mechanism of a Multiscale Convolutional Neural Network. Sensors 2023, 23, 3190. [Google Scholar] [CrossRef]
  38. Li, J.; Wang, H.; Zhang, A.; Liu, Y. Semantic Segmentation of Hyperspectral Remote Sensing Images Based on PSE-UNet Model. Sensors 2022, 22, 9678. [Google Scholar] [CrossRef]
  39. Abidi, A.; Ienco, D.; Abbes, A.B.; Farah, I.R. Combining 2D encoding and convolutional neural network to enhance land cover mapping from Satellite Image Time Series. Eng. Appl. Artif. Intell. 2023, 122, 106152. [Google Scholar] [CrossRef]
  40. Qiu, C.; Mou, L.; Schmitt, M.; Zhu, X.X. Local climate zone-based urban land cover classification from multi-seasonal Sentinel-2 images with a recurrent residual network. ISPRS J. Photogramm. Remote Sens. 2019, 154, 151–162. [Google Scholar] [CrossRef]
  41. Dantas, C.F.; Marcos, D.; Ienco, D. Counterfactual Explanations for Land Cover Mapping in a Multi-class Setting. arXiv 2023, arXiv:cs.LG/2301.01520. [Google Scholar]
  42. Chen, B.; Zheng, H.; Wang, L.; Hellwich, O.; Chen, C.; Yang, L.; Liu, T.; Luo, G.; Bao, A.; Chen, X. A joint learning Im-BiLSTM model for incomplete time-series Sentinel-2A data imputation and crop classification. Int. J. Appl. Earth Obs. Geoinf. 2022, 108, 102762. [Google Scholar] [CrossRef]
  43. Jiang, Z.; Yang, S.; Liu, Z.; Xu, Y.; Xiong, Y.; Qi, S.; Pang, Q.; Xu, J.; Liu, F.; Xu, T. Coupling machine learning and weather forecast to predict farmland flood disaster: A case study in Yangtze River basin. Environ. Model. Softw. 2022, 155, 105436. [Google Scholar] [CrossRef]
  44. Aamir, M.; Ali, T.; Irfan, M.; Shaf, A.; Azam, M.Z.; Glowacz, A.; Brumercik, F.; Glowacz, W.; Alqhtani, S.; Rahman, S. Natural Disasters Intensity Analysis and Classification Based on Multispectral Images Using Multi-Layered Deep Convolutional Neural Network. Sensors 2021, 21, 2648. [Google Scholar] [CrossRef]
  45. Siddiqui, M.K.; Imran, M.; Ahmad, A. On Zagreb indices, Zagreb polynomials of some nanostar dendrimers. Appl. Math. Comput. 2016, 280, 132–139. [Google Scholar] [CrossRef]
  46. Ahmad, A.; Bača, M.; Siddiqui, M.K. On Edge Irregular Total Labeling of Categorical Product of Two Cycles. Theory Comput. Syst. 2013, 54, 1–12. [Google Scholar] [CrossRef]
  47. Azeem, M.; Imran, M.; Nadeem, M.F. Sharp bounds on partition dimension of hexagonal Möbius ladder. J. King Saud Univ. Sci. 2022, 34, 101779. [Google Scholar] [CrossRef]
  48. Pang, H.E.; Biljecki, F. 3D building reconstruction from single street view images using deep learning. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102859. [Google Scholar] [CrossRef]
  49. Ding, Y.; Zhang, Z.; Zhao, X.; Hong, D.; Li, W.; Cai, W.; Zhan, Y. AF2GNN: Graph convolution with adaptive filters and aggregator fusion for hyperspectral image classification. Inf. Sci. 2022, 602, 201–219. [Google Scholar] [CrossRef]
  50. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations, Virtual, 3–7 May 2021. [Google Scholar]
  51. Tan, M.; Le, Q.V. EfficientNetV2: Smaller Models and Faster Training. In Proceedings of the 38th International Conference on Machine Learning, Virtual, 18–24 July 2021. [Google Scholar]
  52. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv 2020, arXiv:cs.LG/1905.11946. [Google Scholar]
  53. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  54. Xie, S.; Girshick, R.; Dollar, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. arXiv 2017, arXiv:1611.05431v2. [Google Scholar]
  55. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. empharXiv 2016, arXiv:1512.00567. [Google Scholar]
  56. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K. Densely Connected Convolutional Networks. arXiv 2017, arXiv:1608.06993. [Google Scholar]
  57. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; Bengio, Y. Graph Attention Networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  58. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive Representation Learning on Large Graphs. In Proceedings of the 31st Conference on Neural Information Processing Systems (NeurIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  59. Kaiser, P.; Wegner, J.; Lucchi, A.; Jaggi, M.; Hofmann, T.; Schindler, K. Learning Aerial Image Segmentation From Online Maps. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6054–6068. [Google Scholar] [CrossRef]
  60. Nasir, S.M.; Kamran, K.V.; Blaschke, T.; Karimzadeh, S. Change of land use / land cover in kurdistan region of Iraq: A semi-automated object-based approach. Remote Sens. Appl. Soc. Environ. 2022, 26, 100713. [Google Scholar] [CrossRef]
  61. Liu, T.; Abd-Elrahman, A. Deep convolutional neural network training enrichment using multi-view object-based analysis of Unmanned Aerial systems imagery for wetlands classification. ISPRS J. Photogramm. Remote Sens. 2018, 139, 154–170. [Google Scholar] [CrossRef]
  62. Herawan, A.; Julzarika, A.; Hakim, P.; Asti, E. Object-Based on Land Cover Classification on LAPAN-A3 Satellite Imagery Using Tree Algorithm (Case Study: Rote Island). Int. J. Adv. Sci. Eng. Inf. Technol.y 2021, 11, 2254–2260. [Google Scholar] [CrossRef]
  63. Vizzari, M. PlanetScope, Sentinel-2, and Sentinel-1 Data Integration for Object-Based Land Cover Classification in Google Earth Engine. Remote Sens. 2022, 14, 2628. [Google Scholar] [CrossRef]
  64. Hedayati, A.; Vahidnia, M.H.; Behzadi, S. Paddy lands detection using Landsat-8 satellite images and object-based classification in Rasht city, Iran. Egypt. J. Remote Sens. Space Sci. 2022, 25, 73–84. [Google Scholar] [CrossRef]
  65. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. In Proceedings of the 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, 24–26 April 2017. [Google Scholar]
  66. Zhao, L.; Song, Y.; Zhang, C.; Liu, Y.; Wang, P.; Lin, T.; Deng, M.; Li, H. T-GCN: A Temporal Graph Convolutional Network for Traffic Prediction. IEEE Trans. Intell. Transp. Syst. 2020, 21, 3848–3858. [Google Scholar] [CrossRef] [Green Version]
  67. Zhang, Z.; Huang, J.; Tan, Q. SR-HGAT: Symmetric Relations Based Heterogeneous Graph Attention Network. IEEE Access 2020, 8, 165631–165645. [Google Scholar] [CrossRef]
  68. Hu, B.; Guo, K.; Wang, X.; Zhang, J.; Zhou, D. RRL-GAT: Graph Attention Network-Driven Multilabel Image Robust Representation Learning. IEEE Internet Things J. 2022, 9, 9167–9178. [Google Scholar] [CrossRef]
  69. Ying, R.; He, R.; Chen, K.; Eksombatchai, P.; Hamilton, W.L.; Leskovec, J. Graph Convolutional Neural Networks for Web-Scale Recommender Systems. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2018, London, UK, 19–23 August 2018; pp. 974–983. [Google Scholar]
  70. Zhao, W.; Peng, S.; Chen, J.; Peng, R. Contextual-Aware Land Cover Classification With U-Shaped Object Graph Neural Network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6510705. [Google Scholar] [CrossRef]
  71. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC Superpixels Compared to State-of-the-Art Superpixel Methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [Green Version]
  72. Stutz, D.; Hermans, A.; Leibe, B. Superpixels: An evaluation of the state-of-the-art. Comput. Vis. Image Underst. 2018, 166, 1–27. [Google Scholar] [CrossRef] [Green Version]
  73. Felzenszwalb, P.F.; Huttenlocher, D.P. Efficient Graph-Based Image Segmentation. Int. J. Comput. Vis. 2004, 59, 167–181. [Google Scholar] [CrossRef]
  74. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  75. Grandini, M.; Bagli, E.; Visani, G. Metrics for Multi-Class Classification: An Overview. arXiv 2020, arXiv:abs/2008.05756. [Google Scholar]
  76. Lever, J.; Krzywinski, M.; Altman, N. Points of Significance: Classification evaluation. Nat. Methods 2016, 13, 603–604. [Google Scholar] [CrossRef]
Figure 1. The proposed method’s workflow with four main steps.
Figure 1. The proposed method’s workflow with four main steps.
Sensors 23 06648 g001
Figure 2. Examples of applying Felzenszwalb’s image segmentation algorithm ( σ = 0.5) on a Sentinel-2 13-layer image with 10 m spatial resolution of an example region of Graz, Austria in July of 2017. A True colour (RGB) composite is shown of the multispectral image.
Figure 2. Examples of applying Felzenszwalb’s image segmentation algorithm ( σ = 0.5) on a Sentinel-2 13-layer image with 10 m spatial resolution of an example region of Graz, Austria in July of 2017. A True colour (RGB) composite is shown of the multispectral image.
Sensors 23 06648 g002
Figure 3. An example of overlap portion calculations between segments, shown in (a), and creation of G with τ = 0.3 in (b) and τ = 0.4 in (c). (a) Overlap portion calculations between the red segment in time t and colored segments in time t + 1 , (b) τ = 0.3, (c) τ = 0.4.
Figure 3. An example of overlap portion calculations between segments, shown in (a), and creation of G with τ = 0.3 in (b) and τ = 0.4 in (c). (a) Overlap portion calculations between the red segment in time t and colored segments in time t + 1 , (b) τ = 0.3, (c) τ = 0.4.
Sensors 23 06648 g003
Figure 4. Examples of G s u b construction for a v t a r g e t (red) in time t at T l o o k b a c k = [0, 2]. The G was constructed for a T S m a s k with 3 images of a small example subregion using τ = 0.2. The orange edges connect all the included (enlarged) nodes in the G s u b . Each included v has a b b o x drawn around the coloured s it represents. G s u b in (a) includes only the v t a r g e t , and the G s u b in (b) contains v t a r g e t and 3 nodes with 3 edges between them.
Figure 4. Examples of G s u b construction for a v t a r g e t (red) in time t at T l o o k b a c k = [0, 2]. The G was constructed for a T S m a s k with 3 images of a small example subregion using τ = 0.2. The orange edges connect all the included (enlarged) nodes in the G s u b . Each included v has a b b o x drawn around the coloured s it represents. G s u b in (a) includes only the v t a r g e t , and the G s u b in (b) contains v t a r g e t and 3 nodes with 3 edges between them.
Sensors 23 06648 g004
Figure 5. Target node classification pipeline, which outputs the land cover class of the s s e l e c t e d by classifying the v t a r g e t based on the input G s u b .
Figure 5. Target node classification pipeline, which outputs the land cover class of the s s e l e c t e d by classifying the v t a r g e t based on the input G s u b .
Sensors 23 06648 g005
Figure 6. Intermonthly T S i m a g e for the region of Graz and the region of Portorož, Izola and Koper. The individual multispectral image contains C = 17 layers. The images in T S i m a g e are visualised with a True colour (RGB) composite.
Figure 6. Intermonthly T S i m a g e for the region of Graz and the region of Portorož, Izola and Koper. The individual multispectral image contains C = 17 layers. The images in T S i m a g e are visualised with a True colour (RGB) composite.
Sensors 23 06648 g006
Figure 7. Examples of the segmented regions. Images (a,c) show the True colour (RGB) composites, while (b,d) show their respective segmentation masks. (a) The region of Graz in January 2019, (b) The m a s k for the region of Graz in January 2019, (c) The region of Portorož, Izola and Koper in November 2018, (d) The m a s k for the region of Portorož, Izola and Koper in November 2018.
Figure 7. Examples of the segmented regions. Images (a,c) show the True colour (RGB) composites, while (b,d) show their respective segmentation masks. (a) The region of Graz in January 2019, (b) The m a s k for the region of Graz in January 2019, (c) The region of Portorož, Izola and Koper in November 2018, (d) The m a s k for the region of Portorož, Izola and Koper in November 2018.
Sensors 23 06648 g007
Figure 8. Examples of CLC level 2 classification outputs, obtained with the UNet model by Esri, are shown in (a,c). The manually corrected ground truth, derived from respective UNet model outputs, are shown in (b,d). (a) Classification output of Esri’s UNet model for the region of Graz in January 2019, (b) Manually corrected ground truth for the region of Graz in January 2019, (c) Classification output of Esri’s UNet model for the region of Portorož, Izola and Koper in November 2018, (d) Manually corrected ground truth for the region of Portorož, Izola and Koper in November 2018.
Figure 8. Examples of CLC level 2 classification outputs, obtained with the UNet model by Esri, are shown in (a,c). The manually corrected ground truth, derived from respective UNet model outputs, are shown in (b,d). (a) Classification output of Esri’s UNet model for the region of Graz in January 2019, (b) Manually corrected ground truth for the region of Graz in January 2019, (c) Classification output of Esri’s UNet model for the region of Portorož, Izola and Koper in November 2018, (d) Manually corrected ground truth for the region of Portorož, Izola and Koper in November 2018.
Sensors 23 06648 g008
Figure 9. Number of pixels per land cover class for each region in the dataset. Images (a,c) show the class distribution in T S t r a i n _ g t _ c o v e r , while (b,d) show the class distribution in T S t e s t _ g t _ c o v e r . (a) Distribution of ground truth land cover labels of pixels in T S t r a i n _ g t _ c o v e r for the region of Graz, (b) Distribution of ground truth land cover labels of pixels in T S t e s t _ g t _ c o v e r for the region of Graz, (c) Distribution of ground truth land cover labels of pixels in T S t r a i n _ g t _ c o v e r for the region of Portorož, Izola and Koper, (d) Distribution of ground truth land cover labels of pixels in T S t e s t _ g t _ c o v e r for the region of Portorož, Izola and Koper.
Figure 9. Number of pixels per land cover class for each region in the dataset. Images (a,c) show the class distribution in T S t r a i n _ g t _ c o v e r , while (b,d) show the class distribution in T S t e s t _ g t _ c o v e r . (a) Distribution of ground truth land cover labels of pixels in T S t r a i n _ g t _ c o v e r for the region of Graz, (b) Distribution of ground truth land cover labels of pixels in T S t e s t _ g t _ c o v e r for the region of Graz, (c) Distribution of ground truth land cover labels of pixels in T S t r a i n _ g t _ c o v e r for the region of Portorož, Izola and Koper, (d) Distribution of ground truth land cover labels of pixels in T S t e s t _ g t _ c o v e r for the region of Portorož, Izola and Koper.
Sensors 23 06648 g009
Figure 10. Number of nodes per land cover class for each region in the dataset. Images (a,c) shows the class distribution in G t r a i n , while (b,d) show the class distribution in G t e s t . (a) Distribution of ground truth land cover labels of nodes in G t r a i n for the region of Graz, (b) Distribution of ground truth land cover labels of nodes in G t e s t for the region of Graz, (c) Distribution of ground truth land cover labels of nodes in G t r a i n for the region of Portorož, Izola and Koper, (d) Distribution of ground truth land cover labels of nodes in G t e s t for the region of Portorož, Izola and Koper.
Figure 10. Number of nodes per land cover class for each region in the dataset. Images (a,c) shows the class distribution in G t r a i n , while (b,d) show the class distribution in G t e s t . (a) Distribution of ground truth land cover labels of nodes in G t r a i n for the region of Graz, (b) Distribution of ground truth land cover labels of nodes in G t e s t for the region of Graz, (c) Distribution of ground truth land cover labels of nodes in G t r a i n for the region of Portorož, Izola and Koper, (d) Distribution of ground truth land cover labels of nodes in G t e s t for the region of Portorož, Izola and Koper.
Sensors 23 06648 g010
Figure 11. CLC level 2 classification results for a region of Portorož, Izola and Koper, depending on the selection of GNN and the value of T l o o k b a c k .
Figure 11. CLC level 2 classification results for a region of Portorož, Izola and Koper, depending on the selection of GNN and the value of T l o o k b a c k .
Sensors 23 06648 g011
Figure 12. Confusion matrices for classification of the region of Graz, obtained with the best performing classification model of the proposed GNN-based method. (a) Confusion matrix for CLC level 2 classification. (b) Confusion matrix for CLC level 1 classification.
Figure 12. Confusion matrices for classification of the region of Graz, obtained with the best performing classification model of the proposed GNN-based method. (a) Confusion matrix for CLC level 2 classification. (b) Confusion matrix for CLC level 1 classification.
Sensors 23 06648 g012aSensors 23 06648 g012b
Figure 13. Confusion matrices for classification of the region of Portorož, Izola and Koper, obtained with the best performing classification model of the proposed GNN-based method. (a) Confusion matrix for CLC level 2 classification. (b) Confusion matrix for CLC level 1 classification.
Figure 13. Confusion matrices for classification of the region of Portorož, Izola and Koper, obtained with the best performing classification model of the proposed GNN-based method. (a) Confusion matrix for CLC level 2 classification. (b) Confusion matrix for CLC level 1 classification.
Sensors 23 06648 g013
Figure 14. Individual weighted F1-scores for CLC level 2 classification for the consecutive images in T S t e s t _ i m a g e for both regions of the dataset, obtained with the best performing corresponding classification model of the proposed GNN-based method. (a) Results for the 11 consecutive images of the region of Graz. (b) Results for the 12 consecutive images of the region of Portorož, Izola and Koper.
Figure 14. Individual weighted F1-scores for CLC level 2 classification for the consecutive images in T S t e s t _ i m a g e for both regions of the dataset, obtained with the best performing corresponding classification model of the proposed GNN-based method. (a) Results for the 11 consecutive images of the region of Graz. (b) Results for the 12 consecutive images of the region of Portorož, Izola and Koper.
Sensors 23 06648 g014
Figure 15. Pixel -based heatmaps of cumulative incorrect CLC level 2 classifications for both regions of the dataset, derived from classifying the T S t e s t _ i m a g e with the best performing corresponding classification model of the proposed GNN-based method. The colours’ transition from dark purple to bright yellow represent the frequency of misclassifications, with intensifying brightness signifying a higher count of errors. (a) Heatmap for the region of Graz. (b) Heatmap for the region of Portorož, Izola and Koper.
Figure 15. Pixel -based heatmaps of cumulative incorrect CLC level 2 classifications for both regions of the dataset, derived from classifying the T S t e s t _ i m a g e with the best performing corresponding classification model of the proposed GNN-based method. The colours’ transition from dark purple to bright yellow represent the frequency of misclassifications, with intensifying brightness signifying a higher count of errors. (a) Heatmap for the region of Graz. (b) Heatmap for the region of Portorož, Izola and Koper.
Sensors 23 06648 g015
Figure 16. Examples of CLC level 2 ground truth ((a,c,e,g)) and predictions ((b,d,f,h)) for both regions of the dataset, obtained with the best performing corresponding classification model of the proposed GNN-based method. (a) Ground truth—May 2020—region of Graz, (b) Predicted land cover—May 2020—region of Graz, (c) Ground truth—June 2021—region of Graz, (d) Predicted land cover—June 2021—region of Graz, (e) Ground truth—January 2020—region of Portorož, Izola and Koper, (f) Predicted land cover—January 2020—region of Portorož, Izola and Koper, (g) Ground truth—May 2021—region of Portorož, Izola and Koper, (h) Predicted land cover—May 2021—region of Portorož, Izola and Koper.
Figure 16. Examples of CLC level 2 ground truth ((a,c,e,g)) and predictions ((b,d,f,h)) for both regions of the dataset, obtained with the best performing corresponding classification model of the proposed GNN-based method. (a) Ground truth—May 2020—region of Graz, (b) Predicted land cover—May 2020—region of Graz, (c) Ground truth—June 2021—region of Graz, (d) Predicted land cover—June 2021—region of Graz, (e) Ground truth—January 2020—region of Portorož, Izola and Koper, (f) Predicted land cover—January 2020—region of Portorož, Izola and Koper, (g) Ground truth—May 2021—region of Portorož, Izola and Koper, (h) Predicted land cover—May 2021—region of Portorož, Izola and Koper.
Sensors 23 06648 g016
Figure 17. Average node count and cumulative spatial coverage (total size of all segments) in a G s u b , along with metric scores, depending on the value of T l o o k b a c k . (a) Subgraph-related statistics for the region of Graz. (b) Subgraph-related statistics for the region of Portorož, Izola and Koper.
Figure 17. Average node count and cumulative spatial coverage (total size of all segments) in a G s u b , along with metric scores, depending on the value of T l o o k b a c k . (a) Subgraph-related statistics for the region of Graz. (b) Subgraph-related statistics for the region of Portorož, Izola and Koper.
Sensors 23 06648 g017
Figure 18. Class-specific CLC level 2 classification accuracies (sourced from the confusion matrices in Figure 12a and Figure 13a) for both dataset regions in relation to the distribution of ground truth land cover labels in G t r a i n , derived from Figure 10a,c.
Figure 18. Class-specific CLC level 2 classification accuracies (sourced from the confusion matrices in Figure 12a and Figure 13a) for both dataset regions in relation to the distribution of ground truth land cover labels in G t r a i n , derived from Figure 10a,c.
Sensors 23 06648 g018
Table 1. Accuracy, precision, recall and F1-score for CLC level 2 classification of T S t e s t _ i m a g e for a region of Graz. The best result for each metric is marked in bold.
Table 1. Accuracy, precision, recall and F1-score for CLC level 2 classification of T S t e s t _ i m a g e for a region of Graz. The best result for each metric is marked in bold.
Metric
Classification T lookback AccuracyPrecisionRecallF1-Score
Method WeightedMacroMicroWeightedMacroMicroWeightedMacroMicro
Esri’s UNet/0.8180.8410.3490.8180.8180.4650.8180.8240.3710.818
Proposed GNN-based method00.821 ± 0.0160.862 ± 0.0040.432 ± 0.0210.821 ± 0.0160.821 ± 0.0160.552 ± 0.0240.821 ± 0.0160.834 ± 0.0120.458 ± 0.0200.821 ± 0.016
10.828 ± 0.0130.865 ± 0.0050.441 ± 0.0300.828 ± 0.0130.828 ± 0.0130.535 ± 0.0290.828 ± 0.0130.841 ± 0.0100.465 ± 0.0320.828 ± 0.013
20.831 ± 0.0040.858 ± 0.0030.442 ± 0.0310.831 ± 0.0040.831 ± 0.0040.534 ± 0.0400.831 ± 0.0040.841 ± 0.0030.468 ± 0.0340.831 ± 0.004
30.823 ± 0.0160.855 ± 0.0050.409 ± 0.0090.823 ± 0.0160.823 ± 0.0160.509 ± 0.0170.823 ± 0.0160.836 ± 0.0100.432 ± 0.0120.823 ± 0.016
40.800 ± 0.0150.840 ± 0.0150.392 ± 0.0090.800 ± 0.0150.800 ± 0.0150.494 ± 0.0160.800 ± 0.0150.815 ± 0.0140.412 ± 0.0120.800 ± 0.015
50.788 ± 0.0200.836 ± 0.0100.401 ± 0.0270.788 ± 0.0200.788 ± 0.0200.508 ± 0.0280.788 ± 0.0200.805 ± 0.0160.422 ± 0.0300.788 ± 0.020
60.737 ± 0.0300.810 ± 0.0220.386 ± 0.0330.737 ± 0.0300.737 ± 0.0300.482 ± 0.0580.737 ± 0.0300.763 ± 0.0300.393 ± 0.0400.737 ± 0.030
70.636 ± 0.0850.737 ± 0.0750.391 ± 0.0450.636 ± 0.0850.636 ± 0.0850.478 ± 0.0440.636 ± 0.0850.665 ± 0.0910.395 ± 0.0500.636 ± 0.085
Table 2. Accuracy, precision, recall and F1-score for CLC level 2 classification of T S t e s t _ i m a g e for a region of Portorož, Izola and Koper. The best result for each metric is marked in bold.
Table 2. Accuracy, precision, recall and F1-score for CLC level 2 classification of T S t e s t _ i m a g e for a region of Portorož, Izola and Koper. The best result for each metric is marked in bold.
Metric
Classification T lookback AccuracyPrecisionRecallF1-Score
Method WeightedMacroMicroWeightedMacroMicroWeightedMacroMicro
Esri’s UNet/0.7920.8270.5780.7920.7920.6770.7920.8010.5890.792
Proposed GNN-based method00.741 ± 0.0210.770 ± 0.0110.520 ± 0.0200.741 ± 0.0210.741 ± 0.0210.577 ± 0.0140.741 ± 0.0210.746 ± 0.0190.529 ± 0.0210.741 ± 0.021
10.727 ± 0.0220.762 ± 0.0130.512 ± 0.0180.727 ± 0.0220.727 ± 0.0220.575 ± 0.0130.727 ± 0.0220.730 ± 0.0220.522 ± 0.0150.727 ± 0.022
20.709 ± 0.0120.736 ± 0.0090.491 ± 0.0290.709 ± 0.0120.709 ± 0.0120.561 ± 0.0110.709 ± 0.0120.709 ± 0.0110.504 ± 0.0210.709 ± 0.012
30.656 ± 0.0510.714 ± 0.0210.439 ± 0.0380.656 ± 0.0510.656 ± 0.0510.531 ± 0.0290.656 ± 0.0510.660 ± 0.0470.448 ± 0.0510.656 ± 0.051
40.584 ± 0.0370.674 ± 0.0220.369 ± 0.0230.584 ± 0.0370.584 ± 0.0370.492 ± 0.0200.584 ± 0.0370.585 ± 0.0330.362 ± 0.0270.584 ± 0.037
50.476 ± 0.0420.595 ± 0.0430.311 ± 0.0280.476 ± 0.0420.476 ± 0.0420.421 ± 0.0230.476 ± 0.0420.466 ± 0.0490.285 ± 0.0250.476 ± 0.042
60.385 ± 0.0890.557 ± 0.0440.249 ± 0.0290.385 ± 0.0890.385 ± 0.0890.279 ± 0.0530.385 ± 0.0890.395 ± 0.0900.199 ± 0.0360.385 ± 0.089
70.369 ± 0.0460.463 ± 0.0400.192 ± 0.0180.369 ± 0.0460.369 ± 0.0460.206 ± 0.0210.369 ± 0.0460.368 ± 0.0350.157 ± 0.0200.369 ± 0.046
Table 3. Accuracy, precision, recall and F1-score for CLC level 1 classification of T S t e s t _ i m a g e for a region of Graz. The best result for each metric is marked in bold.
Table 3. Accuracy, precision, recall and F1-score for CLC level 1 classification of T S t e s t _ i m a g e for a region of Graz. The best result for each metric is marked in bold.
Metric
Classification T lookback AccuracyPrecisionRecallF1-Score
Method WeightedMacroMicroWeightedMacroMicroWeightedMacroMicro
Esri’s UNet/0.8570.8630.5680.8570.8570.5700.8570.8570.5670.857
Proposed GNN-based method00.881 ± 0.0070.883 ± 0.0060.867 ± 0.0060.881 ± 0.0070.881 ± 0.0070.901 ± 0.0100.881 ± 0.0070.880 ± 0.0080.881 ± 0.0070.881 ± 0.007
10.888 ± 0.0060.890 ± 0.0060.877 ± 0.0080.888 ± 0.0060.888 ± 0.0060.907 ± 0.0080.888 ± 0.0060.888 ± 0.0070.889 ± 0.0060.888 ± 0.006
20.886 ± 0.0030.888 ± 0.0020.872 ± 0.0090.886 ± 0.0030.886 ± 0.0030.906 ± 0.0070.886 ± 0.0030.886 ± 0.0030.886 ± 0.0020.886 ± 0.003
30.882 ± 0.0090.883 ± 0.0080.860 ± 0.0100.882 ± 0.0090.882 ± 0.0090.903 ± 0.0090.882 ± 0.0090.882 ± 0.0090.878 ± 0.0090.882 ± 0.009
40.866 ± 0.0150.868 ± 0.0140.846 ± 0.0170.866 ± 0.0150.866 ± 0.0150.885 ± 0.0260.866 ± 0.0150.865 ± 0.0150.861 ± 0.0200.866 ± 0.015
50.859 ± 0.0110.863 ± 0.0100.838 ± 0.0120.859 ± 0.0110.859 ± 0.0110.884 ± 0.0110.859 ± 0.0110.859 ± 0.0110.855 ± 0.0130.859 ± 0.011
60.822 ± 0.0270.832 ± 0.0240.789 ± 0.0350.822 ± 0.0270.822 ± 0.0270.839 ± 0.0310.822 ± 0.0270.821 ± 0.0280.798 ± 0.0420.822 ± 0.027
70.756 ± 0.0710.770 ± 0.0650.723 ± 0.0690.756 ± 0.0710.756 ± 0.0710.796 ± 0.0590.756 ± 0.0710.756 ± 0.0740.738 ± 0.0740.756 ± 0.071
Table 4. Accuracy, precision, recall and F1-score for CLC level 1 classification of T S t e s t _ i m a g e for a region of Portorož, Izola and Koper. The best result for each metric is marked in bold.
Table 4. Accuracy, precision, recall and F1-score for CLC level 1 classification of T S t e s t _ i m a g e for a region of Portorož, Izola and Koper. The best result for each metric is marked in bold.
Metric
Classification T lookback AccuracyPrecisionRecallF1-Score
Method WeightedMacroMicroWeightedMacroMicroWeightedMacroMicro
Esri’s UNet/0.8640.8690.8620.8640.8640.7940.8640.8620.8210.864
Proposed GNN-based method00.864 ± 0.0230.873 ± 0.0110.847 ± 0.0140.864 ± 0.0230.864 ± 0.0230.882 ± 0.0160.864 ± 0.0230.864 ± 0.0240.859 ± 0.0180.864 ± 0.023
10.871 ± 0.0140.876 ± 0.0100.844 ± 0.0200.871 ± 0.0140.871 ± 0.0140.890 ± 0.0100.871 ± 0.0140.872 ± 0.0130.862 ± 0.0160.871 ± 0.014
20.861 ± 0.0120.865 ± 0.0090.830 ± 0.0190.861 ± 0.0120.861 ± 0.0120.882 ± 0.0070.861 ± 0.0120.861 ± 0.0120.850 ± 0.0160.861 ± 0.012
30.841 ± 0.0260.843 ± 0.0240.793 ± 0.0450.841 ± 0.0260.841 ± 0.0260.859 ± 0.0220.841 ± 0.0260.840 ± 0.0260.814 ± 0.0470.841 ± 0.026
40.793 ± 0.0270.803 ± 0.0220.712 ± 0.0370.793 ± 0.0270.793 ± 0.0270.814 ± 0.0190.793 ± 0.0270.788 ± 0.0300.722 ± 0.0400.793 ± 0.027
50.734 ± 0.0330.750 ± 0.0260.646 ± 0.0390.734 ± 0.0330.734 ± 0.0330.752 ± 0.0280.734 ± 0.0330.730 ± 0.0250.644 ± 0.0460.734 ± 0.033
60.623 ± 0.0720.710 ± 0.0330.571 ± 0.0340.623 ± 0.0720.623 ± 0.0720.586 ± 0.0880.623 ± 0.0720.638 ± 0.0560.515 ± 0.0420.623 ± 0.072
70.572 ± 0.0670.615 ± 0.0500.484 ± 0.0470.572 ± 0.0670.572 ± 0.0670.461 ± 0.0480.572 ± 0.0670.572 ± 0.0630.451 ± 0.0530.572 ± 0.067
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kavran, D.; Mongus, D.; Žalik, B.; Lukač, N. Graph Neural Network-Based Method of Spatiotemporal Land Cover Mapping Using Satellite Imagery. Sensors 2023, 23, 6648. https://doi.org/10.3390/s23146648

AMA Style

Kavran D, Mongus D, Žalik B, Lukač N. Graph Neural Network-Based Method of Spatiotemporal Land Cover Mapping Using Satellite Imagery. Sensors. 2023; 23(14):6648. https://doi.org/10.3390/s23146648

Chicago/Turabian Style

Kavran, Domen, Domen Mongus, Borut Žalik, and Niko Lukač. 2023. "Graph Neural Network-Based Method of Spatiotemporal Land Cover Mapping Using Satellite Imagery" Sensors 23, no. 14: 6648. https://doi.org/10.3390/s23146648

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop