remotesensing-logo

Journal Browser

Journal Browser

Remote Sensing Cross-Modal Research: Algorithms and Practices

A special issue of Remote Sensing (ISSN 2072-4292). This special issue belongs to the section "Remote Sensing Image Processing".

Deadline for manuscript submissions: 31 August 2024 | Viewed by 1393

Special Issue Editors


E-Mail Website
Guest Editor
College of Physics and Information Engineering, Fuzhou University, Fuzhou 350108, China.
Interests: cross-domain scene classification; multi-modal image analysis; cross-modal image interpretation
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
Hangzhou Institute of Technology, Xidian University, Hangzhou 311200, China
Interests: cross-modal image recognition; multi-source information joint perception; remote sesing application

Special Issue Information

Dear Colleagues,

With the development of remote sensing technology, multi-modal remote sensing data have become widely available, including, but not limited to, optical, SAR (synthetic aperture radar), thermal images, LIDAR, hyperspectral images and text. Different modalities in remote sensing data are highly correlated in specific events and applications. Remote sensing cross-modal research can help us understand and employ remote sensing data more comprehensively. For example, in urban planning, studying the cross-modal interaction of optical image and LIDAR data can enable us to conduct more accurate three-dimensional modeling and terrain analysis. In environmental monitoring, the cross-modal analysis of optical and thermal imaging data can enable us to conduct surface temperature monitoring. In agricultural resource management, cross-modal research on hyperspectral and SAR data can enable us to conduct crop growth monitoring and soil humidity analysis. Therefore, the current research on remote sensing cross-modal research has received significant attention from academic and industrial circles. Various types of remote sensing cross-modal tasks have been proposed, such as cross-model retrieval between remote sensing images and videos, remote sensing image caption, video abstract extraction, visual question answering and so on. Remote sensing cross-modal data possess heterogeneous features at the bottom and related semantics at the top. Determining how to represent the underlying features of different modal data domains, extract the high-level semantics and model the correlation between different modalities are major challenges faced by remote sensing cross-modal research. New algorithms and methods must be developed in order to process and analyze cross-modal remote sensing data.

This Special Issue aims to promote exchange and cooperation in cross-modal research, and promote the development and application of remote sensing science. Topics may cover anything from remote sensing cross-modal retrieval to more comprehensive aims and scales. Articles may address, but are not limited, to the following topics:

  • Cross-modal remote sensing classification, segmentation, and retrieval;
  • Cross-modal remote sensing data fusion and feature extraction;
  • Remote sensing image caption generation;
  • Remote sensing visual question answering;
  • Remote sensing cross-modal image generation;
  • Remote sensing image-text cross-modal conversion;
  • Geographic knowledge map construction;
  • Optical–SAR image interpretation;
  • Cross-modal remote sensing object detection;
  • HSI–LIDAR cross-modal remote sensing image fusion classification;
  • Application of remote sensing cross-modal research: ecosystem monitoring, underground resource exploration, urban planning and geological hazard warning.

Dr. Xiangtao Zheng
Dr. Xiumei Chen
Prof. Dr. Jinchang Ren
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Remote Sensing is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2700 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • cross-modal representation learning
  • cross-modal detection and identification
  • cross-modal conversion and retrieval
  • cross-modal knowledge reasoning
  • cross-modal image generation
  • cross-modal collaborative learning
  • remote sensing cross-modal application
  • domain adaptation
  • transfer learning
  • cross-modal consistency feature representation

Published Papers (2 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

18 pages, 5494 KiB  
Article
Hierarchical Semantic-Guided Contextual Structure-Aware Network for Spectral Satellite Image Dehazing
by Lei Yang, Jianzhong Cao, Hua Wang, Sen Dong and Hailong Ning
Remote Sens. 2024, 16(9), 1525; https://doi.org/10.3390/rs16091525 - 25 Apr 2024
Viewed by 152
Abstract
Haze or cloud always shrouds satellite images, obscuring valuable geographic information for military surveillance, natural calamity surveillance and mineral resource exploration. Satellite image dehazing (SID) provides the possibility for better applications of satellite images. Most of the existing dehazing methods are tailored for [...] Read more.
Haze or cloud always shrouds satellite images, obscuring valuable geographic information for military surveillance, natural calamity surveillance and mineral resource exploration. Satellite image dehazing (SID) provides the possibility for better applications of satellite images. Most of the existing dehazing methods are tailored for natural images and are not very effective for satellite images with non-homogeneous haze since the semantic structure information and inconsistent attenuation are not fully considered. To tackle this problem, this study proposes a hierarchical semantic-guided contextual structure-aware network (SCSNet) for spectral satellite image dehazing. Specifically, a hybrid CNN–Transformer architecture integrated with a hierarchical semantic guidance (HSG) module is presented to learn semantic structure information by synergetically complementing local representation from non-local features. Furthermore, a cross-layer fusion (CLF) module is specially designed to replace the traditional skip connection during the feature decoding stage so as to reinforce the attention to the spatial regions and feature channels with more serious attenuation. The results on the SateHaze1k, RS-Haze, and RSID datasets demonstrated that the proposed SCSNet can achieve effective dehazing and outperforms existing state-of-the-art methods. Full article
(This article belongs to the Special Issue Remote Sensing Cross-Modal Research: Algorithms and Practices)
24 pages, 12612 KiB  
Article
Multi-Dimensional Fusion of Spectral and Polarimetric Images Followed by Pseudo-Color Algorithm Integration and Mapping in HSI Space
by Fengqi Guo, Jingping Zhu, Liqing Huang, Feng Li, Ning Zhang, Jinxin Deng, Haoxiang Li, Xiangzhe Zhang, Yuanchen Zhao, Huilin Jiang and Xun Hou
Remote Sens. 2024, 16(7), 1119; https://doi.org/10.3390/rs16071119 - 22 Mar 2024
Viewed by 515
Abstract
Spectral–polarization imaging technology plays a crucial role in remote sensing detection, enhancing target identification and tracking capabilities by capturing both spectral and polarization information reflected from object surfaces. However, the acquisition of multi-dimensional data often leads to extensive datasets that necessitate comprehensive analysis, [...] Read more.
Spectral–polarization imaging technology plays a crucial role in remote sensing detection, enhancing target identification and tracking capabilities by capturing both spectral and polarization information reflected from object surfaces. However, the acquisition of multi-dimensional data often leads to extensive datasets that necessitate comprehensive analysis, thereby impeding the convenience and efficiency of remote sensing detection. To address this challenge, we propose a fusion algorithm based on spectral–polarization characteristics, incorporating principal component analysis (PCA) and energy weighting. This algorithm effectively consolidates multi-dimensional features within the scene into a single image, enhancing object details and enriching edge features. The robustness and universality of our proposed algorithm are demonstrated through experimentally obtained datasets and verified with publicly available datasets. Additionally, to meet the requirements of remote sensing tracking, we meticulously designed a pseudo-color mapping scheme consistent with human vision. This scheme maps polarization degree to color saturation, polarization angle to hue, and the fused image to intensity, resulting in a visual display aligned with human visual perception. We also discuss the application of this technique in processing data generated by the Channel-modulated static birefringent Fourier transform imaging spectropolarimeter (CSBFTIS). Experimental results demonstrate a significant enhancement in the information entropy and average gradient of the fused image compared to the optimal image before fusion, achieving maximum increases of 88% and 94%, respectively. This provides a solid foundation for target recognition and tracking in airborne remote sensing detection. Full article
(This article belongs to the Special Issue Remote Sensing Cross-Modal Research: Algorithms and Practices)
Show Figures

Figure 1

Back to TopTop