Logo image
TriGEFNet: A Tri-Stream Multimodal Enhanced Fusion Network for Landslide Segmentation from Remote Sensing Imagery
Journal article   Open access   Peer reviewed

TriGEFNet: A Tri-Stream Multimodal Enhanced Fusion Network for Landslide Segmentation from Remote Sensing Imagery

Zirui Zhang, Qingfeng Hu, Haoran Fang, Wenkai Liu, Ruimin Feng, Shoukai Chen, Qifan Wu, Peng Wang and Weiqiang Lu
Remote sensing (Basel, Switzerland), Vol.18(2), pp.1-26
01/06/2026

Abstract

Landslides are among the most prevalent geological hazards worldwide, posing severe threats to public safety due to their sudden onset and destructive potential. The rapid and accurate automated segmentation of landslide areas is a critical task for enhancing capabilities in disaster risk assessment, emergency response, and post-disaster management. However, existing deep learning models for landslide segmentation predominantly rely on unimodal remote sensing imagery. In complex Karst landscapes characterized by dense vegetation and severe shadow interference, the optical features of landslides are difficult to extract effectively, thereby significantly limiting recognition accuracy. Therefore, synergistically utilizing multimodal data while mitigating information redundancy and noise interference has emerged as a core challenge in this field. To address this challenge, this paper proposes a Triple-Stream Guided Enhancement and Fusion Network (TriGEFNet), designed to efficiently fuse three data sources: RGB imagery, Vegetation Indices (VI), and Slope. The model incorporates an adaptive guidance mechanism within the encoder. This mechanism leverages the terrain constraints provided by slope to compensate for the information loss within optical imagery under shadowing conditions. Simultaneously, it integrates the sensitivity of VIs to surface destruction to collectively calibrate and enhance RGB features, thereby extracting fused features that are highly responsive to landslides. Subsequently, gated skip connections in the decoder refine these features, ensuring the optimal combination of deep semantic information with critical boundary details, thus achieving deep synergy among multimodal features. A systematic performance evaluation of the proposed model was conducted on the self-constructed Zunyi dataset and two publicly available datasets. Experimental results demonstrate that TriGEFNet achieved mean Intersection over Union (mIoU) scores of 86.27% on the Zunyi dataset, 80.26% on the L4S dataset, and 89.53% on the Bijie dataset, respectively. Compared to the multimodal baseline model, TriGEFNet achieved significant improvements, with maximum gains of 7.68% in Recall and 4.37% in F1-score across the three datasets. This study not only presents a novel and effective paradigm for multimodal remote sensing data fusion but also provides a forward-looking solution for constructing more robust and precise intelligent systems for landslide monitoring and assessment.
url
https://doi.org/10.3390/rs18020186View
Published (Version of record) Open

Metrics

1 Record Views

Details

Logo image