DOI: 10.3390/rs17162795 ISSN: 2072-4292

Landslide Detection with MSTA-YOLO in Remote Sensing Images

Bingkun Wang, Jiali Su, Jiangbo Xi, Yuyang Chen, Hanyu Cheng, Honglue Li, Cheng Chen, Haixing Shang, Yun Yang

Deep learning-based landslide detection in optical remote sensing images has been extensively studied. However, several challenges remain. Over time, factors such as vegetation cover and surface weathering can weaken the distinct characteristics of landslides, leading to blurred boundaries and diminished texture features. Furthermore, obtaining landslide samples is challenging in regions with low landslide frequency. Expanding the acquisition range introduces greater variability in the optical characteristics of the samples. As a result, deep learning models often struggle to achieve accurate landslide identification in these regions. To address these challenges, we propose a multi-scale target attention YOLO model (MSTA-YOLO). First, we introduced a receptive field attention (RFA) module, which initially applies channel attention to emphasize the primary features and then simulates the human visual receptive field using convolutions of varying sizes. This design enhances the model’s feature extraction capability, particularly for complex and multi-scale features. Next, we incorporated the normalized Wasserstein distance (NWD) to refine the loss function, thereby enhancing the model’s learning capacity for detecting small-scale landslides. Finally, we streamlined the model by removing redundant structures, achieving a more efficient architecture compared to state-of-the-art YOLO models. Experimental results demonstrated that our proposed MSTA-YOLO outperformed other compared methods in landslide detection and is particularly suitable for wide-area landslide monitoring.

More from our Archive