导航

Acta Aeronautica et Astronautica Sinica ›› 2025, Vol. 46 ›› Issue (23): 632763.doi: 10.7527/S1000-6893.2025.32763

• special column • Previous Articles    

RS-AdaDiff: One-step remote sensing image super-resolution diffusion model with degradation-aware adaptive estimation

Fei WANG1,2,3, Yong LIU1,2,3, Jiawei YAO1,2,3, Xuanlei ZHU1,2,3, Xiaoqiang LU4, Wenxing GUO1,2,3, Xuetao ZHANG1,2,3, Yu GUO1,2,3()   

  1. 1.National Key Laboratory of Human-Machine Hybrid Augmented Intelligence,Xi’an Jiaotong University,Xi’an 710049,China
    2.National Engineering Research Center of Visual Information and Applications,Xi’an Jiaotong University,Xi’an 710049,China
    3.Institute of Artificial Intelligence and Robotics,Xi’an Jiaotong University,Xi’an 710049,China
    4.College of Physics and Information Engineering,Fuzhou University,Fuzhou 350108,China
  • Received:2025-09-06 Revised:2025-09-24 Accepted:2025-10-20 Online:2025-11-20 Published:2025-11-13
  • Contact: Yu GUO E-mail:yu.guo@xjtu.edu.cn
  • Supported by:
    National Major Science and Technology Projects of China(2009XJTU0016)

Abstract:

Diffusion models have demonstrated great potential in generating realistic image details. However, existing diffusion models are primarily trained on natural images, making their application to remote sensing image super-resolution highly challenging. Moreover, these models typically require dozens or even hundreds of iterative sampling steps during inference, resulting in high computational costs and limited practicality. To address these issues, this paper proposes a degradation-aware adaptive estimation-based single-step remote sensing image super-resolution diffusion model (RS-AdaDiff), which balances reconstruction performance and inference efficiency. Specifically, we propose a degradation-aware timestep estimation module that adaptively estimates the diffusion timestep for the diffusion model by assessing the degradation level of the input image. This approach reconstructs the iterative denoising process into a single-step reconstruction from low-resolution to high-resolution images, thereby significantly accelerating inference. Meanwhile, we integrate trainable lightweight LoRA layers into a pre-trained diffusion model and fine-tune it on a remote sensing image dataset to mitigate the domain gap caused by data distribution differences. Additionally, to fully leverage the image priors of the pre-trained model, we introduce distribution contrastive matching distillation. By regularizing the KL divergence, the reconstructed super-resolved images are brought closer to high-resolution images and farther from low-resolution images in the feature space, thereby improving generation quality. Finally, we propose a feature-edge joint perceptual similarity loss to enhance the perception of structural information and mitigate issues such as edge blur and texture distortion. Extensive experimental results demonstrate that the proposed RS-AdaDiff outperforms existing state-of-the-art methods on multiple public remote sensing datasets, achieving significant improvements in both quantitative metrics and visual quality, and producing super-resolved remote sensing images with clearer structures and richer details.

Key words: remote sensing image super-resolution, diffusion model, adaptive estimation, computer vision, aerospace

CLC Number: