欢迎您访问《智慧农业(中英文)》官方网站! English

Smart Agriculture ›› 2025, Vol. 7 ›› Issue (2): 172-182.doi: 10.12133/j.smartag.SA202410011

• 信息处理与决策 • 上一篇    下一篇

基于混合注意力生成对抗网络的遥感图像去雾方法

马六1, 毛克彪2(), 郭中华1   

  1. 1. 宁夏大学 电子与电气工程学院 宁夏 银川 750021,中国
    2. 中国农业科学院农业资源与农业区划研究所,北京 100081,中国
  • 收稿日期:2024-10-14 出版日期:2025-03-30
  • 作者简介:
    马 六,硕士研究生,研究方向为基于对抗神经网络的热红外遥感多参数反演研究。E-mail:
  • 通信作者:
    毛克彪,博士,研究员,研究方向为人工智能在地学和农学中的应用。E-mail:

Defogging Remote Sensing Images Method Based on a Hybrid Attention-Based Generative Adversarial Network

MA Liu1, MAO Kebiao2(), GUO Zhonghua1   

  1. 1. School of Electronic and Electrical Engineering, Ningxia University, Ningxia 750021, China
    2. Institute of Agricultural Resources and Agricultural Zoning, Chinese Academy of Agricultural Sciences, Beijing 100081, China
  • Received:2024-10-14 Online:2025-03-30
  • Foundation items:Key Project of Natural Science Foundation of Ningxia Science and Technology Department(2024AC02032); Central Public-Interest Scientific Institution Basal Research Fund(Y2025YC86)
  • About author:
    MA Liu, E-mail:
  • Corresponding author:
    MAO Kebiao, E-mail:

摘要:

【目的/意义】 近年来,深度学习在遥感图像去雾领域取得了显著进展,尤其是在引入注意力机制以提升特征学习方面。然而,传统的注意力机制大多依赖全局平均池化,导致模型对特定影响点的敏感性过高,难以有效应对遥感图像中的去雾问题。为了提高去雾技术的效果,满足农业、城市规划等领域对图像质量日益增长的需求,现有方法亟需改进。 【方法】 本研究提出了一种混合注意力生成对抗网络(Hybrid Attention-Based Generative Adversarial Network, HAB-GAN)。该模型通过结合高效通道注意力模块与空间注意力模块,嵌入生成对抗网络架构中,实现了对遥感图像去雾效果的显著提升。高效通道注意力模块通过降低全局特征聚合中的冗余信息,既保留了性能,又减少了模型复杂度;空间注意力模块则从局部到全局对遥感图像中的雾化区域进行识别和聚焦,增强了对这些区域的恢复能力。这种方法能够更加有效地应对遥感图像中复杂多变的景观,尤其适用于农业等需要高质量遥感数据的领域。 【结果与讨论】 在RESISC(Remote Sensing Image Scene Classification)45数据集上,与现有的其他注意力机制去雾模型,如SpA GAN和HyA-GAN进行比较,HAB-GAN模型去雾效果更优,其中峰值信噪比(Peak Signal-to-Noise Ratio, PSNR)分别增加了2.64和1.14 dB,结构相似度(Structural Similarity Index, SSIM)分别增加了0.012 2和0.001 9。此外,消融实验验证了混合注意力机制的有效性,去除HAB模块后,HAB-GAN模型的PSNR下降了3.87 dB,SSIM下降了0.033 4。 【结论】 提出的HAB-GAN模型显著提升了遥感图像的去雾效果,使生成的图像更加接近无雾图像,特别是对于复杂的农业、环境监测等场景具有重要应用价值。HAB模块在提升模型性能方面发挥了关键作用,为未来的遥感图像处理和相关领域提供了有力的技术支持。

关键词: 遥感图像, 深度学习, 生成对抗网络, 高效通道注意力模块, 空间注意力模块, 去雾

Abstract:

[Objective] Remote sensing images have become an important data source in fields such as surface observation, environmental monitoring, and natural disaster prediction. However, the acquisition of remote sensing images is often affected by weather phenomena such as fog and clouds, which reduces the image quality and poses challenges to subsequent analysis and processing tasks. In recent years, the introduction of attention mechanisms has enabled models to better capture and utilize important features in images, thereby significantly improving defogging performance. However, traditional channel attention mechanisms usually rely on global average pooling to summarize feature information. Although this method simplifies the complexity of calculations, it is not satisfactory when dealing with images with significant local changes and sensitivity to outliers. In addition, remote sensing images usually cover a wide area, and the diverse terrain makes the fog pattern more complex. Therefore, to address this issue, a hybrid attention-based generative adversarial network hybrid attention-based generative adversarial network (HAB-GAN) was proposed in this research, which integrates an efficient channel attention (ECA) module and a spatial attention block (SAB). [Method] By merging feature extraction from both channel and spatial dimensions, the model effectively enhanced its ability to identify and recover hazy areas in remote sensing images. In HAB-GAN, the ECA module captured local cross-channel interactions, addressing the shortcomings of traditional global averaged pooling in terms of insufficient sensitivity to local detail information. The ECA module used a global average pooling strategy without dimensionality reduction, automatically adapting to the characteristics of each channel without introducing extra parameters, thereby enhancing the inter-channel dependencies. ECA emploied a one-dimensional convolution operation, which used a learnable kernel size to adaptively determine the range of channel interactions. This design effectively avoided the over-smoothing of global features common in traditional pooling layers, allowing the model to more precisely extract local detailed while maintaining low computational complexity. The SAB module introduced a weighted mechanism on the spatial dimension by constructing a spatial attention map to enhance the model's ability to identify hazy areas in the image. This module extracted feature maps through convolution operations and applies attention weighting in both horizontal and vertical directions, highlighting regions with severe haze, allowing the model to better capture spatial information in the image, thereby enhancing dehazing performance. The generator of HAB-GAN combined residual network structures with hybrid attention modules. It first extracted initial features from input images through convolutional layers and then passed these features through several residual blocks. The residual blocks effectively mitigated the vanishing gradient problem in deep neural networks and maintain feature consistency and continuity by passing input features directly to deeper network layers through skip connections. Each residual block incorporated ECA and SAB modules, enabling precise feature learning through weighted processing in both channel and spatial dimensions. After extracting effective features, the generator generated dehazed images through convolution operations. The discriminator adopted a standard convolutional neural network architecture, focusing on extracting local detail features from the images generated by the generator. It consisted of multiple convolutional layers, batch normalization layers, and Leaky ReLU activation functions. By extracting local features layer by layer and down-sampling, the discriminator progressively reduced the spatial resolution of the images, evaluating their realism at both global and local levels. The generator and discriminator were jointly optimized through adversarial training, where the generator aimed to produce increasingly realistic dehazed images, and the discriminator continually improved its ability to distinguish between real and generated images, thereby enhancing the learning effectiveness and image quality of the generator. [Results and Discussions] To validate the effectiveness of HAB-GAN, experiments were conducted on the remote sensing image scene classification 45 (RESISC45) dataset. The experimental results demonstrated that compared to existing dehazing models, HAB-GAN excels in key evaluation metrics such as peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM). Specifically, compared to SpA GAN, HAB-GAN improved PSNR by 2.642 5 dB and SSIM by 0.012 2; Compared to HyA-GAN, PSNR improved by 1.138 dB and SSIM by 0.001 9. Additionally, to assess the generalization capability of HAB-GAN, further experiments were conducted on the RICE2 dataset to verify its performance in cloud removal tasks. The results showed that HAB-GAN also performs exceptionally well in cloud removal tasks, with PSNR improving by 3.593 2 dB and SSIM improving by 0.040 2. Compared to HyA-GAN, PSNR and SSIM increased by 1.854 dB and 0.012 4, respectively. To further explored the impact of different modules on the model's performance, ablation experiments were designed, gradually removing the ECA module, the SAB module, and the entire hybrid attention module. The experimental results showed that removing the ECA module reduced PSNR by 2.642 5 dB and SSIM by 0.012 2; Removing the SAB module reduced PSNR by 2.955 dB and SSIM by 0.008 7, and removing the entire hybrid attention module reduced PSNR and SSIM by 3.866 1 dB and 0.033 4, respectively. [Conclusions] The proposed HAB-GAN model not only performs excellently in dehazing and beclouding tasks but also significantly enhances the clarity and detail recovery of dehazed images through the synergistic effect of the ECA module and the SAB module. Additionally, its strong performance across different remote sensing datasets further validates its effectiveness and generalization ability, showcasing broad application potential particularly in fields such as agriculture, environmental monitoring, and disaster prediction, where high-quality remote sensing data is crucial. HAB-GAN is poised to become a valuable tool for improving data reliability and supporting more accurate decision-making and analysis.

Key words: remote sensing image, deep learning, generative adversarial network, efficient channel attention module, spatial attention module, defogging

中图分类号: