欢迎您访问《智慧农业(中英文)》官方网站! English

Smart Agriculture ›› 2023, Vol. 5 ›› Issue (3): 49-61.doi: 10.12133/j.smartag.SA202308013

• 专刊--作物信息监测技术 • 上一篇    下一篇

基于无人机遥感和深度学习的葡萄卷叶病感染程度诊断方法

刘易雪1,2,3(), 宋育阳4, 崔萍5, 房玉林4, 苏宝峰1,2,3()   

  1. 1. 西北农林科技大学 机械与电子工程学院,陕西杨凌 712100,中国
    2. 农业农村部农业物联网重点实验室,陕西杨凌 712100,中国
    3. 陕西省农业信息感知与智能服务重点实验室,陕西杨凌 712100,中国
    4. 西北农林科技大学 葡萄酒学院,陕西杨凌 712100,中国
    5. 宁夏贺兰山东麓葡萄产业园区管理委员会,宁夏 银川,750002,中国
  • 收稿日期:2023-08-10 出版日期:2023-09-30
  • 基金资助:
    宁夏回族自治区重点研发计划项目(2021BEF02017)
  • 作者简介:
    刘易雪,研究方向为田间植物表型方法研究。E-mail:
  • 通信作者:
    苏宝峰,博士,教授,研究方向为田间植物表型快速获取及应用。E-mail:

Diagnosis of Grapevine Leafroll Disease Severity Infection via UAV Remote Sensing and Deep Learning

LIU Yixue1,2,3(), SONG Yuyang4, CUI Ping5, FANG Yulin4, SU Baofeng1,2,3()   

  1. 1. College of Mechanical and Electronic Engineering, Northwest A&F University, Yangling 712100, China
    2. Key Laboratory of Agricultural Internet of Things, Ministry of Agriculture and Rural Affairs, Yangling 712100, China
    3. Shaanxi Key Laboratory of Agriculture Information Perception and Intelligent Service, Yangling 712100, China
    4. College of Enology, Northwest A&F University, Yangling 712100, China
    5. Ningxia Helan Mountain East Foothill Wine Industry Park Management Committee, Yinchuan 750002, China
  • Received:2023-08-10 Online:2023-09-30
  • Supported by:
    The Key R & D projects of Ningxia Hui Autonomous Region(2021BEF02017)

摘要:

[目的/意义] 葡萄卷叶病是一种严重影响葡萄产量和品质的病害。葡萄卷叶病感染程度类别之间存在严重的数据不平衡,导致无人机遥感技术难以进行精确的诊断。针对此问题,本研究提出一种结合细粒度分类和生成对抗网络(Generative Adversarial Network,GAN)的方法,用于提高无人机遥感图像中葡萄卷叶病感染程度分类的性能。 [方法] 以蛇龙珠品种卷叶病识别诊断为例,使用GANformer分别对每一类的葡萄园正射影像的分块图像进行学习,生成多样化和逼真的图像以增强数据,并以Swin Transformer tiny作为基础模型,提出改进模型CA-Swin Transformer,引入通道注意力机制(Channel Attention,CA)来增强特征表达能力,并使用ArcFace损失函数和实例归一化(Instance Normalization,IN)来改进模型的性能。 [结果和讨论] GANformer可以生成FID score为93.20的蛇龙珠虚拟冠层图像,有效地改善数据不平衡问题。同时,相比基于卷积神经网络(Convolutional Neural Networks,CNN)的深度学习模型,基于Transformer的深度学习模型在卷叶病感染程度诊断的问题上更具优势。最佳模型Swin Transformer在增强数据集上达到83.97%的准确率,比在原始数据集上提高3.86%,且高于GoogLeNet、MobileNetV2、NasNet Mobile、ResNet18、ResNet50、CVT和T2TViT等对照模型。而本研究所提的CA-Swin Transformer在增强数据后的测试集上达到86.65%的分类精度,比在原始的测试集上使用Swin Transformer精度提高6.54%。 [结论] 本研究基于CA-Swin Transformer使用滑动窗口法制作了葡萄园蛇龙珠卷叶病严重程度分布图,为葡萄园卷叶病的防治提供了参考。同时,本研究的方法为无人机遥感监测作物病害提供了一种新的思路和技术手段。

关键词: 无人机遥感, 深度学习, 生成对抗网络, Swin Transformer, 酿酒葡萄卷叶病, 数据增强, 注意力机制

Abstract:

[Objective] Wine grapes are severely affected by leafroll disease, which affects their growth, and reduces the quality of the color, taste, and flavor of wine. Timely and accurate diagnosis of leafroll disease severity is crucial for preventing and controlling the disease, improving the wine grape fruit quality and wine-making potential. Unmanned aerial vehicle (UAV) remote sensing technology provides high-resolution images of wine grape vineyards, which can capture the features of grapevine canopies with different levels of leafroll disease severity. Deep learning networks extract complex and high-level features from UAV remote sensing images and perform fine-grained classification of leafroll disease infection severity. However, the diagnosis of leafroll disease severity is challenging due to the imbalanced data distribution of different infection levels and categories in UAV remote sensing images. [Method] A novel method for diagnosing leafroll disease severity was developed at a canopy scale using UAV remote sensing technology and deep learning. The main challenge of this task was the imbalanced data distribution of different infection levels and categories in UAV remote sensing images. To address this challenge, a method that combined deep learning fine-grained classification and generative adversarial networks (GANs) was proposed. In the first stage, the GANformer, a Transformer-based GAN model was used, to generate diverse and realistic virtual canopy images of grapevines with different levels of leafroll disease severity. To further analyze the image generation effect of GANformer. The t-distributed stochastic neighbor embedding (t-SNE) to visualize the learned features of real and simulated images. In the second stage, the CA-Swin Transformer, an improved image classification model based on the Swin Transformer and channel attention mechanism was used, to classify the patch images into different classes of leafroll disease infection severity. CA-Swin Transformer could also use a self-attention mechanism to capture the long-range dependencies of image patches and enhance the feature representation of the Swin Transformer model by adding a channel attention mechanism after each Transformer layer. The channel attention (CA) mechanism consisted of two fully connected layers and an activation function, which could extract correlations between different channels and amplify the informative features. The ArcFace loss function and instance normalization layer was also used to enhance the fine-grained feature extraction and downsampling ability for grapevine canopy images. The UAV images of wine grape vineyards were collected and processed into orthomosaic images. They labeled into three categories: healthy, moderate infection, and severe infection using the in-field survey data. A sliding window method was used to extract patch images and labels from orthomosaic images for training and testing. The performance of the improved method was compared with the baseline model using different loss functions and normalization methods. The distribution of leafroll disease severity was mapped in vineyards using the trained CA-Swin Transformer model. [Results and Discussions] The experimental results showed that the GANformer could generate high-quality virtual canopy images of grapevines with an FID score of 93.20. The images generated by GANformer were visually very similar to real images and could produce images with different levels of leafroll disease severity. The T-SNE visualization showed that the features of real and simulated images were well clustered and separated in two-dimensional space, indicating that GANformer learned meaningful and diverse features, which enriched the image dataset. Compared to CNN-based deep learning models, Transformer-based deep learning models had more advantages in diagnosing leafroll disease infection. Swin Transformer achieved an optimal accuracy of 83.97% on the enhanced dataset, which was higher than other models such as GoogLeNet, MobileNetV2, NasNet Mobile, ResNet18, ResNet50, CVT, and T2TViT. It was found that replacing the cross entropy loss function with the ArcFace loss function improved the classification accuracy by 1.50%, and applying instance normalization instead of layer normalization further improved the accuracy by 0.30%. Moreover, the proposed channel attention mechanism, named CA-Swin Transformer, enhanced the feature representation of the Swin Transformer model, achieved the highest classification accuracy on the test set, reaching 86.65%, which was 6.54% higher than using the Swin Transformer on the original test dataset. By creating a distribution map of leafroll disease severity in vineyards, it was found that there was a certain correlation between leafroll disease severity and grape rows. Areas with a larger number of severe leafroll diseases caused by Cabernet Sauvignon were more prone to have missing or weak plants. [Conclusions] A novel method for diagnosing grapevine leafroll disease severity at a canopy scale using UAV remote sensing technology and deep learning was proposed. This method can generate diverse and realistic virtual canopy images of grapevines with different levels of leafroll disease severity using GANformer, and classify them into different classes using CA-Swin Transformer. This method can also map the distribution of leafroll disease severity in vineyards using a sliding window method, and provides a new approach for crop disease monitoring based on UAV remote sensing technology.

Key words: UAV remote sensing, deep learning, generate adversarial networks, Swin Transformer, leafroll disease of wine grape, data augmentation, attention mechanism