Welcome to Smart Agriculture 中文

Smart Agriculture ›› 2021, Vol. 3 ›› Issue (1): 109-117.doi: 10.12133/j.smartag.2021.3.1.202009-SA004

• Information Processing and Decision Making • Previous Articles     Next Articles

Distilled-MobileNet Model of Convolutional Neural Network Simplified Structure for Plant Disease Recognition

QIU Wenjie1(), YE Jin1(), HU Liangqing1, YANG Juan2, LI Qili3, MO Jianyou3, YI Wanmao1   

  1. 1.School of Computer and Electronic Information, Guangxi University, Nanning 530004, China
    2.School of Agriculture, Guangxi University, Nanning 530004, China
    3.Institute of Plant Protection, Guangxi Academy of Agricultural Sciences, Nanning 530007, China
  • Received:2020-09-29 Revised:2020-12-03 Online:2021-03-30 Published:2021-06-01
  • corresponding author: Jin YE E-mail:qiuwenjie1997@163.com;yejin@gxu.edu.cn


The development of convolutional neural networks(CNN) has brought a large number of network parameters and huge model volumes, which greatly limites the application on devices with small computing resources, such as single-chip microcomputers and mobile devices. In order to solve the problem, a structured model compression method was studied in this research. Its core idea was using knowledge distillation to transfer the knowledge from the complex integrated model to a lightweight small-scale neural network. Firstly, VGG16 was used to train a teacher model with a higher recognition rate, whose volume was much larger than the student model. Then the knowledge in the model was transfered to MobileNet by using distillation. The parameters number of the VGG16 model was greatly reduced. The knowledge-distilled model was named Distilled-MobileNet, and was applied to the classification task of 38 common diseases (powdery mildew, Huanglong disease, etc.) of 14 crops (soybean, cucumber, tomato, etc.). The performance test of knowledge distillation on four different network structures of VGG16, AlexNet, GoogleNet, and ResNet showed that when VGG16 was used as a teacher model, the accuracy of the model was improved to 97.54%. Using single disease recognition rate, average accuracy rate, model memory and average recognition time as 4 indicators to evaluate the accuracy of the trained Distilled-MobileNet model in a real environment, the results showed that, the average accuracy of the model reached 97.62%, and the average recognition time was shortened to 0.218 s, only accounts for 13.20% of the VGG16 model, and the model size was reduced to only 19.83 MB, which was 93.60% smaller than VGG16. Compared with traditional neural networks, distilled-mobile model has a significant improvement in reducing size and shorting recognition time, and can provide a new idea for disease recognition on devices with limited memory and computing resources.

Key words: disease identification, deep learning, model compression, knowledge distillation, convolutional neural network

CLC Number: