Distilled-MobileNet Model of Convolutional Neural Network Simplified Structure for Plant Disease Recognition

The development of convolutional neural networks(CNN) has brought a large number of network parameters and huge model volumes, which greatly limites the application on devices with small computing resources, such as single-chip microcomputers and mobile devices. In order to solve the problem, a stru...

Descripción completa

Guardado en:
Detalles Bibliográficos
Autores principales: QIU Wenjie, YE Jin, HU Liangqing, YANG Juan, LI Qili, MO Jianyou, YI Wanmao
Formato: article
Lenguaje:EN
ZH
Publicado: Editorial Office of Smart Agriculture 2021
Materias:
Acceso en línea:https://doaj.org/article/f84d60be63634dc9b10de6e32d98663f
Etiquetas: Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!
id oai:doaj.org-article:f84d60be63634dc9b10de6e32d98663f
record_format dspace
spelling oai:doaj.org-article:f84d60be63634dc9b10de6e32d98663f2021-11-17T07:52:00ZDistilled-MobileNet Model of Convolutional Neural Network Simplified Structure for Plant Disease Recognition2096-809410.12133/j.smartag.2021.3.1.202009-SA004https://doaj.org/article/f84d60be63634dc9b10de6e32d98663f2021-03-01T00:00:00Zhttp://www.smartag.net.cn/article/2021/2096-8094/2096-8094-2021-3-1-109.shtmlhttps://doaj.org/toc/2096-8094The development of convolutional neural networks(CNN) has brought a large number of network parameters and huge model volumes, which greatly limites the application on devices with small computing resources, such as single-chip microcomputers and mobile devices. In order to solve the problem, a structured model compression method was studied in this research. Its core idea was using knowledge distillation to transfer the knowledge from the complex integrated model to a lightweight small-scale neural network. Firstly, VGG16 was used to train a teacher model with a higher recognition rate, whose volume was much larger than the student model. Then the knowledge in the model was transfered to MobileNet by using distillation. The parameters number of the VGG16 model was greatly reduced. The knowledge-distilled model was named Distilled-MobileNet, and was applied to the classification task of 38 common diseases (powdery mildew, Huanglong disease, etc.) of 14 crops (soybean, cucumber, tomato, etc.). The performance test of knowledge distillation on four different network structures of VGG16, AlexNet, GoogleNet, and ResNet showed that when VGG16 was used as a teacher model, the accuracy of the model was improved to 97.54%. Using single disease recognition rate, average accuracy rate, model memory and average recognition time as 4 indicators to evaluate the accuracy of the trained Distilled-MobileNet model in a real environment, the results showed that, the average accuracy of the model reached 97.62%, and the average recognition time was shortened to 0.218 s, only accounts for 13.20% of the VGG16 model, and the model size was reduced to only 19.83 MB, which was 93.60% smaller than VGG16. Compared with traditional neural networks, distilled-mobile model has a significant improvement in reducing size and shorting recognition time, and can provide a new idea for disease recognition on devices with limited memory and computing resources.QIU WenjieYE JinHU LiangqingYANG JuanLI QiliMO JianyouYI WanmaoEditorial Office of Smart Agriculturearticledisease identificationdeep learningmodel compressionknowledge distillationconvolutional neural networkAgriculture (General)S1-972Technology (General)T1-995ENZH智慧农业, Vol 3, Iss 1, Pp 109-117 (2021)
institution DOAJ
collection DOAJ
language EN
ZH
topic disease identification
deep learning
model compression
knowledge distillation
convolutional neural network
Agriculture (General)
S1-972
Technology (General)
T1-995
spellingShingle disease identification
deep learning
model compression
knowledge distillation
convolutional neural network
Agriculture (General)
S1-972
Technology (General)
T1-995
QIU Wenjie
YE Jin
HU Liangqing
YANG Juan
LI Qili
MO Jianyou
YI Wanmao
Distilled-MobileNet Model of Convolutional Neural Network Simplified Structure for Plant Disease Recognition
description The development of convolutional neural networks(CNN) has brought a large number of network parameters and huge model volumes, which greatly limites the application on devices with small computing resources, such as single-chip microcomputers and mobile devices. In order to solve the problem, a structured model compression method was studied in this research. Its core idea was using knowledge distillation to transfer the knowledge from the complex integrated model to a lightweight small-scale neural network. Firstly, VGG16 was used to train a teacher model with a higher recognition rate, whose volume was much larger than the student model. Then the knowledge in the model was transfered to MobileNet by using distillation. The parameters number of the VGG16 model was greatly reduced. The knowledge-distilled model was named Distilled-MobileNet, and was applied to the classification task of 38 common diseases (powdery mildew, Huanglong disease, etc.) of 14 crops (soybean, cucumber, tomato, etc.). The performance test of knowledge distillation on four different network structures of VGG16, AlexNet, GoogleNet, and ResNet showed that when VGG16 was used as a teacher model, the accuracy of the model was improved to 97.54%. Using single disease recognition rate, average accuracy rate, model memory and average recognition time as 4 indicators to evaluate the accuracy of the trained Distilled-MobileNet model in a real environment, the results showed that, the average accuracy of the model reached 97.62%, and the average recognition time was shortened to 0.218 s, only accounts for 13.20% of the VGG16 model, and the model size was reduced to only 19.83 MB, which was 93.60% smaller than VGG16. Compared with traditional neural networks, distilled-mobile model has a significant improvement in reducing size and shorting recognition time, and can provide a new idea for disease recognition on devices with limited memory and computing resources.
format article
author QIU Wenjie
YE Jin
HU Liangqing
YANG Juan
LI Qili
MO Jianyou
YI Wanmao
author_facet QIU Wenjie
YE Jin
HU Liangqing
YANG Juan
LI Qili
MO Jianyou
YI Wanmao
author_sort QIU Wenjie
title Distilled-MobileNet Model of Convolutional Neural Network Simplified Structure for Plant Disease Recognition
title_short Distilled-MobileNet Model of Convolutional Neural Network Simplified Structure for Plant Disease Recognition
title_full Distilled-MobileNet Model of Convolutional Neural Network Simplified Structure for Plant Disease Recognition
title_fullStr Distilled-MobileNet Model of Convolutional Neural Network Simplified Structure for Plant Disease Recognition
title_full_unstemmed Distilled-MobileNet Model of Convolutional Neural Network Simplified Structure for Plant Disease Recognition
title_sort distilled-mobilenet model of convolutional neural network simplified structure for plant disease recognition
publisher Editorial Office of Smart Agriculture
publishDate 2021
url https://doaj.org/article/f84d60be63634dc9b10de6e32d98663f
work_keys_str_mv AT qiuwenjie distilledmobilenetmodelofconvolutionalneuralnetworksimplifiedstructureforplantdiseaserecognition
AT yejin distilledmobilenetmodelofconvolutionalneuralnetworksimplifiedstructureforplantdiseaserecognition
AT huliangqing distilledmobilenetmodelofconvolutionalneuralnetworksimplifiedstructureforplantdiseaserecognition
AT yangjuan distilledmobilenetmodelofconvolutionalneuralnetworksimplifiedstructureforplantdiseaserecognition
AT liqili distilledmobilenetmodelofconvolutionalneuralnetworksimplifiedstructureforplantdiseaserecognition
AT mojianyou distilledmobilenetmodelofconvolutionalneuralnetworksimplifiedstructureforplantdiseaserecognition
AT yiwanmao distilledmobilenetmodelofconvolutionalneuralnetworksimplifiedstructureforplantdiseaserecognition
_version_ 1718425903216721920