Professional Documents
Culture Documents
com
A R T I C L E I N F O A B S T R A C T
Article history: Grape diseases are main factors causing serious grapes reduction. So it is urgent to develop
Received 17 April 2019 an automatic identification method for grape leaf diseases. Deep learning techniques have
Received in revised form recently achieved impressive successes in various computer vision problems, which
11 October 2019 inspires us to apply them to grape diseases identification task. In this paper, a united con-
Accepted 15 October 2019 volutional neural networks (CNNs) architecture based on an integrated method is pro-
Available online 23 October 2019 posed. The proposed CNNs architecture, i.e., UnitedModel is designed to distinguish
leaves with common grape diseases i.e., black rot, esca and isariopsis leaf spot from
Keywords: healthy leaves. The combination of multiple CNNs enables the proposed UnitedModel to
Grape leaf diseases extract complementary discriminative features. Thus the representative ability of United-
Identification Model has been enhanced. The UnitedModel has been evaluated on the hold-out PlantVil-
Multi-network integration method lage dataset and has been compared with several state-of-the-art CNN models. The
Convolutional neural network experimental results have shown that UnitedModel achieves the best performance on var-
Deep learning ious evaluation metrics. The UnitedModel achieves an average validation accuracy of
99.17% and a test accuracy of 98.57%, which can serve as a decision support tool to help
farmers identify grape diseases.
Ó 2019 China Agricultural University. Production and hosting by Elsevier B.V. on behalf of
KeAi. This is an open access article under the CC BY-NC-ND license (http://creativecommons.
org/licenses/by-nc-nd/4.0/).
previously unidentified and, inherently, where there is no Athanikar and Badar [9] applied Neural Network to categorize
local expertise to combat them [3]. As a result, an automatic the potato leaf image as either healthy or diseased. Their
identification method for grape disease identification is in results showed that BPNN could effectively detect the disease
urgent demand. spots and could classify the particular disease type with an
The development of sophisticated instruments and fast accuracy of 92%. The deployment of deep CNNs has especially
computational techniques have paved the way for real-time led to a breakthrough in plant disease classification, which
scanning and automatic detection of anomalies in a crop can find very high variance of pathological symptoms in
[4]. Although traditional machine learning methods have visual appearance, even high intra-class dissimilarity and
gained some valuable experience in identification and diag- low inter-class similarity that may be only noticed by the
nosis of crop diseases, they are limited to following the pipe- botanists [25]. Lee et al. [26] proposed a CNN approach to iden-
lined procedures of image segmentation (such as clustering tify leaf images and reported an average accuracy of 99. 7% on
method [5], threshold method [6], etc.), feature extraction a dataset covering 44 species, but the scale of datasets was
(such as shape, texture, color features, etc. [7]), and pattern very small. Zhang et al. [27] used GoogLeNet to address the
recognition (such as k-nearest neighbor method (KNN) [8], detection of cherry leaf powdery mildew disease and obtained
support vector machine (SVM) [1], back propagation neural an accuracy of 99.6%. Their results also demonstrated trans-
network (BPNN) [9], etc.). It is difficult to select and extract fer learning can boost the performances of deep learning
the optimal visible pathological features and thus highly model in crop disease identification. Mohanty et al. [15]
skilled engineers and experienced experts are demanded, fine-tuned deep learning models pre-trained on ImageNet to
which is not only to a considerable extent subjective but also identify 14 crop species and 26 diseases. The models were
leads to a great waste of manpower and financial resources. evaluated on a publicly available dataset including 54,306
In contrast, deep learning techniques can automatically learn images of diseased and healthy plant leaves collected under
the hierarchical feature of pathologies and do not need to controlled conditions. They achieved the best accuracy of
manually design the feature extraction and classifier. Deep 99.35% on a hold-out test dataset.
learning method has so excellent generalization ability and Although the basic CNN frameworks, such as AlexNet [28],
robustness that it excels in many areas: signal processing VGGNet [29], GoogLeNet [30], DenseNet [31] and ResNet [32]
[10], pedestrian detection [11], face recognition [12], road have been demonstrated effective and widely used in crop
crack detection [13], biomedical image analysis [14], etc. In diseases classification, most previous works had troubling
addition, deep learning techniques have also achieved boosting up the classification accuracy rate to some extent.
impressive results in the field of agriculture and benefit more As a matter of fact, a single model can’t meet the further
smallholders and horticultural workers, including diagnosis requests in terms of precision. In large machine learning
of crop diseases [15], recognition of weeds [16], selection of competitions, the best results were usually achieved by the
fine seeds [17], pest identification [18], fruit counting [19], integration of multiple models rather than by a single model.
research on land cover [20] etc, which has contributed to deal- For instance, the well-known Inception-ResNet-v2 [33] was
ing with image classification of areas of interest. Further- born out of the fusion of two excellent deep CNNs, as the
more, some applications focused on predicting future name suggests. Inspired by network in network concept
parameters, such as crop yield [21], weather conditions [22] [34], we think that integration is the most straightforward
and soil moisture content in the field [23]. Inspired by the and effective way when the basic models are significantly dif-
great success of CNN-based methods in image classification, ferent. We propose the UnitedModel in this study. The United-
we propose an integrated model, denoted UnitedModel, for Model is a integration of the powerful and popular deep
automatic grape leaf disease identification. learning architectures of GoogLeNet and ResNet. We also take
The rest of this paper is organized as follows. Section 2 advantage of transfer learning to boost accuracy as well as
starts with an overview of related works. Section 3 introduces reduce the training time.
dataset as well as data preprocessing and covers details of the
proposed UnitedModel. In Section 4, we conduct all the exper- 3. Materials and methods
iments, discuss the limitation of the proposed UnitedModel
and prospect the future work. In Section 5, we conclude this 3.1. Dataset and preprocessing
paper.
Our dataset comes from an open access repository named
2. Related works PlantVillage which focuses on plant health [35]. The dataset
used for evaluating the proposed method is composed of
Machine learning techniques were applied widely in plant healthy (171 images) and symptom images including black
disease classification in the early stage. Li et al. [24] proposed rot (pathogen: Guignardia bidwellii, 476 images), esca (patho-
a method based on K-means clustering segmentation of the gen: Phaeomoniella spp, 552 images) and isariopsis leaf spot
grape disease images and a SVM classifier was designed (pathogen: Pseudocercospor a-vitis, 420 images) (see Table 1).
based on thirty-one effective selected features to identify The identification of grape diseases is based on its leaf, not
grape downy mildew disease and grape powdery mildew dis- flower, fruit or stem. For one thing, the flower and fruit of
ease with testing recognition rates of 90% and 93.33%, respec- grape only appear in a limited time while the leaf presents
tively. Significant progress has been made in the use of image most of the year. For another thing, the stem of grape can
processing approaches to detect various diseases in crops. hardly present the symptoms of the diseases timely, while
420 Information Processing in Agriculture 7 (2 0 2 0) 4 1 8–42 6
1 Black rot 476 288 72 116 Appear small, See Fig. 1 first row
brown circular lesions
2 Esca 552 288 72 192 Appear dark red or yellow stipes See Fig. 1 second row
3 Isariopsis 420 288 72 60 Appear many small rounded, See Fig. 1 third row
leaf spot polygonal
or irregular brown spots
4 Health 171 96 24 51 – See Fig. 1 fourth row
Total samples 1619 960 240 419 – –
the leaf is generally sensitive to the state of plants, whose 3.2. Identification model of grape leaf diseases
shape, texture and color usually contains richer information.
The raw images are divided into training dataset and test In this paper, the grape leaf diseases identification can be for-
dataset. 360 Symptom and 120 healthy images are selected mulated as a multi-class classification problem. The classifi-
for training and the rest images are used for testing. To pre- cation process of this work can be described as follows
vent over-fitting, the training dataset is further split into (Fig. 2). Assuming a mapping function from grape leaf images
training (80%) and validation data (20%). Thus the training to corresponding predicted labels (i.e. ‘‘Label”, ‘‘Label2”,
dataset is 960 samples in total, validation dataset is 240 sam- ‘‘Label3” and ‘‘Label4”) f : X ! Y: Then the grape leaf training
ples in total and test dataset is 419 samples in total [36]. The dataset is denoted as ðx1 ; y1 Þ; ðx2 ; y2 Þ ;:::; ðxn ; yn Þ. The train-
original images in the PlantVillage dataset are RGB images of ing dataset is the input of the CNN model A. During training,
arbitrary size. Data preprocessing is necessary and all the the weights of A are updated continuously to minimize the
images are resized to the expected input size of the respective average loss function which is defined as:
networks, i.e. 224 224 for VGGNet, DenseNet and ResNet,
1X n
and 229 229 for GoogLeNet. Model optimization and predic- JðxÞ ¼ y logðhx ðxi ÞÞ ð1Þ
n i¼1 i
tion are both performed on these rescaled images. To avoid
over-fitting and to boost the generalizability of the CNNs, data where x is the training sample, n is the number of input data,
augmentation techniques which include rotating, flipping, yi is the true label and hxðxi Þ is the predicted label of the CNN
shearing, zooming and colour changing are randomly per- model A given current weights x. Then, the end-to-end opti-
formed in training phase as shown in Fig. 1. mization can be performed using the stochastic gradient des-
Considering that the number of healthy samples in our cent (SGD) method. The gradient of the loss function can be
work is relatively small with 171 samples in total, the propor- computed with the following equation:
tion between healthy samples and diseased samples is set to
xtþ1 ¼ lxt arJðxt Þ ð2Þ
1: 3 while the class_weight ratio of them is set to 3: 1. By this
way, the loss function can be adjusted in the training process where l is the momentum weight for the current weights xt
and healthy samples can obtain more attention. and a is the learning rate. Noted that the iterative operation
Information Processing in Agriculture 7 ( 2 0 2 0 ) 4 1 8 –4 2 6 421
Fig. 5 – Validation accuracy and validation loss of various architectures are compared and the number of epoch is varied in the
case of early stopping mechanism. VGGNet, GoogLeNet, DenseNet, ResNet and UnitedModel stop training with the epoch 11,
18, 11, 9 and 12, respectively.
4. Results and discussion score (94.40%), which once again proves the superiority of
integrated CNNs.
As illustrated in Fig. 5, UnitedModel has a tendency to consis- The confusion matrices of different models on test dataset
tently improve in accuracy with growing number of epochs, are shown in Fig. 7. The threshold is set to 0.5 and the fraction
with no signs of performance deterioration. On the other of accurately predicted images for each class is displayed in
hand, loss curve of UnitedModel converges rapidly and has detail. The performance of VGGNet is poor. VGGNet tends to
a big gap with other single models in terms of error rate. Con- distinguish diseased samples as healthy, which is extremely
sistent with Fig. 6, the performance of GoogLeNet, VGGNet, harmful in actual agriculture produce. All the models except
DenseNet and ResNet is similar and all inferior to VGGNet can distinguish label3 and label4 easily, but label1
UnitedModel. and label2 are prone to be misclassified. The large green area
As we can see in Table 3, the average precision (99.05%), of healthy leaves and the golden appearance of leaves
recall (98.88%) and F1-score (98.96%) of UnitedModel are the infected with isariopsis leaf spot disease make them easier
highest among all these models. And scores of VGGNet are to be distinguished. Meanwhile, the confusion between label1
the lowest with precision (94.16%), recall (93.32%) and F1- and label2 is due to their similar pathological features.
424 Information Processing in Agriculture 7 (2 0 2 0) 4 1 8–42 6
Table 3 – Precision, recall and F1-score for the corresponding experimental configurations (best overall performance is
displayed in boldface).
Fig. 7 – Confusion matrices for the grape diseases dataset are presented with an overall accuracy associated for individual
classes. The confusion matrices are given in terms of percentages, not absolute numbers.
However, UnitedModel improves the test accuracy of label1 The loss and accuracy curves of the UnitedModel during
(97%) and label2 (99%). The UnitedModel is a combination of training phase are presented in Fig. 8. The green and black
InceptionV3 and ResNet50 and can improve the representa- curves are the loss function values on the training dataset
tive capability by enriching features using multi-network and validation dataset, respectively. The two curves drop
integration method. In this study, we conjecture different rapidly in the first two epochs and start to converge after
CNN models to generate complementary features and we around 6 epochs. And then, the training loss continues to des-
boost the performance by combining the complementary fea- cend slightly until roughly the final epoch while validation
tures. In addition, the use of transfer learning makes this pro- loss is no longer significantly reduced. The red and blue
cess more concise and efficient. curves represent the training accuracy and validation
Information Processing in Agriculture 7 ( 2 0 2 0 ) 4 1 8 –4 2 6 425
[9] Athanikar G, Badar P. Potato leaf diseases detection and machine. In: Proc international conference on computer and
classification system. Int J Comp Sci Mob Comput 2016;5 computing technologies in agriculture. Beijing, China. p.
(2):76–88. 151–62.
[10] Xie D, Zhang L, Bai L. Deep learning in visual computing and [25] Zhu H, Liu Q, Qi Y, Huang X, Jiang F, Zhang S. Plant
signal processing. Appl Comput Intell Soft Comput identification based on very deep convolutional neural
2017;2017:1–13. networks. Multimed Tools Appl 2018;77(22):29779–97.
[11] Ouyang W, Wang X. Joint deep learning for pedestrian [26] Lee SH, Chan CS, Wilkin P, Remagnino P. Deep-plant: plant
detection. In: Proc. IEEE international conference on identification with convolutional neural networks. In: Proc
computer vision. Sydney, Australia. p. 2056–63. IEEE international conference on image processing. Quebec,
[12] Sun Y, Wang X, Tang X. Hybrid deep learning for face Canada. p. 452–6.
verification. IEEE Trans Pattern Anal. Mach Intell 2016;38 [27] Zhang K, Zhang L, Wu Q. Identification of cherry leaf disease
(10):1997–2009. infected by Podosphaera Pannosa via convolutional neural
[13] Zhang L, Yang F, Zhang Y, Zhu Y. Road crack detection using network. Int J Agric Environ Inf Syst 2019;10(2):98–110.
deep convolutional neural network. In: Proc. IEEE [28] Krizhevsky A, Sutskever I, Hinton G. ImageNet classification
international conference on image processing. Arizona, USA. with deep convolutional neural networks. Adv Neural Inf
p. 3708–12. Process Syst 2012;25:1106–14.
[14] Zhou Z, Shin J, Zhang L, Gurudu S, Liang J. Fine-Tuning [29] Simonyan K, Zisserman A. Very deep convolutional networks
convolutional neural networks for biomedical image for large-scale image recognition. 2014. link: http://arxiv.org/
analysis: actively and incrementally. In: Proc. IEEE conference abs/1409.1556.
on computer vision and pattern recognition. Honolulu, [30] Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, et al.
Hawaii. p. 7340–9. Going deeper with convolutions. Proc IEEE conference on
[15] Mohanty S, Hughes D, Salathé M. Using deep learning for computer vision and pattern recognition. Columbus, USA,
image-based plant disease detection. Front Plant Sci 2014.
2016;7:1419. [31] Iandola F, Moskewicz M, Karayev S, Girshick R, Darrell T,
[16] Lottes P, Behley J, Milioto A, Cyrill Stachniss C. Fully Keutzer, K. Densenet: implementing efficient convnet
convolutional networks with sequential information for descriptor pyramids. 2014. link: https://arxiv.org/abs/1404.
robust crop and weed detection in precision farming. IEEE 1869v1.
Robot Autom Lett 2018;3(4):2870–7. [32] He K, Zhang X, Ren S, Sun J. Deep residual learning for image
[17] Wang X, Cai C. Weed seeds classification based on PCANet recognition. In: Proc IEEE conference on computer vision and
deep learning baseline. In: Proc IEEE Asia-Pacific signal and pattern recognition. Boston, Massachusetts. p. 770–8.
information processing association annual summit and [33] Szegedy C, Ioffe S, Vanhoucke V, Alemi A. Inception-v4,
conference. Hong Kong, China. p. 408–15. inception-resnet and the impact of residual connections on
[18] Cheng X, Zhang Y, Chen Y, Wu Y, Yue Y. Pest identification via learning. link: http://arxiv.org/abs/1602.07261.
deep residual learning in complex background. Comput [34] Lin M, Chen Q, Yan S. Network in network. link: http://arxiv.
Electron Agric 2017;141:351–6. org/abs/1312.4400.
[19] Rahnemoonfar M, Sheppard C. Deep count: fruit counting [35] Hughes DP, Salathe M. An open access repository of images
based on deep simulated learning. Sensors 2017;17(4):905. on plant health to enable the development of mobile disease
[20] Kussul N, Lavreniuk M, Skakun S, Shelestov A. Deep learning diagnostics through machine learning and crowdsourcing.
classification of land cover and crop types using remote 2015. link: http://arxiv.org/abs/1511.08060.
sensing data. IEEE Geosci Remote S 2017;14(5):778–82. [36] Too EC, Yujian L, Njuki S, Yingchun L. A comparative study of
[21] Kuwata K, Shibasaki R. Estimating crop yields with deep fine-tuning deep learning models for plant disease
learning and remotely sensed data. In: Proc. IEEE identification. Comput Electron Agric 2018;161:272–9.
international geoscience and remote sensing symposium. [37] Lin Z, Mu S, Huang F, Khattak AM, Wang M, Gao W, et al. A
Milan, Italy. p. 858–61. united matrix-based convolutional neural network for fine-
[22] Zhao B, Li X, Lu X, Wang Z. A CNN-RNN architecture for grained image classification of wheat leaf diseases. IEEE
multi-label weather recognition. Neurocomputing Access 2019;7:11570–90.
2018;322:47–57. [38] Anthimopoulos M, Christodoulidis S, Ebner L, Christe A,
[23] Song X, Zhang G, Liu F, Li D, Zhao Y, Yang J. Modeling spatio- Mougiakakou S. Lung pattern classification for interstitial
temporal distribution of soil moisture by deep learning-based lung diseases using a deep convolutional neural network.
cellular automata model. J Arid Land 2018;8(5):734–48. IEEE T Med Imaging 2016;35(5):1207–16.
[24] Li G, Ma Z, Wang H. Image recognition of grape downy
mildew and grape powdery mildew based on support vector