researcharticle can deep learning identify tomato leaf

11
Research Article Can Deep Learning Identify Tomato Leaf Disease? Keke Zhang , 1 Qiufeng Wu , 2 Anwang Liu , 1 and Xiangyan Meng 2 1 College of Engineering, Northeast Agricultural University, Harbin 150030, China 2 College of Science, Northeast Agricultural University, Harbin 150030, China Correspondence should be addressed to Qiufeng Wu; [email protected] Received 9 June 2018; Accepted 30 August 2018; Published 26 September 2018 Academic Editor: Alexander Loui Copyright © 2018 Keke Zhang et al. is is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. is paper applies deep convolutional neural network (CNN) to identify tomato leaf disease by transfer learning. AlexNet, GoogLeNet, and ResNet were used as backbone of the CNN. e best combined model was utilized to change the structure, aiming at exploring the performance of full training and fine-tuning of CNN. e highest accuracy of 97.28% for identifying tomato leaf disease is achieved by the optimal model ResNet with stochastic gradient descent (SGD), the number of batch size of 16, the number of iterations of 4992, and the training layers from the 37 layer to the fully connected layer (denote as “fc”). e experimental results show that the proposed technique is effective in identifying tomato leaf disease and could be generalized to identify other plant diseases. 1. Introduction Tomato is a widely cultivated crop throughout the world, which contains rich nutrition, unique taste, and health effects, so it plays an important role in the agricultural production and trade around the world. Given the importance of tomato in the economic context, it is necessary to maximize produc- tivity and product quality by using techniques. Corynespora leaf spot disease, early blight, late blight, leaf mold disease, septoria leaf spot, two-spotted spider mite, virus disease, and yellow leaf curl disease are 8 common diseases in tomato [1– 8]; thus, a real time and precise recognition technology is essential. Recently, since CNN has the self-learned mechanism, that is, extracting features and classifying images in the one procedure [9], CNN has been successfully applied in various applications, such as writer identification [10], salient object detection [11, 12], scene text detection [13, 14], truncated inference learning [15], road crack detection [16, 17], biomed- ical image analysis [18], predicting face attributes from web images [19], and pedestrian detection [20], and achieved the better performance. In addition, CNN is able to extract more robust and discriminative features with considering the global context information of regions [10], and CNN is scarcely affected by the shadow, distortion, and brightness of the natural images. With the rapid development of CNN, many powerful architectures of CNN emerged, such as AlexNet [21], GoogLeNet [22], VGGNet [23], Inception-V3 [24], Inception-V4 [25], ResNet [26], and DenseNets [27]. Training deep neural networks from scratch needs amounts of data and expensive computational resources. Meanwhile, we sometimes have a classification task in one domain, but we only have enough data in other domains. Fortunately, transfer learning can improve the performance of deep neural networks by avoiding complex data mining and data-labeling efforts [28]. In practice, transfer learning consists of two ways [29]. One option is to fine-tune the networks weights by using our data as input; it is worth nothing that the new data must be resized to the input size of the pretrained network. Another way is to obtain the learned weights from the pretrained network and apply the weights to the target network. In this work, first, we compared the performance between SGD [30] and Adaptive Moment Estimation (Adam) [30, 31] in identifying tomato leaf disease. ese optimization methods are based on the pretrained networks AlexNet [21], GoogLeNet [22], and ResNet [26]. en, the network architecture with the highest performance was selected and experiments on effect of two hyperparameters (i.e., batch size and number of iterations) on accuracy were carried out. Next, Hindawi Advances in Multimedia Volume 2018, Article ID 6710865, 10 pages https://doi.org/10.1155/2018/6710865

Upload: others

Post on 28-Jul-2022

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: ResearchArticle Can Deep Learning Identify Tomato Leaf

Research ArticleCan Deep Learning Identify Tomato Leaf Disease

Keke Zhang 1 QiufengWu 2 Anwang Liu 1 and XiangyanMeng 2

1College of Engineering Northeast Agricultural University Harbin 150030 China2College of Science Northeast Agricultural University Harbin 150030 China

Correspondence should be addressed to Qiufeng Wu qfwuneaueducn

Received 9 June 2018 Accepted 30 August 2018 Published 26 September 2018

Academic Editor Alexander Loui

Copyright copy 2018 Keke Zhang et al This is an open access article distributed under the Creative Commons Attribution Licensewhich permits unrestricted use distribution and reproduction in any medium provided the original work is properly cited

This paper applies deep convolutional neural network (CNN) to identify tomato leaf disease by transfer learning AlexNetGoogLeNet and ResNet were used as backbone of the CNNThe best combinedmodel was utilized to change the structure aimingat exploring the performance of full training and fine-tuning of CNN The highest accuracy of 9728 for identifying tomato leafdisease is achieved by the optimal model ResNet with stochastic gradient descent (SGD) the number of batch size of 16 the numberof iterations of 4992 and the training layers from the 37 layer to the fully connected layer (denote as ldquofcrdquo)The experimental resultsshow that the proposed technique is effective in identifying tomato leaf disease and could be generalized to identify other plantdiseases

1 Introduction

Tomato is a widely cultivated crop throughout the worldwhich contains rich nutrition unique taste and health effectsso it plays an important role in the agricultural productionand trade around the world Given the importance of tomatoin the economic context it is necessary to maximize produc-tivity and product quality by using techniques Corynesporaleaf spot disease early blight late blight leaf mold diseaseseptoria leaf spot two-spotted spider mite virus disease andyellow leaf curl disease are 8 common diseases in tomato [1ndash8] thus a real time and precise recognition technology isessential

Recently since CNN has the self-learned mechanismthat is extracting features and classifying images in the oneprocedure [9] CNN has been successfully applied in variousapplications such as writer identification [10] salient objectdetection [11 12] scene text detection [13 14] truncatedinference learning [15] road crack detection [16 17] biomed-ical image analysis [18] predicting face attributes from webimages [19] and pedestrian detection [20] and achievedthe better performance In addition CNN is able to extractmore robust and discriminative features with consideringthe global context information of regions [10] and CNN isscarcely affected by the shadow distortion and brightness

of the natural images With the rapid development of CNNmany powerful architectures of CNN emerged such asAlexNet [21] GoogLeNet [22] VGGNet [23] Inception-V3[24] Inception-V4 [25] ResNet [26] and DenseNets [27]

Training deep neural networks from scratch needsamounts of data and expensive computational resourcesMeanwhile we sometimes have a classification task in onedomain but we only have enough data in other domainsFortunately transfer learning can improve the performanceof deep neural networks by avoiding complex data miningand data-labeling efforts [28] In practice transfer learningconsists of two ways [29] One option is to fine-tune thenetworks weights by using our data as input it is worthnothing that the new data must be resized to the input size ofthe pretrained network Another way is to obtain the learnedweights from the pretrained network and apply the weightsto the target network

In this work first we compared the performance betweenSGD [30] and Adaptive Moment Estimation (Adam) [3031] in identifying tomato leaf disease These optimizationmethods are based on the pretrained networks AlexNet[21] GoogLeNet [22] and ResNet [26] Then the networkarchitecture with the highest performance was selected andexperiments on effect of two hyperparameters (ie batch sizeand number of iterations) on accuracy were carried out Next

HindawiAdvances in MultimediaVolume 2018 Article ID 6710865 10 pageshttpsdoiorg10115520186710865

2 Advances in Multimedia

Unknown objective functionf X rarr Y

Tomato leaves training samples(x(1) y(1)) (x(2) y(2))middot middot middot (x(m) y(m))

Learning algorithmA

Hypothesis setH

Final assumptiong asymp f

Figure 1 Proposed workflow diagram

we utilized the network with the suitable hyperparameterswhich was obtained from the previous experiments to dis-cuss the impact of different network structures on recognitiontasksWe believe thismakes sense for researchers who chooseto fine-tune pretrained systems for other similar issues

The rest of this paper is organized as follows Section 2displays an overview of related works Section 3 introducesthe dataset and three deep convolutional neural networksie AlexNet GoogLeNet and ResNet Section 4 presents theexperiments and results in this work Section 5 concludes thepaper

2 Related Work

The research of agricultural disease identification based oncomputer vision has been a hot topic In the early years thetraditional machine learning methods and shallow networkswere extensively adopted in the agricultural field

Sannakki et al [32] proposed to use k-means basedclustering performed on each image pixel to isolate theinfected spot They obtained the result that the GradingSystem they built by machine vision and fuzzy logic is veryuseful for grading the plant disease Samanta et al [33]proposed a novel histogram based scab diseases detectionof potato and applied color image segmentation techniqueto exact intensity pattern They got the best classificationaccuracy of 975 Pedro et al [34] applied fuzzy decision-making to identify weed shape with fuzzy multicriteriadecision-making strategy they achieved the best accuracyof 929 Cheng and Matson [35] adopted Decision TreeSupport Vector Machine (SVM) and Neural Network toidentify weed and rice the best accuracy they achieved is982 by using Decision Tree Sankaran and Ehsani [36]used quadratic discriminant analysis (QDA) and k-nearestneighbour (kNN) to classify citrus leaves infectedwith cankerand Huanglongbing (HLB) from healthy citrus leaves theygot the highest overall accuracy of 999 by kNN

Recently deep learning methods have been applied inidentifying plant disease widely Cheng et al [37] used

ResNet and AlexNet to identify agricultural pests At thesame time they carried out comparative experiments withSVM and BP neural networks finally they got the bestaccuracy of 9867byResNet-101 Ferreiraa et al [38] utilizedConvNets to perform weed detection in soybean crop imagesand classify these weeds among grass and broadleaf Thebest accuracy they achieved is 995 Sladojevic et al [39]built a deep convolutional neural network to automati-cally classify and detect 15 categories of plant leaf diseasesMeanwhile their model was able to distinguish plants fromtheir surroundings They got an average accuracy of 963Mohanty et al [40] trained a deep convolutional neuralnetwork based on the pretrained AlexNet and GoogLeNetto identify 14 crop species and 26 diseases They achievedan accuracy of 9935 on a held-out test set Sa et al[41] proposed a novel approach to fruit detection by usingdeep convolutional neural networks They adapted FasterRegion-based CNN (Faster R-CNN)model through transferlearning They got the F1 score with 083 in a field farmdataset

3 Materials and Methods

This paper concentrates on identifying tomato leaf diseaseby deep learning In this section the abstract mathematicalmodel about identifying tomato leaf disease is displayed atfirst Meanwhile the process of typical CNN is describedwith formulas Then the dataset and data augmentation arepresented Finally we introduced three powerful deep neuralnetworks adopted in this paper ie AlexNet GoogLeNetand ResNet

The main process of tomato leaf disease identification inthis work can be abstracted as a mathematical model (seeFigure 1) First we assume themapping function from tomatoleaves to diseases is 119891 119883 997888rarr 119884 and then send the trainingsamples to the optimization method The hypothesis set 119867means possible objective functions with different parametersthrough a series of parameters update we can get the finalassumption 119892 asymp 119891

Advances in Multimedia 3

Figure 2 Raw tomato leaf images

The typical CNN process can be represented with fol-lowing formulas Firstly send the training samples (ietraining tomato leaf images) to the classifier (ie AlexNetGoogLeNet and ResNet) Then convolution operation iscarried out that is a number of filters slide over the featuremap of the previous layer and the weight matrices do dotproduct

119872119897119895 = 119891(sum119894isin119873119895

119872119897minus1119894 lowast 119908119897119895 + 119887119897119895) (1)

where 119891(sdot) is activation function typically a Rectifier LinearUnit (ReLU) [42] function

119891 (119909) = max (119909 0) (2)

119873119895 is the number of kernels of the certain layer119872119897minus1119894 repre-sents the feature map of the previous layer 119908119897119895 is the weightmatrix and 119887119897119895 is the bias term

Max-pooling or average pooling is conducted after theconvolution operation Furthermore the learned features aresent to the fully connected layer The softmax regressionalways follows the final fully connected layer an input 119909 willget the probability of belonging to class 119894

119901 (119910 = 119894 | 119909 120579) = 119890120579119879119894 119909sum119896119895=1 119890120579119879119894 119909 (3)

where 119910 is the response variable (ie predict label) 119896 is thenumber of categories and 120579 is the parameters of our model

31 Raw Dataset The raw tomato leaf dataset utilized inthis work comes from an open access repository of imageswhich focus on plant health [43] Health and other 8 diseasescategories are included (see Table 1 Figure 2) ie early blight(pathogen Alternaria solani) [1] yellow leaf curl disease(pathogen Tomato Yellow Leaf Curl Virus (Tylcv) FamilyGeminiviridae Genus Begomovirus) [2] corynespora leafspot disease (pathogen Corynespora cassiicola) [3] leafmolddisease (pathogen Fulvia fulva) [4] virus disease (pathogenTomato Mosaic Virus) [5] late blight (pathogen Phytoph-thora Infestans)[6] septoria leaf spot (pathogen Septorialycopersici) [7] and two-spotted spider mite (pathogenTetranychus urticae) [8] The total dataset is 5550

32 Data Augmentation Deep convolutional neural net-works contain millions of parameters thus massive amountsof data is required Otherwise the deep neural network maybe overfitting or not robust The most common method toreduce overfitting on image dataset is to enlarge the datasetmanually and conduct label-preserving transformations [2144]

In this work at first the raw image dataset was dividedinto 80 training samples and 20 testing samples and thenthe data augmentation procedure was conducted (1) flipping

4 Advances in Multimedia

Table 1 The raw tomato leaf dataset

Label Category Number Leaf symptoms Illustration

1 Corynesporaleaf spot disease 547

Small brown spotsappear leaf spots have

yellow halo

See Figure 1 first rowNo1-No5

2 Early blight 405

Black or brown spotsappear leaf spots oftenhave yellow or green

concentric ring pattern

See Figure 1 first rowNo6-No10

4 Late blight 726

Water-soaked areaappears and rapidlyenlarges to formpurple-brown

oily-appearing blotches

See Figure 1 second rowNo1-No5

5 Leaf molddisease 480 Irregular yellow or green

area appearsSee Figure 1 second row

No6-No10

6 Septoria leafspot 734

Round spots marginalbrown chlorotic yellow

appear

See Figure 1 third rowNo1-No5

7 Two-spottedspider mite 720 Show white or yellow

spots blade back nettingSee Figure 1 third row

No6-No10

8 Virus disease 481 Develop yellow or greenslightly shrinking

See Figure 1 forth rowNo1-No5

9 Yellow leaf curldisease 814

Develop small and curlupward crumpling andmarginal yellowingbushy appearance

See Figure 1 forth rowNo6-No10

3 Health 643 See Figure 1 fifth rowTotal 5550

the image from left to right (2) flipping the image from topto bottom (3) flipping the image diagonally (4) adjusting thebrightness of image setting the max delta to 04 (5) adjustingthe contrast of image setting the ratio from 02 to 15 (6)adjusting the hue of image setting the max delta to 05 (7)adjusting the saturation of image setting the ratio from 02 to15 (8) rotating the image by 90∘ and 270∘ respectively Thefinal dataset is shown in Table 2 and the label in the first rowrepresents the disease categories which are given in Table 1

33 Deep Learning Models

331 AlexNet AlexNet is the winner of ImageNet Large-Scale Visual Recognition Challenge (ILSVRC) 2012 a deepconvolutional neural network which has 60 million param-eters and 650000 neurons [21] The architecture of AlexNetutilized in this paper is displayed in Figure 3 The AlexNetarchitecture consists of five convolutional layers (ie conv1conv2 and so on) some of which are followed by max-pooling layers (ie pool1 pool2 and pool5) three fullyconnected layers (ie fc6 fc7 and fc8) and a liner layer withsoftmax activation in output In order to reduce overfittingin the fully connected layers a regularization method calledldquodropoutrdquo is used (ie drop6 drop7) [21] The ReLU activa-tion function is applied to each of the first seven layers (ierelu1 relu2 and so on) [45] In Figure 3 the notation119898times119898times119899in each convolutional layer represents the size of the feature

map for each layer 4096 represents the number of neurons ofthe first two fully connected layers The number of neuronsof the final fully connected layer was modified to 9 sincethe classification problem in this work has 9 categories Inaddition the size of input imagesmust be shaped to 227times227which meets the input pixel size requirement of AlexNet

332 GoogLeNet GoogLeNet is an inception architecture[22] which is the winner of ILSVRC 2014 and owns roughly68 million parameters The architecture of GoogLeNet ispresented in Figure 4 The inception module is inspired bythe network in network [46] and uses a parallel combinationof 1 times 1 3 times 3 and 5 times 5 convolutional layer along with3 times 3 max-pooling layer [45] the 1 times 1 convolutional layerbefore 3 times 3 and 5 times 5 convolutional layer reduces the spatialdimension and limits the size of GoogLeNet The wholearchitecture of GoogLeNet is stacked by inception moduleon top of each other (See Figure 4) which has nine inceptionmodules two convolutional layers four max-pooling layersone average pooling layer one fully connected layer and alinear layer with softmax function in the output GoogLeNetuses dropout regularization in the fully connected layerand applies the ReLU activation function in all of theconvolutional layers [29] In this work the last three layersof GoogLeNet were replaced by a fully connected layer asoftmax layer and a classification layer the fully connectedlayer was modified to 9 neurons which is equal to the

Advances in Multimedia 5

Table 2 The final tomato leaf dataset

Labels Label1 Label2 Label3 Label4 Label5 Label6 Label7 Label8 Label9 TotalTraining set 3933 2916 4626 5229 3456 5283 5184 3465 5859 39951Testing set 110 81 129 145 96 147 144 161 163 1176

Figure 3 The architecture of AlexNet in this work

categories in the tomato leaf disease identification problemThe size requested of input image of GoogLeNet is 224times 224333 ResNet The deep residual learning framework is pro-posed for addressing the degradation problem ResNet con-sists of many stacked residual units which won the first placein ILSVRC 2015 and COCO 2015 classification challenge witherror rate of 357 [26] Each unit can be expressed in thefollowing formulas [47]

119910119897 = ℎ (119909119897) + 119865 (119909119897119882119897) (4)

119909119897+1 = 119891 (119910119897) (5)

where 119909119897 and 119909119897+1 are input and output of the l-th unit and 119865is a residual function In [26]ℎ(119909119897) = 119909119897 is an identity mappingand 119891 is a ReLU function [42] A ldquobottleneckrdquo building blockis designed for ResNet (See Figure 5) and comprises two 1times1convolutions with a 3 times 3 convolution in between and adirect skip connection bypassing input and output The 1 times 1layers are responsible for changing in dimensions ResNetmodel has three types of layers with 50 101 and 152 Forsaving computing resources and training time we choose theResNet50 which also has high performance In this work atfirst the last three layers of ResNet were modified by a fullyconnected layer a softmax layer and a classification layerthe fully connected layer was replaced to 9 neurons which isequal to the categories of the tomato leaf disease We changedthe structure of ResNet subsequently The size of input imageof ResNet should satisfy 224 times 2244 Experiments and Results

In this section we reveal the experiments and discuss theexperimental results All the experiments were implementedin Matlab under Windows 10 using the GPU NVIDIAGTX1050 with 4G video memory or NVIDIA GTX1080Tiwith 11G video memory In this paper overall accuracy wasregarded as the evaluation metric in every experiment on

tomato leaf disease detection which means the percentage ofsamples that are correctly classified

accuracy = true positive + true negativepositive + negative (6)

where ldquotrue positiverdquo is the number of instances that are pos-itive and classified as positive ldquotrue negativerdquo is the numberof instances that are negative and classified as negative andthe denominator represents the total number of samples Inaddition the training time was regarded as an additionalperformance metric of the network structure experiment

41 Experiments on Optimization Methods The first exper-iment is designed for seeking the suitable optimizationmethod between SGD [30] and Adam [30 31] in identifyingtomato leaf diseases combining with the pretrained net-work AlexNet GoogLeNet and ResNet respectively In thisexperiment the hyperparameters were set as follows for eachnetwork the batch size was set to 32 the initial learning ratewas set to 0001 and dropped by a factor of 05 every 2 epochsand the max epoch was set to 5 ie the number of iterationsis 6240 So far as SGD optimization method the momentumwas set to 09 For Adam the gradient decay rate 1205731 was set to09 the squared gradient decay rate 1205732 was set to 0999 andthe denominator offset 120576 was set to 10minus8 [31] The accuracyof different networks is displayed in Table 3 In addition wechoose the better results in each deep neural network to showthe training loss against number of iterations during the fine-tuning process (See Figure 6) The words inside parenthesisindicate the corresponding optimization method

In Table 3 the ResNet with SGD optimization methodgets the highest test accuracy 9651 In identifying tomatoleaf diseases the performance of Adam optimization methodis inferior to the SGD optimization method especially incombining with AlexNet In the following paper AlexNet(SGD) GoogLeNet (SGD) and ResNet (SGD) are referred toas AlexNet GoogLeNet and ResNet respectively

As it can be seen in Figure 6 the training loss of ResNetdrops rapidly in the earlier iterations and tends to stable after3000 iterations Consistent with Table 3 the performance of

6 Advances in Multimedia

Inputimage

Conv1 Conv2

MaxPool MaxPool

MaxPool

MaxPool

Inception 3a

Inception 3b

Inception 4a

Inception 4b

Inception 4c

Inception 4d

Inception 4e

Inception 5a

Inception 5b

AvgPool

Dropout40

FC

somax

output

Figure 4 The architecture of GoogLeNet [22 45]

1times1 conv 64

3times3 conv 64

1times1 conv 256

relu

relu

relu

+

Figure 5 ResNet bottleneck residual building block [26]

Table 3 Model recognition accuracy

Model AccuracyAlexNet (SGD) 9583AlexNet (Adam) 1386GoogLeNet (SGD) 9566GoogLeNet (Adam) 9406ResNet (SGD) 9651ResNet (Adam) 9439

AlexNet and GoogLeNet is similar and both inferior to theResNet

42 Experiments on Batch Size and Number of IterationsFrom the experiment on optimization methods the ResNetobtains the highest classification accuracy Next we evaluatedthe effects of batch size and the number of iterations on theperformance of the ResNet The batch size was set to 16 32and 64 respectively Meanwhile the number of iterations wasset to 2496 4992 and 9984 The classification accuracy ofdifferent training scenarios is given in Table 4 At the sametime the classification accuracy of each labels representativeleaf disease category (See Table 1) is given In this experimentthe initial learning rate was set to 0001 and dropped by afactor of 05 every 2496 iterations

In Table 4 the best overall classification accuracy 9719is got by the ResNet combining with batch size 16 and

Trai

ning

Los

s

Number of Iterations

AlexNetGoogLeNetResNet

3

25

2

15

1

05

00 1000 2000 3000 4000 5000 6000

Figure 6 The training loss during the fine-tuning process

iterations 4992 As shown in Table 4 whether increasingthe number of iterations or batch size the performance ofcorresponding models has not been improved significantlyin identifying tomato leaf disease A small batch size witha medium number of iterations is quite effective in thiswork Moreover a larger batch size and number of iterationsincreases the training duration We have not tried higher orlower values for the attempted parameters since differentclassification task may have various suitable parameters andit is hard to give a certain rule in setting hyperparameters

43 Experiments on Full Training and Fine-Tuning of ResNetThis section is designed for exploring the performance ofCNN by changing the structure of the models In practicala deep CNN always owns a large size which means a largenumber of parameters Thus full training of a deep CNNrequires extensive computational resources and is time-consuming In addition full training of a deep CNN mayled to overfitting when the training data is limited So wecompared the performance of the pretrained CNN throughfull training and fine-tuning their structures

We changed the structure of ResNet and combination ofthe best parameters from the front experiments was utilized

Advances in Multimedia 7

Table 4 Classification accuracies with different parameters during fine-tuning of the ResNetThe numbers inside parenthesis indicate batchsize and number of iterations

Networks label1 label2 label3 label4 label5 label6 label7 label8 label9 overallResNet (162496) 9091 8889 100 100 9688 100 9028 8820 9755 9498ResNet (164992) 9818 9877 100 9862 9688 100 9653 8882 9877 9719ResNet (169984) 9818 9753 100 9862 9688 100 9722 8696 9877 9694ResNet (322496) 9727 9506 100 9793 9688 100 9514 8634 9939 9634ResNet (324992) 9727 9506 100 9724 9688 100 9653 8696 9939 9651ResNet (329984) 9636 9630 100 9931 9688 100 9444 8820 9877 9660ResNet (642496) 9364 9383 100 9724 9688 100 9444 8758 9939 9592ResNet (644992) 9455 9529 100 9655 9583 100 9583 8696 9939 9583ResNet (649984) 9545 9383 100 9793 9688 100 9444 8758 9939 9617

ResNet50 has 177 layers if the layers for each building blockand connection are calculated In this experiment the lastthree layers of ResNet were modified to a fully connectedlayer (denoted as ldquofcrdquo) a softmax layer and a classificationlayer and the fully connected layer owns 9 neurons Thestructure was changed by freezing the weights of a certainnumber of layers in the network by setting the learning ratein those layers to zero During training the parameters ofthe frozen layers are not updated Full training and fine-tuning are defined by the number of training layers ie fulltraining (1-ldquofcrdquo) fine-tuning (37-ldquofcrdquo 79-ldquofcrdquo 111-ldquofcrdquo 141-ldquofcrdquo 163-ldquofcrdquo) The accuracy and training time of differentnetwork structure are presented in Table 5 At first the batchsize and 4992 iterations were combined the initial learningrate was set to 0001 and dropped by a factor of 01 every2496 iterations In order to get more convincing conclusionsResNet (16 9984) which gets the second place in Table 4 wasalso used to execute the experiments

In Table 5 the accuracy and training time of differentnetwork structures are presented In two cases ie the 4992iterations and 9984 iterations of ResNet the accuracy of themodel from the 37 layer fine-tuning structure are higherthan that of the full training model In the case where thenumber of iterations is 4992 the accuracy of the model fromthe 79 layer fine-tuning structure is equal to that of the fulltraining modelThe final column of the Table 5 represents thetraining time of the corresponding network and it is clearthat the training time of the fine-tuning models is greatlylowered than the full training model Because the gradientsof the frozen layers do not need to be computed freezingthe weights of initial layers can speed up network trainingWe observe that the moderate fine-tuning models (37-ldquofcrdquo79-ldquofcrdquo 111-ldquofcrdquo) always led to a performance superior orapproximately equal to the full training models Thus wesuggest that for practical application the moderate fine-tuning models may be a good choice Especially for theresearcher who holds massive data the fine-tuning modelsmay achieve good performance while saving computationalresources and time

Moreover the features of the final fully connected layerof ResNet (16 4992 37-ldquofcrdquo) were examined by utilizingthe t-distributed Stochastic Neighbour Embedding (t-SNE)algorithm (see Figure 7) [48] 1176 test images were used to

label1 label2 label3 label4 label5

label6 label7 label8 label9

Figure 7 Two-dimensional scatter plot of high-dimensional fea-tures generated with t-SNE

extract the features In Figure 7 different colors representdifferent labels the corresponding disease categories of thelabels were listed in Table 1 As shown in Figure 7 9 differentcolor points are clearly separated which indicates that thefeatures learned from the ResNet with the optimal structurecan be used to classify the tomato leaf disease precisely

5 Conclusion

This paper concentrates on identifying tomato leaf diseaseusing deep convolutional neural networks by transfer learn-ing The utilized networks are based on the pretrained deeplearning models of AlexNet GoogLeNet and ResNet Firstwe compared the relative performance of these networksby using SGD and Adam optimization method revealingthat the ResNet with SGD optimization method obtainsthe highest result with the best accuracy 9651 Thenthe performance evaluation of batch size and number of

8 Advances in Multimedia

Table 5 Accuracies and training time in different network structuresThe values inside parenthesis denote batch size number of iterationsand training layers

Network topology Accuracy Time (minsec)ResNet (16 4992 1-ldquofcrdquo) 9643 59min 30secResNet (16 4992 37-ldquofcrdquo) 9728 44min 13secResNet (16 4992 79-ldquofcrdquo) 9643 37min 27secResNet (16 4992 111-ldquofcrdquo) 9575 30min 6secResNet (16 4992 141-ldquofcrdquo) 9532 24min 15secResNet (16 4992 163-ldquofcrdquo) 9269 19min 31secResNet (16 9984 1-ldquofcrdquo) 9694 118min 32secResNet (16 9984 37-ldquofcrdquo) 9702 92min 53secResNet (16 9984 79-ldquofcrdquo) 9677 72min 23secResNet (16 9984 111-ldquofcrdquo) 9626 58min 40secResNet (16 9984 141-ldquofcrdquo) 9575 47min 22secResNet (16 9984 163-ldquofcrdquo) 9396 39min 32sec

iterations affecting the transfer learning of the ResNet wasconducted A small batch size of 16 combining a moderatenumber of iterations of 4992 is the optimal choice in thiswork Our findings suggest that for a particular task neitherlarge batch size nor large number of iterations may improvethe accuracy of the target modelThe setting of batch size andnumber of iterations depends on your data set and the utilizednetwork Next the best combined model was used to fine-tune the structure Fine-tuning ResNet layers from 37 to ldquofcrdquoobtained the highest accuracy 9728 in identifying tomatoleaf disease Based on the amount of available data layer-wisefine-tuning may provide a practical way to achieve the bestperformance of the application at hand We believe that theresults obtained in this work will bring some inspiration toother similar visual recognition problems and the practicalstudy of this work can be easily extended to other plant leafdisease identification problems

Data Availability

Thetomato leaf data supporting this work are frompreviouslyreported studies which have been cited The processed dataare available from the corresponding author request

Conflicts of Interest

The authors declare that there are no conflicts of interestregarding the publication of this paper

Acknowledgments

This study was supported by the National Science and tech-nology support program (2014BAD12B01-1-3) PublicWelfareIndustry (Agriculture) Research Projects Level-2 (201503116-04-06) Postdoctoral Foundation of Heilongjiang Province(LBHZ15020) Harbin Applied Technology Research andDevelopment Program (2017RAQXJ096) and EconomicDecision Making and Early Warning of Soybean Industryin Technology Collaborative Innovation System of SoybeanIndustry in Heilongjiang Province (20170401)

References

[1] RChaerani andR EVoorrips ldquoTomato early blight (Alternariasolani) The pathogen genetics and breeding for resistancerdquoJournal of General Plant Pathology vol 72 no 6 pp 335ndash3472006

[2] A M Dickey L S Osborne and C L Mckenzie ldquoPapaya(Carica papaya Brassicales Caricaceae) is not a host plant oftomato yellow leaf curl virus (TYLCV family Geminiviridaegenus Begomovirus)rdquo Florida Entomologist vol 95 no 1 pp211ndash213 2012

[3] G Wei L Baoju S Yanxia and X Xuewen ldquoStudies on path-ogenicity differentiation of corynespora cassiicola isolatesagainst cucumber tomato and eggplantrdquo Acta HorticulturaeSinica vol 38 no 3 pp 465ndash470 2011

[4] P Lindhout W Korta M Cislik I Vos and T GerlaghldquoFurther identification of races of Cladosporium fulvum (Fulviafulva) on tomato originating from the Netherlands France andPolandrdquo Netherlands Journal of Plant Pathology vol 95 no 3pp 143ndash148 1989

[5] K Kubota S Tsuda A Tamai and T Meshi ldquoTomato mosaicvirus replication protein suppresses virus-targeted posttran-scriptional gene silencingrdquo Journal of Virology vol 77 no 20pp 11016ndash11026 2003

[6] M Tian B Benedetti and S Kamoun ldquoA second Kazal-like protease inhibitor from Phytophthora infestans inhibitsand interacts with the apoplastic pathogenesis-related proteaseP69B of tomatordquo Plant Physiology vol 138 no 3 pp 1785ndash17932005

[7] L E Blum ldquoReduction of incidence and severity of Septorialycopersici leaf spot of tomatowith bacteria and yeastsrdquoCienciaRural vol 30 no 5 pp 761ndash765 2000

[8] E A Chatzivasileiadis and M W Sabelis ldquoToxicity of methylketones from tomato trichomes to Tetranychus urticae KochrdquoExperimental and Applied Acarology vol 21 no 6-7 pp 473ndash484 1997

[9] M Anthimopoulos S Christodoulidis L Ebner A Christeand S Mougiakakou ldquoLung Pattern Classification for Inter-stitial Lung Diseases Using a Deep Convolutional NeuralNetworkrdquo IEEE Transactions on Medical Imaging vol 35 no5 pp 1207ndash1216 2016

[10] Y Tang and X Wu ldquoText-independent writer identification viaCNN features and joint Bayesianrdquo in Proceedings of the 15th

Advances in Multimedia 9

International Conference on Frontiers in Handwriting Recogni-tion ICFHR 2016 pp 566ndash571 Shenzhen China October 2016

[11] Y Tang and XWu ldquoSaliency Detection via Combining Region-Level and Pixel-Level Predictions with CNNsrdquo in Proceedings ofthe EuropeanConference onComputer Vision (ECCV) pp 1608ndash05186 2016

[12] Y Tang and X Wu ldquoSalient object detection with chainedmulti-scale fully convolutional networkrdquo ACM Multimedia(ACMMM) pp 618ndash626 2017

[13] Y Tang and X Wu ldquoScene text detection and segmentationbased on cascaded convolution neural networksrdquo IEEE Trans-actions on Image Processing vol 26 no 3 pp 1509ndash1520 2017

[14] Y Tang and X Wu ldquoScene Text Detection using Superpixelbased Stroke Feature Transform and Deep Learning basedRegion Classificationrdquo IEEE Transactions on Multimedia vol20 no 9 pp 2276ndash2288 2018

[15] Y Yao XWu Z Lei S Shan andW Zuo ldquoJoint Representationand Truncated Inference Learning for Correlation Filter basedTrackingrdquo in Proceedings of the European Conference on Com-puter Vision (ECCV) pp 1ndash14 2018

[16] L Zhang F Yang Y Daniel Zhang and Y J Zhu ldquoRoadcrack detection using deep convolutional neural networkrdquoin Proceedings of the 23rd IEEE International Conference onImage Processing ICIP 2016 pp 3708ndash3712 Phoenix AZ USASeptember 2016

[17] D Xie L Zhang andL Bai ldquoDeep learning in visual computingand signal processingrdquo Applied Computational Intelligence andSoft Computing vol 2017 Article ID 1320780 13 pages 2017

[18] Z Zhou J Shin L Zhang S Gurudu M Gotway and J LiangldquoFine-tuning convolutional neural networks for biomedicalimage analysis Actively and incrementallyrdquo inProceedings of the30th IEEE Conference on Computer Vision and Pattern Recog-nition CVPR 2017 pp 4761ndash4772 USA July 2017

[19] Z Liu P Luo X Wang and X Tang ldquoDeep learning face attri-butes in the wildrdquo in Proceedings of the 15th IEEE InternationalConference on Computer Vision ICCV 2015 pp 3730ndash3738Santiago Chile 2015

[20] W Ouyang and X Wang ldquoJoint deep learning for pedestriandetectionrdquo in Proceedings of the 14th IEEE International Con-ference on Computer Vision (ICCV rsquo13) pp 2056ndash2063 SydneyAustralia December 2013

[21] A Krizhevsky I Sutskever andG EHinton ldquoImagenet classifi-cation with deep convolutional neural networksrdquo in Proceedingsof the 26th Annual Conference on Neural Information ProcessingSystems (NIPS rsquo12) pp 1097ndash1105 Lake Tahoe Nev USADecember 2012

[22] C Szegedy W Liu Y Jia et al ldquoGoing deeper with convolu-tionsrdquo inProceedings of the IEEEConference onComputer Visionand Pattern Recognition (CVPR rsquo15) pp 1ndash9 IEEE BostonMass USA June 2015

[23] K Simonyan and A Zisserman ldquoVery deep convolutional net-works for large-scale image recognitionrdquo httpsarxivorgabs14091556 2015

[24] C Szegedy V Vanhoucke S Ioffe J Shlens and Z WojnaldquoRethinking the inception architecture for computer visionrdquohttpsarxivorgabs151200567 2015

[25] C Szegedy S Ioffe V Vanhoucke and A Alemi ldquoInception-v4 inception-ResNet and the impact of residual connections onlearningrdquo httpsarxivorgabs160207261 2016

[26] K He X Zhang S Ren and J Sun ldquoDeep residual learning forimage recognitionrdquo in Proceedings of the 2016 IEEE Conference

on Computer Vision and Pattern Recognition CVPR 2016 pp770ndash778 July 2016

[27] G Huang Z Liu K Q Weinberger and L van der MaatenldquoDensely connected convolutional networksrdquo in Proceedingsof the 2017 IEEE Conference on Computer Vision and PatternRecognition (CVPR) Honolulu HI USA 2016

[28] S J Pan and Q Yang ldquoA survey on transfer learningrdquo IEEETransactions on Knowledge and Data Engineering vol 22 no10 pp 1345ndash1359 2010

[29] M Mehdipour Ghazi B Yanikoglu and E Aptoula ldquoPlantidentification using deep neural networks via optimization oftransfer learning parametersrdquo Neurocomputing vol 235 pp228ndash235 2017

[30] S Ruder ldquoAn overview of gradient descent optimization algo-rithmsrdquo httpsarxivorgabs160904747 2017

[31] D P Kingma and J L Ba ldquoAdam a method for stochasticoptimizationrdquo httpsarxivorgabs14126980 2017

[32] S S Sannakki V S Rajpurohit V B Nargund R Arun Kumarand S Prema Yallur ldquoLeaf disease grading by machine visionand fuzzy logicrdquo International Journal of Computer Technologyand Applications vol 2 no 5 pp 1709ndash1716 2011

[33] D Samanta P P Chaudhury and A Ghosh ldquoScab diseasesdetection of potato using image processingrdquo InternationalJournal of Computer Trends and Technology vol 3 pp 109ndash1132012

[34] P J Herrera J Dorado and A Ribeiro ldquoA novel approach forweed type classification based on shape descriptors and a fuzzydecision-making methodrdquo Sensors vol 14 no 8 pp 15304ndash15324 2014

[35] B Cheng and E T Matson ldquoA feature-based machine learningagent for automatic rice andweed discriminationrdquo InternationalConference onArtificial Intelligence and SoftComputing pp 517ndash527 2015

[36] S Sankaran and R Ehsani ldquoComparison of visible-near infra-red and mid-infrared spectroscopy for classification of Huang-longbing and citrus canker infected leavesrdquo Agricultural Engi-neering International CIGR Journal vol 15 no 3 pp 75ndash792013

[37] X Cheng Y Zhang Y Chen Y Wu and Y Yue ldquoPest iden-tification via deep residual learning in complex backgroundrdquoComputers and Electronics in Agriculture vol 141 pp 351ndash3562017

[38] A dos Santos Ferreira D Matte Freitas G Goncalves daSilva H Pistori and M Theophilo Folhes ldquoWeed detection insoybean crops using ConvNetsrdquo Computers and Electronics inAgriculture vol 143 pp 314ndash324 2017

[39] S Sladojevic M Arsenovic A Anderla D Culibrk andD Stefanovic ldquoDeep Neural Networks Based Recognition ofPlant Diseases by Leaf Image Classificationrdquo ComputationalIntelligence and Neuroscience vol 2016 Article ID 3289801 11pages 2016

[40] S P Mohanty D P Hughes and M Salathe ldquoUsing deep learn-ing for image-based plant disease detectionrdquo Frontiers in PlantScience vol 7 article no 1419 2016

[41] I Sa Z Ge F Dayoub B Upcroft T Perez and C McCoolldquoDeepfruits A fruit detection system using deep neural net-worksrdquo Sensors vol 16 article no 1222 no 8 2016

[42] V Nair and G E Hinton ldquoRectified linear units improveRestricted Boltzmann machinesrdquo in Proceedings of the 27thInternational Conference on Machine Learning (ICML rsquo10) pp807ndash814 Haifa Israel June 2010

10 Advances in Multimedia

[43] D P Hughes and M Salathe ldquoAn open access repository ofimages on plant health to enable the development of mobiledisease diagnosticsrdquo httpsarxivorgabs151108060 2016

[44] D Ciresan U Meier J Masci and J Schmidhuber ldquoA com-mittee of neural networks for traffic sign classificationrdquo inProceedings of the 2011 International Joint Conference on NeuralNetworks (IJCNN 2011 - San Jose) San Jose CA USA July 2011

[45] P Pawara E Okafor O Surinta L Schomaker andMWieringldquoComparing Local Descriptors and Bags of Visual Words toDeep Convolutional Neural Networks for Plant Recognitionrdquoin Proceedings of the 6th International Conference on PatternRecognition Applications and Methods pp 479ndash486 PortoPortugal Feburary 2017

[46] M Lin ldquoNetwork inNnetworkrdquo httpsarxivorgabs131244002014

[47] K He X Zhang S Ren and J Sun ldquoIdentity mappings in deepresidual networksrdquo in Proceedings of the European Conferenceon Computer Vision pp 630ndash645 2016

[48] L van der Maaten and G Hinton ldquoVisualizing data using t-SNErdquo Journal of Machine Learning Research vol 9 pp 2579ndash2625 2008

International Journal of

AerospaceEngineeringHindawiwwwhindawicom Volume 2018

RoboticsJournal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Active and Passive Electronic Components

VLSI Design

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Shock and Vibration

Hindawiwwwhindawicom Volume 2018

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawiwwwhindawicom

Volume 2018

Hindawi Publishing Corporation httpwwwhindawicom Volume 2013Hindawiwwwhindawicom

The Scientific World Journal

Volume 2018

Control Scienceand Engineering

Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom

Journal ofEngineeringVolume 2018

SensorsJournal of

Hindawiwwwhindawicom Volume 2018

International Journal of

RotatingMachinery

Hindawiwwwhindawicom Volume 2018

Modelling ampSimulationin EngineeringHindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Navigation and Observation

International Journal of

Hindawi

wwwhindawicom Volume 2018

Advances in

Multimedia

Submit your manuscripts atwwwhindawicom

Page 2: ResearchArticle Can Deep Learning Identify Tomato Leaf

2 Advances in Multimedia

Unknown objective functionf X rarr Y

Tomato leaves training samples(x(1) y(1)) (x(2) y(2))middot middot middot (x(m) y(m))

Learning algorithmA

Hypothesis setH

Final assumptiong asymp f

Figure 1 Proposed workflow diagram

we utilized the network with the suitable hyperparameterswhich was obtained from the previous experiments to dis-cuss the impact of different network structures on recognitiontasksWe believe thismakes sense for researchers who chooseto fine-tune pretrained systems for other similar issues

The rest of this paper is organized as follows Section 2displays an overview of related works Section 3 introducesthe dataset and three deep convolutional neural networksie AlexNet GoogLeNet and ResNet Section 4 presents theexperiments and results in this work Section 5 concludes thepaper

2 Related Work

The research of agricultural disease identification based oncomputer vision has been a hot topic In the early years thetraditional machine learning methods and shallow networkswere extensively adopted in the agricultural field

Sannakki et al [32] proposed to use k-means basedclustering performed on each image pixel to isolate theinfected spot They obtained the result that the GradingSystem they built by machine vision and fuzzy logic is veryuseful for grading the plant disease Samanta et al [33]proposed a novel histogram based scab diseases detectionof potato and applied color image segmentation techniqueto exact intensity pattern They got the best classificationaccuracy of 975 Pedro et al [34] applied fuzzy decision-making to identify weed shape with fuzzy multicriteriadecision-making strategy they achieved the best accuracyof 929 Cheng and Matson [35] adopted Decision TreeSupport Vector Machine (SVM) and Neural Network toidentify weed and rice the best accuracy they achieved is982 by using Decision Tree Sankaran and Ehsani [36]used quadratic discriminant analysis (QDA) and k-nearestneighbour (kNN) to classify citrus leaves infectedwith cankerand Huanglongbing (HLB) from healthy citrus leaves theygot the highest overall accuracy of 999 by kNN

Recently deep learning methods have been applied inidentifying plant disease widely Cheng et al [37] used

ResNet and AlexNet to identify agricultural pests At thesame time they carried out comparative experiments withSVM and BP neural networks finally they got the bestaccuracy of 9867byResNet-101 Ferreiraa et al [38] utilizedConvNets to perform weed detection in soybean crop imagesand classify these weeds among grass and broadleaf Thebest accuracy they achieved is 995 Sladojevic et al [39]built a deep convolutional neural network to automati-cally classify and detect 15 categories of plant leaf diseasesMeanwhile their model was able to distinguish plants fromtheir surroundings They got an average accuracy of 963Mohanty et al [40] trained a deep convolutional neuralnetwork based on the pretrained AlexNet and GoogLeNetto identify 14 crop species and 26 diseases They achievedan accuracy of 9935 on a held-out test set Sa et al[41] proposed a novel approach to fruit detection by usingdeep convolutional neural networks They adapted FasterRegion-based CNN (Faster R-CNN)model through transferlearning They got the F1 score with 083 in a field farmdataset

3 Materials and Methods

This paper concentrates on identifying tomato leaf diseaseby deep learning In this section the abstract mathematicalmodel about identifying tomato leaf disease is displayed atfirst Meanwhile the process of typical CNN is describedwith formulas Then the dataset and data augmentation arepresented Finally we introduced three powerful deep neuralnetworks adopted in this paper ie AlexNet GoogLeNetand ResNet

The main process of tomato leaf disease identification inthis work can be abstracted as a mathematical model (seeFigure 1) First we assume themapping function from tomatoleaves to diseases is 119891 119883 997888rarr 119884 and then send the trainingsamples to the optimization method The hypothesis set 119867means possible objective functions with different parametersthrough a series of parameters update we can get the finalassumption 119892 asymp 119891

Advances in Multimedia 3

Figure 2 Raw tomato leaf images

The typical CNN process can be represented with fol-lowing formulas Firstly send the training samples (ietraining tomato leaf images) to the classifier (ie AlexNetGoogLeNet and ResNet) Then convolution operation iscarried out that is a number of filters slide over the featuremap of the previous layer and the weight matrices do dotproduct

119872119897119895 = 119891(sum119894isin119873119895

119872119897minus1119894 lowast 119908119897119895 + 119887119897119895) (1)

where 119891(sdot) is activation function typically a Rectifier LinearUnit (ReLU) [42] function

119891 (119909) = max (119909 0) (2)

119873119895 is the number of kernels of the certain layer119872119897minus1119894 repre-sents the feature map of the previous layer 119908119897119895 is the weightmatrix and 119887119897119895 is the bias term

Max-pooling or average pooling is conducted after theconvolution operation Furthermore the learned features aresent to the fully connected layer The softmax regressionalways follows the final fully connected layer an input 119909 willget the probability of belonging to class 119894

119901 (119910 = 119894 | 119909 120579) = 119890120579119879119894 119909sum119896119895=1 119890120579119879119894 119909 (3)

where 119910 is the response variable (ie predict label) 119896 is thenumber of categories and 120579 is the parameters of our model

31 Raw Dataset The raw tomato leaf dataset utilized inthis work comes from an open access repository of imageswhich focus on plant health [43] Health and other 8 diseasescategories are included (see Table 1 Figure 2) ie early blight(pathogen Alternaria solani) [1] yellow leaf curl disease(pathogen Tomato Yellow Leaf Curl Virus (Tylcv) FamilyGeminiviridae Genus Begomovirus) [2] corynespora leafspot disease (pathogen Corynespora cassiicola) [3] leafmolddisease (pathogen Fulvia fulva) [4] virus disease (pathogenTomato Mosaic Virus) [5] late blight (pathogen Phytoph-thora Infestans)[6] septoria leaf spot (pathogen Septorialycopersici) [7] and two-spotted spider mite (pathogenTetranychus urticae) [8] The total dataset is 5550

32 Data Augmentation Deep convolutional neural net-works contain millions of parameters thus massive amountsof data is required Otherwise the deep neural network maybe overfitting or not robust The most common method toreduce overfitting on image dataset is to enlarge the datasetmanually and conduct label-preserving transformations [2144]

In this work at first the raw image dataset was dividedinto 80 training samples and 20 testing samples and thenthe data augmentation procedure was conducted (1) flipping

4 Advances in Multimedia

Table 1 The raw tomato leaf dataset

Label Category Number Leaf symptoms Illustration

1 Corynesporaleaf spot disease 547

Small brown spotsappear leaf spots have

yellow halo

See Figure 1 first rowNo1-No5

2 Early blight 405

Black or brown spotsappear leaf spots oftenhave yellow or green

concentric ring pattern

See Figure 1 first rowNo6-No10

4 Late blight 726

Water-soaked areaappears and rapidlyenlarges to formpurple-brown

oily-appearing blotches

See Figure 1 second rowNo1-No5

5 Leaf molddisease 480 Irregular yellow or green

area appearsSee Figure 1 second row

No6-No10

6 Septoria leafspot 734

Round spots marginalbrown chlorotic yellow

appear

See Figure 1 third rowNo1-No5

7 Two-spottedspider mite 720 Show white or yellow

spots blade back nettingSee Figure 1 third row

No6-No10

8 Virus disease 481 Develop yellow or greenslightly shrinking

See Figure 1 forth rowNo1-No5

9 Yellow leaf curldisease 814

Develop small and curlupward crumpling andmarginal yellowingbushy appearance

See Figure 1 forth rowNo6-No10

3 Health 643 See Figure 1 fifth rowTotal 5550

the image from left to right (2) flipping the image from topto bottom (3) flipping the image diagonally (4) adjusting thebrightness of image setting the max delta to 04 (5) adjustingthe contrast of image setting the ratio from 02 to 15 (6)adjusting the hue of image setting the max delta to 05 (7)adjusting the saturation of image setting the ratio from 02 to15 (8) rotating the image by 90∘ and 270∘ respectively Thefinal dataset is shown in Table 2 and the label in the first rowrepresents the disease categories which are given in Table 1

33 Deep Learning Models

331 AlexNet AlexNet is the winner of ImageNet Large-Scale Visual Recognition Challenge (ILSVRC) 2012 a deepconvolutional neural network which has 60 million param-eters and 650000 neurons [21] The architecture of AlexNetutilized in this paper is displayed in Figure 3 The AlexNetarchitecture consists of five convolutional layers (ie conv1conv2 and so on) some of which are followed by max-pooling layers (ie pool1 pool2 and pool5) three fullyconnected layers (ie fc6 fc7 and fc8) and a liner layer withsoftmax activation in output In order to reduce overfittingin the fully connected layers a regularization method calledldquodropoutrdquo is used (ie drop6 drop7) [21] The ReLU activa-tion function is applied to each of the first seven layers (ierelu1 relu2 and so on) [45] In Figure 3 the notation119898times119898times119899in each convolutional layer represents the size of the feature

map for each layer 4096 represents the number of neurons ofthe first two fully connected layers The number of neuronsof the final fully connected layer was modified to 9 sincethe classification problem in this work has 9 categories Inaddition the size of input imagesmust be shaped to 227times227which meets the input pixel size requirement of AlexNet

332 GoogLeNet GoogLeNet is an inception architecture[22] which is the winner of ILSVRC 2014 and owns roughly68 million parameters The architecture of GoogLeNet ispresented in Figure 4 The inception module is inspired bythe network in network [46] and uses a parallel combinationof 1 times 1 3 times 3 and 5 times 5 convolutional layer along with3 times 3 max-pooling layer [45] the 1 times 1 convolutional layerbefore 3 times 3 and 5 times 5 convolutional layer reduces the spatialdimension and limits the size of GoogLeNet The wholearchitecture of GoogLeNet is stacked by inception moduleon top of each other (See Figure 4) which has nine inceptionmodules two convolutional layers four max-pooling layersone average pooling layer one fully connected layer and alinear layer with softmax function in the output GoogLeNetuses dropout regularization in the fully connected layerand applies the ReLU activation function in all of theconvolutional layers [29] In this work the last three layersof GoogLeNet were replaced by a fully connected layer asoftmax layer and a classification layer the fully connectedlayer was modified to 9 neurons which is equal to the

Advances in Multimedia 5

Table 2 The final tomato leaf dataset

Labels Label1 Label2 Label3 Label4 Label5 Label6 Label7 Label8 Label9 TotalTraining set 3933 2916 4626 5229 3456 5283 5184 3465 5859 39951Testing set 110 81 129 145 96 147 144 161 163 1176

Figure 3 The architecture of AlexNet in this work

categories in the tomato leaf disease identification problemThe size requested of input image of GoogLeNet is 224times 224333 ResNet The deep residual learning framework is pro-posed for addressing the degradation problem ResNet con-sists of many stacked residual units which won the first placein ILSVRC 2015 and COCO 2015 classification challenge witherror rate of 357 [26] Each unit can be expressed in thefollowing formulas [47]

119910119897 = ℎ (119909119897) + 119865 (119909119897119882119897) (4)

119909119897+1 = 119891 (119910119897) (5)

where 119909119897 and 119909119897+1 are input and output of the l-th unit and 119865is a residual function In [26]ℎ(119909119897) = 119909119897 is an identity mappingand 119891 is a ReLU function [42] A ldquobottleneckrdquo building blockis designed for ResNet (See Figure 5) and comprises two 1times1convolutions with a 3 times 3 convolution in between and adirect skip connection bypassing input and output The 1 times 1layers are responsible for changing in dimensions ResNetmodel has three types of layers with 50 101 and 152 Forsaving computing resources and training time we choose theResNet50 which also has high performance In this work atfirst the last three layers of ResNet were modified by a fullyconnected layer a softmax layer and a classification layerthe fully connected layer was replaced to 9 neurons which isequal to the categories of the tomato leaf disease We changedthe structure of ResNet subsequently The size of input imageof ResNet should satisfy 224 times 2244 Experiments and Results

In this section we reveal the experiments and discuss theexperimental results All the experiments were implementedin Matlab under Windows 10 using the GPU NVIDIAGTX1050 with 4G video memory or NVIDIA GTX1080Tiwith 11G video memory In this paper overall accuracy wasregarded as the evaluation metric in every experiment on

tomato leaf disease detection which means the percentage ofsamples that are correctly classified

accuracy = true positive + true negativepositive + negative (6)

where ldquotrue positiverdquo is the number of instances that are pos-itive and classified as positive ldquotrue negativerdquo is the numberof instances that are negative and classified as negative andthe denominator represents the total number of samples Inaddition the training time was regarded as an additionalperformance metric of the network structure experiment

41 Experiments on Optimization Methods The first exper-iment is designed for seeking the suitable optimizationmethod between SGD [30] and Adam [30 31] in identifyingtomato leaf diseases combining with the pretrained net-work AlexNet GoogLeNet and ResNet respectively In thisexperiment the hyperparameters were set as follows for eachnetwork the batch size was set to 32 the initial learning ratewas set to 0001 and dropped by a factor of 05 every 2 epochsand the max epoch was set to 5 ie the number of iterationsis 6240 So far as SGD optimization method the momentumwas set to 09 For Adam the gradient decay rate 1205731 was set to09 the squared gradient decay rate 1205732 was set to 0999 andthe denominator offset 120576 was set to 10minus8 [31] The accuracyof different networks is displayed in Table 3 In addition wechoose the better results in each deep neural network to showthe training loss against number of iterations during the fine-tuning process (See Figure 6) The words inside parenthesisindicate the corresponding optimization method

In Table 3 the ResNet with SGD optimization methodgets the highest test accuracy 9651 In identifying tomatoleaf diseases the performance of Adam optimization methodis inferior to the SGD optimization method especially incombining with AlexNet In the following paper AlexNet(SGD) GoogLeNet (SGD) and ResNet (SGD) are referred toas AlexNet GoogLeNet and ResNet respectively

As it can be seen in Figure 6 the training loss of ResNetdrops rapidly in the earlier iterations and tends to stable after3000 iterations Consistent with Table 3 the performance of

6 Advances in Multimedia

Inputimage

Conv1 Conv2

MaxPool MaxPool

MaxPool

MaxPool

Inception 3a

Inception 3b

Inception 4a

Inception 4b

Inception 4c

Inception 4d

Inception 4e

Inception 5a

Inception 5b

AvgPool

Dropout40

FC

somax

output

Figure 4 The architecture of GoogLeNet [22 45]

1times1 conv 64

3times3 conv 64

1times1 conv 256

relu

relu

relu

+

Figure 5 ResNet bottleneck residual building block [26]

Table 3 Model recognition accuracy

Model AccuracyAlexNet (SGD) 9583AlexNet (Adam) 1386GoogLeNet (SGD) 9566GoogLeNet (Adam) 9406ResNet (SGD) 9651ResNet (Adam) 9439

AlexNet and GoogLeNet is similar and both inferior to theResNet

42 Experiments on Batch Size and Number of IterationsFrom the experiment on optimization methods the ResNetobtains the highest classification accuracy Next we evaluatedthe effects of batch size and the number of iterations on theperformance of the ResNet The batch size was set to 16 32and 64 respectively Meanwhile the number of iterations wasset to 2496 4992 and 9984 The classification accuracy ofdifferent training scenarios is given in Table 4 At the sametime the classification accuracy of each labels representativeleaf disease category (See Table 1) is given In this experimentthe initial learning rate was set to 0001 and dropped by afactor of 05 every 2496 iterations

In Table 4 the best overall classification accuracy 9719is got by the ResNet combining with batch size 16 and

Trai

ning

Los

s

Number of Iterations

AlexNetGoogLeNetResNet

3

25

2

15

1

05

00 1000 2000 3000 4000 5000 6000

Figure 6 The training loss during the fine-tuning process

iterations 4992 As shown in Table 4 whether increasingthe number of iterations or batch size the performance ofcorresponding models has not been improved significantlyin identifying tomato leaf disease A small batch size witha medium number of iterations is quite effective in thiswork Moreover a larger batch size and number of iterationsincreases the training duration We have not tried higher orlower values for the attempted parameters since differentclassification task may have various suitable parameters andit is hard to give a certain rule in setting hyperparameters

43 Experiments on Full Training and Fine-Tuning of ResNetThis section is designed for exploring the performance ofCNN by changing the structure of the models In practicala deep CNN always owns a large size which means a largenumber of parameters Thus full training of a deep CNNrequires extensive computational resources and is time-consuming In addition full training of a deep CNN mayled to overfitting when the training data is limited So wecompared the performance of the pretrained CNN throughfull training and fine-tuning their structures

We changed the structure of ResNet and combination ofthe best parameters from the front experiments was utilized

Advances in Multimedia 7

Table 4 Classification accuracies with different parameters during fine-tuning of the ResNetThe numbers inside parenthesis indicate batchsize and number of iterations

Networks label1 label2 label3 label4 label5 label6 label7 label8 label9 overallResNet (162496) 9091 8889 100 100 9688 100 9028 8820 9755 9498ResNet (164992) 9818 9877 100 9862 9688 100 9653 8882 9877 9719ResNet (169984) 9818 9753 100 9862 9688 100 9722 8696 9877 9694ResNet (322496) 9727 9506 100 9793 9688 100 9514 8634 9939 9634ResNet (324992) 9727 9506 100 9724 9688 100 9653 8696 9939 9651ResNet (329984) 9636 9630 100 9931 9688 100 9444 8820 9877 9660ResNet (642496) 9364 9383 100 9724 9688 100 9444 8758 9939 9592ResNet (644992) 9455 9529 100 9655 9583 100 9583 8696 9939 9583ResNet (649984) 9545 9383 100 9793 9688 100 9444 8758 9939 9617

ResNet50 has 177 layers if the layers for each building blockand connection are calculated In this experiment the lastthree layers of ResNet were modified to a fully connectedlayer (denoted as ldquofcrdquo) a softmax layer and a classificationlayer and the fully connected layer owns 9 neurons Thestructure was changed by freezing the weights of a certainnumber of layers in the network by setting the learning ratein those layers to zero During training the parameters ofthe frozen layers are not updated Full training and fine-tuning are defined by the number of training layers ie fulltraining (1-ldquofcrdquo) fine-tuning (37-ldquofcrdquo 79-ldquofcrdquo 111-ldquofcrdquo 141-ldquofcrdquo 163-ldquofcrdquo) The accuracy and training time of differentnetwork structure are presented in Table 5 At first the batchsize and 4992 iterations were combined the initial learningrate was set to 0001 and dropped by a factor of 01 every2496 iterations In order to get more convincing conclusionsResNet (16 9984) which gets the second place in Table 4 wasalso used to execute the experiments

In Table 5 the accuracy and training time of differentnetwork structures are presented In two cases ie the 4992iterations and 9984 iterations of ResNet the accuracy of themodel from the 37 layer fine-tuning structure are higherthan that of the full training model In the case where thenumber of iterations is 4992 the accuracy of the model fromthe 79 layer fine-tuning structure is equal to that of the fulltraining modelThe final column of the Table 5 represents thetraining time of the corresponding network and it is clearthat the training time of the fine-tuning models is greatlylowered than the full training model Because the gradientsof the frozen layers do not need to be computed freezingthe weights of initial layers can speed up network trainingWe observe that the moderate fine-tuning models (37-ldquofcrdquo79-ldquofcrdquo 111-ldquofcrdquo) always led to a performance superior orapproximately equal to the full training models Thus wesuggest that for practical application the moderate fine-tuning models may be a good choice Especially for theresearcher who holds massive data the fine-tuning modelsmay achieve good performance while saving computationalresources and time

Moreover the features of the final fully connected layerof ResNet (16 4992 37-ldquofcrdquo) were examined by utilizingthe t-distributed Stochastic Neighbour Embedding (t-SNE)algorithm (see Figure 7) [48] 1176 test images were used to

label1 label2 label3 label4 label5

label6 label7 label8 label9

Figure 7 Two-dimensional scatter plot of high-dimensional fea-tures generated with t-SNE

extract the features In Figure 7 different colors representdifferent labels the corresponding disease categories of thelabels were listed in Table 1 As shown in Figure 7 9 differentcolor points are clearly separated which indicates that thefeatures learned from the ResNet with the optimal structurecan be used to classify the tomato leaf disease precisely

5 Conclusion

This paper concentrates on identifying tomato leaf diseaseusing deep convolutional neural networks by transfer learn-ing The utilized networks are based on the pretrained deeplearning models of AlexNet GoogLeNet and ResNet Firstwe compared the relative performance of these networksby using SGD and Adam optimization method revealingthat the ResNet with SGD optimization method obtainsthe highest result with the best accuracy 9651 Thenthe performance evaluation of batch size and number of

8 Advances in Multimedia

Table 5 Accuracies and training time in different network structuresThe values inside parenthesis denote batch size number of iterationsand training layers

Network topology Accuracy Time (minsec)ResNet (16 4992 1-ldquofcrdquo) 9643 59min 30secResNet (16 4992 37-ldquofcrdquo) 9728 44min 13secResNet (16 4992 79-ldquofcrdquo) 9643 37min 27secResNet (16 4992 111-ldquofcrdquo) 9575 30min 6secResNet (16 4992 141-ldquofcrdquo) 9532 24min 15secResNet (16 4992 163-ldquofcrdquo) 9269 19min 31secResNet (16 9984 1-ldquofcrdquo) 9694 118min 32secResNet (16 9984 37-ldquofcrdquo) 9702 92min 53secResNet (16 9984 79-ldquofcrdquo) 9677 72min 23secResNet (16 9984 111-ldquofcrdquo) 9626 58min 40secResNet (16 9984 141-ldquofcrdquo) 9575 47min 22secResNet (16 9984 163-ldquofcrdquo) 9396 39min 32sec

iterations affecting the transfer learning of the ResNet wasconducted A small batch size of 16 combining a moderatenumber of iterations of 4992 is the optimal choice in thiswork Our findings suggest that for a particular task neitherlarge batch size nor large number of iterations may improvethe accuracy of the target modelThe setting of batch size andnumber of iterations depends on your data set and the utilizednetwork Next the best combined model was used to fine-tune the structure Fine-tuning ResNet layers from 37 to ldquofcrdquoobtained the highest accuracy 9728 in identifying tomatoleaf disease Based on the amount of available data layer-wisefine-tuning may provide a practical way to achieve the bestperformance of the application at hand We believe that theresults obtained in this work will bring some inspiration toother similar visual recognition problems and the practicalstudy of this work can be easily extended to other plant leafdisease identification problems

Data Availability

Thetomato leaf data supporting this work are frompreviouslyreported studies which have been cited The processed dataare available from the corresponding author request

Conflicts of Interest

The authors declare that there are no conflicts of interestregarding the publication of this paper

Acknowledgments

This study was supported by the National Science and tech-nology support program (2014BAD12B01-1-3) PublicWelfareIndustry (Agriculture) Research Projects Level-2 (201503116-04-06) Postdoctoral Foundation of Heilongjiang Province(LBHZ15020) Harbin Applied Technology Research andDevelopment Program (2017RAQXJ096) and EconomicDecision Making and Early Warning of Soybean Industryin Technology Collaborative Innovation System of SoybeanIndustry in Heilongjiang Province (20170401)

References

[1] RChaerani andR EVoorrips ldquoTomato early blight (Alternariasolani) The pathogen genetics and breeding for resistancerdquoJournal of General Plant Pathology vol 72 no 6 pp 335ndash3472006

[2] A M Dickey L S Osborne and C L Mckenzie ldquoPapaya(Carica papaya Brassicales Caricaceae) is not a host plant oftomato yellow leaf curl virus (TYLCV family Geminiviridaegenus Begomovirus)rdquo Florida Entomologist vol 95 no 1 pp211ndash213 2012

[3] G Wei L Baoju S Yanxia and X Xuewen ldquoStudies on path-ogenicity differentiation of corynespora cassiicola isolatesagainst cucumber tomato and eggplantrdquo Acta HorticulturaeSinica vol 38 no 3 pp 465ndash470 2011

[4] P Lindhout W Korta M Cislik I Vos and T GerlaghldquoFurther identification of races of Cladosporium fulvum (Fulviafulva) on tomato originating from the Netherlands France andPolandrdquo Netherlands Journal of Plant Pathology vol 95 no 3pp 143ndash148 1989

[5] K Kubota S Tsuda A Tamai and T Meshi ldquoTomato mosaicvirus replication protein suppresses virus-targeted posttran-scriptional gene silencingrdquo Journal of Virology vol 77 no 20pp 11016ndash11026 2003

[6] M Tian B Benedetti and S Kamoun ldquoA second Kazal-like protease inhibitor from Phytophthora infestans inhibitsand interacts with the apoplastic pathogenesis-related proteaseP69B of tomatordquo Plant Physiology vol 138 no 3 pp 1785ndash17932005

[7] L E Blum ldquoReduction of incidence and severity of Septorialycopersici leaf spot of tomatowith bacteria and yeastsrdquoCienciaRural vol 30 no 5 pp 761ndash765 2000

[8] E A Chatzivasileiadis and M W Sabelis ldquoToxicity of methylketones from tomato trichomes to Tetranychus urticae KochrdquoExperimental and Applied Acarology vol 21 no 6-7 pp 473ndash484 1997

[9] M Anthimopoulos S Christodoulidis L Ebner A Christeand S Mougiakakou ldquoLung Pattern Classification for Inter-stitial Lung Diseases Using a Deep Convolutional NeuralNetworkrdquo IEEE Transactions on Medical Imaging vol 35 no5 pp 1207ndash1216 2016

[10] Y Tang and X Wu ldquoText-independent writer identification viaCNN features and joint Bayesianrdquo in Proceedings of the 15th

Advances in Multimedia 9

International Conference on Frontiers in Handwriting Recogni-tion ICFHR 2016 pp 566ndash571 Shenzhen China October 2016

[11] Y Tang and XWu ldquoSaliency Detection via Combining Region-Level and Pixel-Level Predictions with CNNsrdquo in Proceedings ofthe EuropeanConference onComputer Vision (ECCV) pp 1608ndash05186 2016

[12] Y Tang and X Wu ldquoSalient object detection with chainedmulti-scale fully convolutional networkrdquo ACM Multimedia(ACMMM) pp 618ndash626 2017

[13] Y Tang and X Wu ldquoScene text detection and segmentationbased on cascaded convolution neural networksrdquo IEEE Trans-actions on Image Processing vol 26 no 3 pp 1509ndash1520 2017

[14] Y Tang and X Wu ldquoScene Text Detection using Superpixelbased Stroke Feature Transform and Deep Learning basedRegion Classificationrdquo IEEE Transactions on Multimedia vol20 no 9 pp 2276ndash2288 2018

[15] Y Yao XWu Z Lei S Shan andW Zuo ldquoJoint Representationand Truncated Inference Learning for Correlation Filter basedTrackingrdquo in Proceedings of the European Conference on Com-puter Vision (ECCV) pp 1ndash14 2018

[16] L Zhang F Yang Y Daniel Zhang and Y J Zhu ldquoRoadcrack detection using deep convolutional neural networkrdquoin Proceedings of the 23rd IEEE International Conference onImage Processing ICIP 2016 pp 3708ndash3712 Phoenix AZ USASeptember 2016

[17] D Xie L Zhang andL Bai ldquoDeep learning in visual computingand signal processingrdquo Applied Computational Intelligence andSoft Computing vol 2017 Article ID 1320780 13 pages 2017

[18] Z Zhou J Shin L Zhang S Gurudu M Gotway and J LiangldquoFine-tuning convolutional neural networks for biomedicalimage analysis Actively and incrementallyrdquo inProceedings of the30th IEEE Conference on Computer Vision and Pattern Recog-nition CVPR 2017 pp 4761ndash4772 USA July 2017

[19] Z Liu P Luo X Wang and X Tang ldquoDeep learning face attri-butes in the wildrdquo in Proceedings of the 15th IEEE InternationalConference on Computer Vision ICCV 2015 pp 3730ndash3738Santiago Chile 2015

[20] W Ouyang and X Wang ldquoJoint deep learning for pedestriandetectionrdquo in Proceedings of the 14th IEEE International Con-ference on Computer Vision (ICCV rsquo13) pp 2056ndash2063 SydneyAustralia December 2013

[21] A Krizhevsky I Sutskever andG EHinton ldquoImagenet classifi-cation with deep convolutional neural networksrdquo in Proceedingsof the 26th Annual Conference on Neural Information ProcessingSystems (NIPS rsquo12) pp 1097ndash1105 Lake Tahoe Nev USADecember 2012

[22] C Szegedy W Liu Y Jia et al ldquoGoing deeper with convolu-tionsrdquo inProceedings of the IEEEConference onComputer Visionand Pattern Recognition (CVPR rsquo15) pp 1ndash9 IEEE BostonMass USA June 2015

[23] K Simonyan and A Zisserman ldquoVery deep convolutional net-works for large-scale image recognitionrdquo httpsarxivorgabs14091556 2015

[24] C Szegedy V Vanhoucke S Ioffe J Shlens and Z WojnaldquoRethinking the inception architecture for computer visionrdquohttpsarxivorgabs151200567 2015

[25] C Szegedy S Ioffe V Vanhoucke and A Alemi ldquoInception-v4 inception-ResNet and the impact of residual connections onlearningrdquo httpsarxivorgabs160207261 2016

[26] K He X Zhang S Ren and J Sun ldquoDeep residual learning forimage recognitionrdquo in Proceedings of the 2016 IEEE Conference

on Computer Vision and Pattern Recognition CVPR 2016 pp770ndash778 July 2016

[27] G Huang Z Liu K Q Weinberger and L van der MaatenldquoDensely connected convolutional networksrdquo in Proceedingsof the 2017 IEEE Conference on Computer Vision and PatternRecognition (CVPR) Honolulu HI USA 2016

[28] S J Pan and Q Yang ldquoA survey on transfer learningrdquo IEEETransactions on Knowledge and Data Engineering vol 22 no10 pp 1345ndash1359 2010

[29] M Mehdipour Ghazi B Yanikoglu and E Aptoula ldquoPlantidentification using deep neural networks via optimization oftransfer learning parametersrdquo Neurocomputing vol 235 pp228ndash235 2017

[30] S Ruder ldquoAn overview of gradient descent optimization algo-rithmsrdquo httpsarxivorgabs160904747 2017

[31] D P Kingma and J L Ba ldquoAdam a method for stochasticoptimizationrdquo httpsarxivorgabs14126980 2017

[32] S S Sannakki V S Rajpurohit V B Nargund R Arun Kumarand S Prema Yallur ldquoLeaf disease grading by machine visionand fuzzy logicrdquo International Journal of Computer Technologyand Applications vol 2 no 5 pp 1709ndash1716 2011

[33] D Samanta P P Chaudhury and A Ghosh ldquoScab diseasesdetection of potato using image processingrdquo InternationalJournal of Computer Trends and Technology vol 3 pp 109ndash1132012

[34] P J Herrera J Dorado and A Ribeiro ldquoA novel approach forweed type classification based on shape descriptors and a fuzzydecision-making methodrdquo Sensors vol 14 no 8 pp 15304ndash15324 2014

[35] B Cheng and E T Matson ldquoA feature-based machine learningagent for automatic rice andweed discriminationrdquo InternationalConference onArtificial Intelligence and SoftComputing pp 517ndash527 2015

[36] S Sankaran and R Ehsani ldquoComparison of visible-near infra-red and mid-infrared spectroscopy for classification of Huang-longbing and citrus canker infected leavesrdquo Agricultural Engi-neering International CIGR Journal vol 15 no 3 pp 75ndash792013

[37] X Cheng Y Zhang Y Chen Y Wu and Y Yue ldquoPest iden-tification via deep residual learning in complex backgroundrdquoComputers and Electronics in Agriculture vol 141 pp 351ndash3562017

[38] A dos Santos Ferreira D Matte Freitas G Goncalves daSilva H Pistori and M Theophilo Folhes ldquoWeed detection insoybean crops using ConvNetsrdquo Computers and Electronics inAgriculture vol 143 pp 314ndash324 2017

[39] S Sladojevic M Arsenovic A Anderla D Culibrk andD Stefanovic ldquoDeep Neural Networks Based Recognition ofPlant Diseases by Leaf Image Classificationrdquo ComputationalIntelligence and Neuroscience vol 2016 Article ID 3289801 11pages 2016

[40] S P Mohanty D P Hughes and M Salathe ldquoUsing deep learn-ing for image-based plant disease detectionrdquo Frontiers in PlantScience vol 7 article no 1419 2016

[41] I Sa Z Ge F Dayoub B Upcroft T Perez and C McCoolldquoDeepfruits A fruit detection system using deep neural net-worksrdquo Sensors vol 16 article no 1222 no 8 2016

[42] V Nair and G E Hinton ldquoRectified linear units improveRestricted Boltzmann machinesrdquo in Proceedings of the 27thInternational Conference on Machine Learning (ICML rsquo10) pp807ndash814 Haifa Israel June 2010

10 Advances in Multimedia

[43] D P Hughes and M Salathe ldquoAn open access repository ofimages on plant health to enable the development of mobiledisease diagnosticsrdquo httpsarxivorgabs151108060 2016

[44] D Ciresan U Meier J Masci and J Schmidhuber ldquoA com-mittee of neural networks for traffic sign classificationrdquo inProceedings of the 2011 International Joint Conference on NeuralNetworks (IJCNN 2011 - San Jose) San Jose CA USA July 2011

[45] P Pawara E Okafor O Surinta L Schomaker andMWieringldquoComparing Local Descriptors and Bags of Visual Words toDeep Convolutional Neural Networks for Plant Recognitionrdquoin Proceedings of the 6th International Conference on PatternRecognition Applications and Methods pp 479ndash486 PortoPortugal Feburary 2017

[46] M Lin ldquoNetwork inNnetworkrdquo httpsarxivorgabs131244002014

[47] K He X Zhang S Ren and J Sun ldquoIdentity mappings in deepresidual networksrdquo in Proceedings of the European Conferenceon Computer Vision pp 630ndash645 2016

[48] L van der Maaten and G Hinton ldquoVisualizing data using t-SNErdquo Journal of Machine Learning Research vol 9 pp 2579ndash2625 2008

International Journal of

AerospaceEngineeringHindawiwwwhindawicom Volume 2018

RoboticsJournal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Active and Passive Electronic Components

VLSI Design

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Shock and Vibration

Hindawiwwwhindawicom Volume 2018

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawiwwwhindawicom

Volume 2018

Hindawi Publishing Corporation httpwwwhindawicom Volume 2013Hindawiwwwhindawicom

The Scientific World Journal

Volume 2018

Control Scienceand Engineering

Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom

Journal ofEngineeringVolume 2018

SensorsJournal of

Hindawiwwwhindawicom Volume 2018

International Journal of

RotatingMachinery

Hindawiwwwhindawicom Volume 2018

Modelling ampSimulationin EngineeringHindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Navigation and Observation

International Journal of

Hindawi

wwwhindawicom Volume 2018

Advances in

Multimedia

Submit your manuscripts atwwwhindawicom

Page 3: ResearchArticle Can Deep Learning Identify Tomato Leaf

Advances in Multimedia 3

Figure 2 Raw tomato leaf images

The typical CNN process can be represented with fol-lowing formulas Firstly send the training samples (ietraining tomato leaf images) to the classifier (ie AlexNetGoogLeNet and ResNet) Then convolution operation iscarried out that is a number of filters slide over the featuremap of the previous layer and the weight matrices do dotproduct

119872119897119895 = 119891(sum119894isin119873119895

119872119897minus1119894 lowast 119908119897119895 + 119887119897119895) (1)

where 119891(sdot) is activation function typically a Rectifier LinearUnit (ReLU) [42] function

119891 (119909) = max (119909 0) (2)

119873119895 is the number of kernels of the certain layer119872119897minus1119894 repre-sents the feature map of the previous layer 119908119897119895 is the weightmatrix and 119887119897119895 is the bias term

Max-pooling or average pooling is conducted after theconvolution operation Furthermore the learned features aresent to the fully connected layer The softmax regressionalways follows the final fully connected layer an input 119909 willget the probability of belonging to class 119894

119901 (119910 = 119894 | 119909 120579) = 119890120579119879119894 119909sum119896119895=1 119890120579119879119894 119909 (3)

where 119910 is the response variable (ie predict label) 119896 is thenumber of categories and 120579 is the parameters of our model

31 Raw Dataset The raw tomato leaf dataset utilized inthis work comes from an open access repository of imageswhich focus on plant health [43] Health and other 8 diseasescategories are included (see Table 1 Figure 2) ie early blight(pathogen Alternaria solani) [1] yellow leaf curl disease(pathogen Tomato Yellow Leaf Curl Virus (Tylcv) FamilyGeminiviridae Genus Begomovirus) [2] corynespora leafspot disease (pathogen Corynespora cassiicola) [3] leafmolddisease (pathogen Fulvia fulva) [4] virus disease (pathogenTomato Mosaic Virus) [5] late blight (pathogen Phytoph-thora Infestans)[6] septoria leaf spot (pathogen Septorialycopersici) [7] and two-spotted spider mite (pathogenTetranychus urticae) [8] The total dataset is 5550

32 Data Augmentation Deep convolutional neural net-works contain millions of parameters thus massive amountsof data is required Otherwise the deep neural network maybe overfitting or not robust The most common method toreduce overfitting on image dataset is to enlarge the datasetmanually and conduct label-preserving transformations [2144]

In this work at first the raw image dataset was dividedinto 80 training samples and 20 testing samples and thenthe data augmentation procedure was conducted (1) flipping

4 Advances in Multimedia

Table 1 The raw tomato leaf dataset

Label Category Number Leaf symptoms Illustration

1 Corynesporaleaf spot disease 547

Small brown spotsappear leaf spots have

yellow halo

See Figure 1 first rowNo1-No5

2 Early blight 405

Black or brown spotsappear leaf spots oftenhave yellow or green

concentric ring pattern

See Figure 1 first rowNo6-No10

4 Late blight 726

Water-soaked areaappears and rapidlyenlarges to formpurple-brown

oily-appearing blotches

See Figure 1 second rowNo1-No5

5 Leaf molddisease 480 Irregular yellow or green

area appearsSee Figure 1 second row

No6-No10

6 Septoria leafspot 734

Round spots marginalbrown chlorotic yellow

appear

See Figure 1 third rowNo1-No5

7 Two-spottedspider mite 720 Show white or yellow

spots blade back nettingSee Figure 1 third row

No6-No10

8 Virus disease 481 Develop yellow or greenslightly shrinking

See Figure 1 forth rowNo1-No5

9 Yellow leaf curldisease 814

Develop small and curlupward crumpling andmarginal yellowingbushy appearance

See Figure 1 forth rowNo6-No10

3 Health 643 See Figure 1 fifth rowTotal 5550

the image from left to right (2) flipping the image from topto bottom (3) flipping the image diagonally (4) adjusting thebrightness of image setting the max delta to 04 (5) adjustingthe contrast of image setting the ratio from 02 to 15 (6)adjusting the hue of image setting the max delta to 05 (7)adjusting the saturation of image setting the ratio from 02 to15 (8) rotating the image by 90∘ and 270∘ respectively Thefinal dataset is shown in Table 2 and the label in the first rowrepresents the disease categories which are given in Table 1

33 Deep Learning Models

331 AlexNet AlexNet is the winner of ImageNet Large-Scale Visual Recognition Challenge (ILSVRC) 2012 a deepconvolutional neural network which has 60 million param-eters and 650000 neurons [21] The architecture of AlexNetutilized in this paper is displayed in Figure 3 The AlexNetarchitecture consists of five convolutional layers (ie conv1conv2 and so on) some of which are followed by max-pooling layers (ie pool1 pool2 and pool5) three fullyconnected layers (ie fc6 fc7 and fc8) and a liner layer withsoftmax activation in output In order to reduce overfittingin the fully connected layers a regularization method calledldquodropoutrdquo is used (ie drop6 drop7) [21] The ReLU activa-tion function is applied to each of the first seven layers (ierelu1 relu2 and so on) [45] In Figure 3 the notation119898times119898times119899in each convolutional layer represents the size of the feature

map for each layer 4096 represents the number of neurons ofthe first two fully connected layers The number of neuronsof the final fully connected layer was modified to 9 sincethe classification problem in this work has 9 categories Inaddition the size of input imagesmust be shaped to 227times227which meets the input pixel size requirement of AlexNet

332 GoogLeNet GoogLeNet is an inception architecture[22] which is the winner of ILSVRC 2014 and owns roughly68 million parameters The architecture of GoogLeNet ispresented in Figure 4 The inception module is inspired bythe network in network [46] and uses a parallel combinationof 1 times 1 3 times 3 and 5 times 5 convolutional layer along with3 times 3 max-pooling layer [45] the 1 times 1 convolutional layerbefore 3 times 3 and 5 times 5 convolutional layer reduces the spatialdimension and limits the size of GoogLeNet The wholearchitecture of GoogLeNet is stacked by inception moduleon top of each other (See Figure 4) which has nine inceptionmodules two convolutional layers four max-pooling layersone average pooling layer one fully connected layer and alinear layer with softmax function in the output GoogLeNetuses dropout regularization in the fully connected layerand applies the ReLU activation function in all of theconvolutional layers [29] In this work the last three layersof GoogLeNet were replaced by a fully connected layer asoftmax layer and a classification layer the fully connectedlayer was modified to 9 neurons which is equal to the

Advances in Multimedia 5

Table 2 The final tomato leaf dataset

Labels Label1 Label2 Label3 Label4 Label5 Label6 Label7 Label8 Label9 TotalTraining set 3933 2916 4626 5229 3456 5283 5184 3465 5859 39951Testing set 110 81 129 145 96 147 144 161 163 1176

Figure 3 The architecture of AlexNet in this work

categories in the tomato leaf disease identification problemThe size requested of input image of GoogLeNet is 224times 224333 ResNet The deep residual learning framework is pro-posed for addressing the degradation problem ResNet con-sists of many stacked residual units which won the first placein ILSVRC 2015 and COCO 2015 classification challenge witherror rate of 357 [26] Each unit can be expressed in thefollowing formulas [47]

119910119897 = ℎ (119909119897) + 119865 (119909119897119882119897) (4)

119909119897+1 = 119891 (119910119897) (5)

where 119909119897 and 119909119897+1 are input and output of the l-th unit and 119865is a residual function In [26]ℎ(119909119897) = 119909119897 is an identity mappingand 119891 is a ReLU function [42] A ldquobottleneckrdquo building blockis designed for ResNet (See Figure 5) and comprises two 1times1convolutions with a 3 times 3 convolution in between and adirect skip connection bypassing input and output The 1 times 1layers are responsible for changing in dimensions ResNetmodel has three types of layers with 50 101 and 152 Forsaving computing resources and training time we choose theResNet50 which also has high performance In this work atfirst the last three layers of ResNet were modified by a fullyconnected layer a softmax layer and a classification layerthe fully connected layer was replaced to 9 neurons which isequal to the categories of the tomato leaf disease We changedthe structure of ResNet subsequently The size of input imageof ResNet should satisfy 224 times 2244 Experiments and Results

In this section we reveal the experiments and discuss theexperimental results All the experiments were implementedin Matlab under Windows 10 using the GPU NVIDIAGTX1050 with 4G video memory or NVIDIA GTX1080Tiwith 11G video memory In this paper overall accuracy wasregarded as the evaluation metric in every experiment on

tomato leaf disease detection which means the percentage ofsamples that are correctly classified

accuracy = true positive + true negativepositive + negative (6)

where ldquotrue positiverdquo is the number of instances that are pos-itive and classified as positive ldquotrue negativerdquo is the numberof instances that are negative and classified as negative andthe denominator represents the total number of samples Inaddition the training time was regarded as an additionalperformance metric of the network structure experiment

41 Experiments on Optimization Methods The first exper-iment is designed for seeking the suitable optimizationmethod between SGD [30] and Adam [30 31] in identifyingtomato leaf diseases combining with the pretrained net-work AlexNet GoogLeNet and ResNet respectively In thisexperiment the hyperparameters were set as follows for eachnetwork the batch size was set to 32 the initial learning ratewas set to 0001 and dropped by a factor of 05 every 2 epochsand the max epoch was set to 5 ie the number of iterationsis 6240 So far as SGD optimization method the momentumwas set to 09 For Adam the gradient decay rate 1205731 was set to09 the squared gradient decay rate 1205732 was set to 0999 andthe denominator offset 120576 was set to 10minus8 [31] The accuracyof different networks is displayed in Table 3 In addition wechoose the better results in each deep neural network to showthe training loss against number of iterations during the fine-tuning process (See Figure 6) The words inside parenthesisindicate the corresponding optimization method

In Table 3 the ResNet with SGD optimization methodgets the highest test accuracy 9651 In identifying tomatoleaf diseases the performance of Adam optimization methodis inferior to the SGD optimization method especially incombining with AlexNet In the following paper AlexNet(SGD) GoogLeNet (SGD) and ResNet (SGD) are referred toas AlexNet GoogLeNet and ResNet respectively

As it can be seen in Figure 6 the training loss of ResNetdrops rapidly in the earlier iterations and tends to stable after3000 iterations Consistent with Table 3 the performance of

6 Advances in Multimedia

Inputimage

Conv1 Conv2

MaxPool MaxPool

MaxPool

MaxPool

Inception 3a

Inception 3b

Inception 4a

Inception 4b

Inception 4c

Inception 4d

Inception 4e

Inception 5a

Inception 5b

AvgPool

Dropout40

FC

somax

output

Figure 4 The architecture of GoogLeNet [22 45]

1times1 conv 64

3times3 conv 64

1times1 conv 256

relu

relu

relu

+

Figure 5 ResNet bottleneck residual building block [26]

Table 3 Model recognition accuracy

Model AccuracyAlexNet (SGD) 9583AlexNet (Adam) 1386GoogLeNet (SGD) 9566GoogLeNet (Adam) 9406ResNet (SGD) 9651ResNet (Adam) 9439

AlexNet and GoogLeNet is similar and both inferior to theResNet

42 Experiments on Batch Size and Number of IterationsFrom the experiment on optimization methods the ResNetobtains the highest classification accuracy Next we evaluatedthe effects of batch size and the number of iterations on theperformance of the ResNet The batch size was set to 16 32and 64 respectively Meanwhile the number of iterations wasset to 2496 4992 and 9984 The classification accuracy ofdifferent training scenarios is given in Table 4 At the sametime the classification accuracy of each labels representativeleaf disease category (See Table 1) is given In this experimentthe initial learning rate was set to 0001 and dropped by afactor of 05 every 2496 iterations

In Table 4 the best overall classification accuracy 9719is got by the ResNet combining with batch size 16 and

Trai

ning

Los

s

Number of Iterations

AlexNetGoogLeNetResNet

3

25

2

15

1

05

00 1000 2000 3000 4000 5000 6000

Figure 6 The training loss during the fine-tuning process

iterations 4992 As shown in Table 4 whether increasingthe number of iterations or batch size the performance ofcorresponding models has not been improved significantlyin identifying tomato leaf disease A small batch size witha medium number of iterations is quite effective in thiswork Moreover a larger batch size and number of iterationsincreases the training duration We have not tried higher orlower values for the attempted parameters since differentclassification task may have various suitable parameters andit is hard to give a certain rule in setting hyperparameters

43 Experiments on Full Training and Fine-Tuning of ResNetThis section is designed for exploring the performance ofCNN by changing the structure of the models In practicala deep CNN always owns a large size which means a largenumber of parameters Thus full training of a deep CNNrequires extensive computational resources and is time-consuming In addition full training of a deep CNN mayled to overfitting when the training data is limited So wecompared the performance of the pretrained CNN throughfull training and fine-tuning their structures

We changed the structure of ResNet and combination ofthe best parameters from the front experiments was utilized

Advances in Multimedia 7

Table 4 Classification accuracies with different parameters during fine-tuning of the ResNetThe numbers inside parenthesis indicate batchsize and number of iterations

Networks label1 label2 label3 label4 label5 label6 label7 label8 label9 overallResNet (162496) 9091 8889 100 100 9688 100 9028 8820 9755 9498ResNet (164992) 9818 9877 100 9862 9688 100 9653 8882 9877 9719ResNet (169984) 9818 9753 100 9862 9688 100 9722 8696 9877 9694ResNet (322496) 9727 9506 100 9793 9688 100 9514 8634 9939 9634ResNet (324992) 9727 9506 100 9724 9688 100 9653 8696 9939 9651ResNet (329984) 9636 9630 100 9931 9688 100 9444 8820 9877 9660ResNet (642496) 9364 9383 100 9724 9688 100 9444 8758 9939 9592ResNet (644992) 9455 9529 100 9655 9583 100 9583 8696 9939 9583ResNet (649984) 9545 9383 100 9793 9688 100 9444 8758 9939 9617

ResNet50 has 177 layers if the layers for each building blockand connection are calculated In this experiment the lastthree layers of ResNet were modified to a fully connectedlayer (denoted as ldquofcrdquo) a softmax layer and a classificationlayer and the fully connected layer owns 9 neurons Thestructure was changed by freezing the weights of a certainnumber of layers in the network by setting the learning ratein those layers to zero During training the parameters ofthe frozen layers are not updated Full training and fine-tuning are defined by the number of training layers ie fulltraining (1-ldquofcrdquo) fine-tuning (37-ldquofcrdquo 79-ldquofcrdquo 111-ldquofcrdquo 141-ldquofcrdquo 163-ldquofcrdquo) The accuracy and training time of differentnetwork structure are presented in Table 5 At first the batchsize and 4992 iterations were combined the initial learningrate was set to 0001 and dropped by a factor of 01 every2496 iterations In order to get more convincing conclusionsResNet (16 9984) which gets the second place in Table 4 wasalso used to execute the experiments

In Table 5 the accuracy and training time of differentnetwork structures are presented In two cases ie the 4992iterations and 9984 iterations of ResNet the accuracy of themodel from the 37 layer fine-tuning structure are higherthan that of the full training model In the case where thenumber of iterations is 4992 the accuracy of the model fromthe 79 layer fine-tuning structure is equal to that of the fulltraining modelThe final column of the Table 5 represents thetraining time of the corresponding network and it is clearthat the training time of the fine-tuning models is greatlylowered than the full training model Because the gradientsof the frozen layers do not need to be computed freezingthe weights of initial layers can speed up network trainingWe observe that the moderate fine-tuning models (37-ldquofcrdquo79-ldquofcrdquo 111-ldquofcrdquo) always led to a performance superior orapproximately equal to the full training models Thus wesuggest that for practical application the moderate fine-tuning models may be a good choice Especially for theresearcher who holds massive data the fine-tuning modelsmay achieve good performance while saving computationalresources and time

Moreover the features of the final fully connected layerof ResNet (16 4992 37-ldquofcrdquo) were examined by utilizingthe t-distributed Stochastic Neighbour Embedding (t-SNE)algorithm (see Figure 7) [48] 1176 test images were used to

label1 label2 label3 label4 label5

label6 label7 label8 label9

Figure 7 Two-dimensional scatter plot of high-dimensional fea-tures generated with t-SNE

extract the features In Figure 7 different colors representdifferent labels the corresponding disease categories of thelabels were listed in Table 1 As shown in Figure 7 9 differentcolor points are clearly separated which indicates that thefeatures learned from the ResNet with the optimal structurecan be used to classify the tomato leaf disease precisely

5 Conclusion

This paper concentrates on identifying tomato leaf diseaseusing deep convolutional neural networks by transfer learn-ing The utilized networks are based on the pretrained deeplearning models of AlexNet GoogLeNet and ResNet Firstwe compared the relative performance of these networksby using SGD and Adam optimization method revealingthat the ResNet with SGD optimization method obtainsthe highest result with the best accuracy 9651 Thenthe performance evaluation of batch size and number of

8 Advances in Multimedia

Table 5 Accuracies and training time in different network structuresThe values inside parenthesis denote batch size number of iterationsand training layers

Network topology Accuracy Time (minsec)ResNet (16 4992 1-ldquofcrdquo) 9643 59min 30secResNet (16 4992 37-ldquofcrdquo) 9728 44min 13secResNet (16 4992 79-ldquofcrdquo) 9643 37min 27secResNet (16 4992 111-ldquofcrdquo) 9575 30min 6secResNet (16 4992 141-ldquofcrdquo) 9532 24min 15secResNet (16 4992 163-ldquofcrdquo) 9269 19min 31secResNet (16 9984 1-ldquofcrdquo) 9694 118min 32secResNet (16 9984 37-ldquofcrdquo) 9702 92min 53secResNet (16 9984 79-ldquofcrdquo) 9677 72min 23secResNet (16 9984 111-ldquofcrdquo) 9626 58min 40secResNet (16 9984 141-ldquofcrdquo) 9575 47min 22secResNet (16 9984 163-ldquofcrdquo) 9396 39min 32sec

iterations affecting the transfer learning of the ResNet wasconducted A small batch size of 16 combining a moderatenumber of iterations of 4992 is the optimal choice in thiswork Our findings suggest that for a particular task neitherlarge batch size nor large number of iterations may improvethe accuracy of the target modelThe setting of batch size andnumber of iterations depends on your data set and the utilizednetwork Next the best combined model was used to fine-tune the structure Fine-tuning ResNet layers from 37 to ldquofcrdquoobtained the highest accuracy 9728 in identifying tomatoleaf disease Based on the amount of available data layer-wisefine-tuning may provide a practical way to achieve the bestperformance of the application at hand We believe that theresults obtained in this work will bring some inspiration toother similar visual recognition problems and the practicalstudy of this work can be easily extended to other plant leafdisease identification problems

Data Availability

Thetomato leaf data supporting this work are frompreviouslyreported studies which have been cited The processed dataare available from the corresponding author request

Conflicts of Interest

The authors declare that there are no conflicts of interestregarding the publication of this paper

Acknowledgments

This study was supported by the National Science and tech-nology support program (2014BAD12B01-1-3) PublicWelfareIndustry (Agriculture) Research Projects Level-2 (201503116-04-06) Postdoctoral Foundation of Heilongjiang Province(LBHZ15020) Harbin Applied Technology Research andDevelopment Program (2017RAQXJ096) and EconomicDecision Making and Early Warning of Soybean Industryin Technology Collaborative Innovation System of SoybeanIndustry in Heilongjiang Province (20170401)

References

[1] RChaerani andR EVoorrips ldquoTomato early blight (Alternariasolani) The pathogen genetics and breeding for resistancerdquoJournal of General Plant Pathology vol 72 no 6 pp 335ndash3472006

[2] A M Dickey L S Osborne and C L Mckenzie ldquoPapaya(Carica papaya Brassicales Caricaceae) is not a host plant oftomato yellow leaf curl virus (TYLCV family Geminiviridaegenus Begomovirus)rdquo Florida Entomologist vol 95 no 1 pp211ndash213 2012

[3] G Wei L Baoju S Yanxia and X Xuewen ldquoStudies on path-ogenicity differentiation of corynespora cassiicola isolatesagainst cucumber tomato and eggplantrdquo Acta HorticulturaeSinica vol 38 no 3 pp 465ndash470 2011

[4] P Lindhout W Korta M Cislik I Vos and T GerlaghldquoFurther identification of races of Cladosporium fulvum (Fulviafulva) on tomato originating from the Netherlands France andPolandrdquo Netherlands Journal of Plant Pathology vol 95 no 3pp 143ndash148 1989

[5] K Kubota S Tsuda A Tamai and T Meshi ldquoTomato mosaicvirus replication protein suppresses virus-targeted posttran-scriptional gene silencingrdquo Journal of Virology vol 77 no 20pp 11016ndash11026 2003

[6] M Tian B Benedetti and S Kamoun ldquoA second Kazal-like protease inhibitor from Phytophthora infestans inhibitsand interacts with the apoplastic pathogenesis-related proteaseP69B of tomatordquo Plant Physiology vol 138 no 3 pp 1785ndash17932005

[7] L E Blum ldquoReduction of incidence and severity of Septorialycopersici leaf spot of tomatowith bacteria and yeastsrdquoCienciaRural vol 30 no 5 pp 761ndash765 2000

[8] E A Chatzivasileiadis and M W Sabelis ldquoToxicity of methylketones from tomato trichomes to Tetranychus urticae KochrdquoExperimental and Applied Acarology vol 21 no 6-7 pp 473ndash484 1997

[9] M Anthimopoulos S Christodoulidis L Ebner A Christeand S Mougiakakou ldquoLung Pattern Classification for Inter-stitial Lung Diseases Using a Deep Convolutional NeuralNetworkrdquo IEEE Transactions on Medical Imaging vol 35 no5 pp 1207ndash1216 2016

[10] Y Tang and X Wu ldquoText-independent writer identification viaCNN features and joint Bayesianrdquo in Proceedings of the 15th

Advances in Multimedia 9

International Conference on Frontiers in Handwriting Recogni-tion ICFHR 2016 pp 566ndash571 Shenzhen China October 2016

[11] Y Tang and XWu ldquoSaliency Detection via Combining Region-Level and Pixel-Level Predictions with CNNsrdquo in Proceedings ofthe EuropeanConference onComputer Vision (ECCV) pp 1608ndash05186 2016

[12] Y Tang and X Wu ldquoSalient object detection with chainedmulti-scale fully convolutional networkrdquo ACM Multimedia(ACMMM) pp 618ndash626 2017

[13] Y Tang and X Wu ldquoScene text detection and segmentationbased on cascaded convolution neural networksrdquo IEEE Trans-actions on Image Processing vol 26 no 3 pp 1509ndash1520 2017

[14] Y Tang and X Wu ldquoScene Text Detection using Superpixelbased Stroke Feature Transform and Deep Learning basedRegion Classificationrdquo IEEE Transactions on Multimedia vol20 no 9 pp 2276ndash2288 2018

[15] Y Yao XWu Z Lei S Shan andW Zuo ldquoJoint Representationand Truncated Inference Learning for Correlation Filter basedTrackingrdquo in Proceedings of the European Conference on Com-puter Vision (ECCV) pp 1ndash14 2018

[16] L Zhang F Yang Y Daniel Zhang and Y J Zhu ldquoRoadcrack detection using deep convolutional neural networkrdquoin Proceedings of the 23rd IEEE International Conference onImage Processing ICIP 2016 pp 3708ndash3712 Phoenix AZ USASeptember 2016

[17] D Xie L Zhang andL Bai ldquoDeep learning in visual computingand signal processingrdquo Applied Computational Intelligence andSoft Computing vol 2017 Article ID 1320780 13 pages 2017

[18] Z Zhou J Shin L Zhang S Gurudu M Gotway and J LiangldquoFine-tuning convolutional neural networks for biomedicalimage analysis Actively and incrementallyrdquo inProceedings of the30th IEEE Conference on Computer Vision and Pattern Recog-nition CVPR 2017 pp 4761ndash4772 USA July 2017

[19] Z Liu P Luo X Wang and X Tang ldquoDeep learning face attri-butes in the wildrdquo in Proceedings of the 15th IEEE InternationalConference on Computer Vision ICCV 2015 pp 3730ndash3738Santiago Chile 2015

[20] W Ouyang and X Wang ldquoJoint deep learning for pedestriandetectionrdquo in Proceedings of the 14th IEEE International Con-ference on Computer Vision (ICCV rsquo13) pp 2056ndash2063 SydneyAustralia December 2013

[21] A Krizhevsky I Sutskever andG EHinton ldquoImagenet classifi-cation with deep convolutional neural networksrdquo in Proceedingsof the 26th Annual Conference on Neural Information ProcessingSystems (NIPS rsquo12) pp 1097ndash1105 Lake Tahoe Nev USADecember 2012

[22] C Szegedy W Liu Y Jia et al ldquoGoing deeper with convolu-tionsrdquo inProceedings of the IEEEConference onComputer Visionand Pattern Recognition (CVPR rsquo15) pp 1ndash9 IEEE BostonMass USA June 2015

[23] K Simonyan and A Zisserman ldquoVery deep convolutional net-works for large-scale image recognitionrdquo httpsarxivorgabs14091556 2015

[24] C Szegedy V Vanhoucke S Ioffe J Shlens and Z WojnaldquoRethinking the inception architecture for computer visionrdquohttpsarxivorgabs151200567 2015

[25] C Szegedy S Ioffe V Vanhoucke and A Alemi ldquoInception-v4 inception-ResNet and the impact of residual connections onlearningrdquo httpsarxivorgabs160207261 2016

[26] K He X Zhang S Ren and J Sun ldquoDeep residual learning forimage recognitionrdquo in Proceedings of the 2016 IEEE Conference

on Computer Vision and Pattern Recognition CVPR 2016 pp770ndash778 July 2016

[27] G Huang Z Liu K Q Weinberger and L van der MaatenldquoDensely connected convolutional networksrdquo in Proceedingsof the 2017 IEEE Conference on Computer Vision and PatternRecognition (CVPR) Honolulu HI USA 2016

[28] S J Pan and Q Yang ldquoA survey on transfer learningrdquo IEEETransactions on Knowledge and Data Engineering vol 22 no10 pp 1345ndash1359 2010

[29] M Mehdipour Ghazi B Yanikoglu and E Aptoula ldquoPlantidentification using deep neural networks via optimization oftransfer learning parametersrdquo Neurocomputing vol 235 pp228ndash235 2017

[30] S Ruder ldquoAn overview of gradient descent optimization algo-rithmsrdquo httpsarxivorgabs160904747 2017

[31] D P Kingma and J L Ba ldquoAdam a method for stochasticoptimizationrdquo httpsarxivorgabs14126980 2017

[32] S S Sannakki V S Rajpurohit V B Nargund R Arun Kumarand S Prema Yallur ldquoLeaf disease grading by machine visionand fuzzy logicrdquo International Journal of Computer Technologyand Applications vol 2 no 5 pp 1709ndash1716 2011

[33] D Samanta P P Chaudhury and A Ghosh ldquoScab diseasesdetection of potato using image processingrdquo InternationalJournal of Computer Trends and Technology vol 3 pp 109ndash1132012

[34] P J Herrera J Dorado and A Ribeiro ldquoA novel approach forweed type classification based on shape descriptors and a fuzzydecision-making methodrdquo Sensors vol 14 no 8 pp 15304ndash15324 2014

[35] B Cheng and E T Matson ldquoA feature-based machine learningagent for automatic rice andweed discriminationrdquo InternationalConference onArtificial Intelligence and SoftComputing pp 517ndash527 2015

[36] S Sankaran and R Ehsani ldquoComparison of visible-near infra-red and mid-infrared spectroscopy for classification of Huang-longbing and citrus canker infected leavesrdquo Agricultural Engi-neering International CIGR Journal vol 15 no 3 pp 75ndash792013

[37] X Cheng Y Zhang Y Chen Y Wu and Y Yue ldquoPest iden-tification via deep residual learning in complex backgroundrdquoComputers and Electronics in Agriculture vol 141 pp 351ndash3562017

[38] A dos Santos Ferreira D Matte Freitas G Goncalves daSilva H Pistori and M Theophilo Folhes ldquoWeed detection insoybean crops using ConvNetsrdquo Computers and Electronics inAgriculture vol 143 pp 314ndash324 2017

[39] S Sladojevic M Arsenovic A Anderla D Culibrk andD Stefanovic ldquoDeep Neural Networks Based Recognition ofPlant Diseases by Leaf Image Classificationrdquo ComputationalIntelligence and Neuroscience vol 2016 Article ID 3289801 11pages 2016

[40] S P Mohanty D P Hughes and M Salathe ldquoUsing deep learn-ing for image-based plant disease detectionrdquo Frontiers in PlantScience vol 7 article no 1419 2016

[41] I Sa Z Ge F Dayoub B Upcroft T Perez and C McCoolldquoDeepfruits A fruit detection system using deep neural net-worksrdquo Sensors vol 16 article no 1222 no 8 2016

[42] V Nair and G E Hinton ldquoRectified linear units improveRestricted Boltzmann machinesrdquo in Proceedings of the 27thInternational Conference on Machine Learning (ICML rsquo10) pp807ndash814 Haifa Israel June 2010

10 Advances in Multimedia

[43] D P Hughes and M Salathe ldquoAn open access repository ofimages on plant health to enable the development of mobiledisease diagnosticsrdquo httpsarxivorgabs151108060 2016

[44] D Ciresan U Meier J Masci and J Schmidhuber ldquoA com-mittee of neural networks for traffic sign classificationrdquo inProceedings of the 2011 International Joint Conference on NeuralNetworks (IJCNN 2011 - San Jose) San Jose CA USA July 2011

[45] P Pawara E Okafor O Surinta L Schomaker andMWieringldquoComparing Local Descriptors and Bags of Visual Words toDeep Convolutional Neural Networks for Plant Recognitionrdquoin Proceedings of the 6th International Conference on PatternRecognition Applications and Methods pp 479ndash486 PortoPortugal Feburary 2017

[46] M Lin ldquoNetwork inNnetworkrdquo httpsarxivorgabs131244002014

[47] K He X Zhang S Ren and J Sun ldquoIdentity mappings in deepresidual networksrdquo in Proceedings of the European Conferenceon Computer Vision pp 630ndash645 2016

[48] L van der Maaten and G Hinton ldquoVisualizing data using t-SNErdquo Journal of Machine Learning Research vol 9 pp 2579ndash2625 2008

International Journal of

AerospaceEngineeringHindawiwwwhindawicom Volume 2018

RoboticsJournal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Active and Passive Electronic Components

VLSI Design

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Shock and Vibration

Hindawiwwwhindawicom Volume 2018

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawiwwwhindawicom

Volume 2018

Hindawi Publishing Corporation httpwwwhindawicom Volume 2013Hindawiwwwhindawicom

The Scientific World Journal

Volume 2018

Control Scienceand Engineering

Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom

Journal ofEngineeringVolume 2018

SensorsJournal of

Hindawiwwwhindawicom Volume 2018

International Journal of

RotatingMachinery

Hindawiwwwhindawicom Volume 2018

Modelling ampSimulationin EngineeringHindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Navigation and Observation

International Journal of

Hindawi

wwwhindawicom Volume 2018

Advances in

Multimedia

Submit your manuscripts atwwwhindawicom

Page 4: ResearchArticle Can Deep Learning Identify Tomato Leaf

4 Advances in Multimedia

Table 1 The raw tomato leaf dataset

Label Category Number Leaf symptoms Illustration

1 Corynesporaleaf spot disease 547

Small brown spotsappear leaf spots have

yellow halo

See Figure 1 first rowNo1-No5

2 Early blight 405

Black or brown spotsappear leaf spots oftenhave yellow or green

concentric ring pattern

See Figure 1 first rowNo6-No10

4 Late blight 726

Water-soaked areaappears and rapidlyenlarges to formpurple-brown

oily-appearing blotches

See Figure 1 second rowNo1-No5

5 Leaf molddisease 480 Irregular yellow or green

area appearsSee Figure 1 second row

No6-No10

6 Septoria leafspot 734

Round spots marginalbrown chlorotic yellow

appear

See Figure 1 third rowNo1-No5

7 Two-spottedspider mite 720 Show white or yellow

spots blade back nettingSee Figure 1 third row

No6-No10

8 Virus disease 481 Develop yellow or greenslightly shrinking

See Figure 1 forth rowNo1-No5

9 Yellow leaf curldisease 814

Develop small and curlupward crumpling andmarginal yellowingbushy appearance

See Figure 1 forth rowNo6-No10

3 Health 643 See Figure 1 fifth rowTotal 5550

the image from left to right (2) flipping the image from topto bottom (3) flipping the image diagonally (4) adjusting thebrightness of image setting the max delta to 04 (5) adjustingthe contrast of image setting the ratio from 02 to 15 (6)adjusting the hue of image setting the max delta to 05 (7)adjusting the saturation of image setting the ratio from 02 to15 (8) rotating the image by 90∘ and 270∘ respectively Thefinal dataset is shown in Table 2 and the label in the first rowrepresents the disease categories which are given in Table 1

33 Deep Learning Models

331 AlexNet AlexNet is the winner of ImageNet Large-Scale Visual Recognition Challenge (ILSVRC) 2012 a deepconvolutional neural network which has 60 million param-eters and 650000 neurons [21] The architecture of AlexNetutilized in this paper is displayed in Figure 3 The AlexNetarchitecture consists of five convolutional layers (ie conv1conv2 and so on) some of which are followed by max-pooling layers (ie pool1 pool2 and pool5) three fullyconnected layers (ie fc6 fc7 and fc8) and a liner layer withsoftmax activation in output In order to reduce overfittingin the fully connected layers a regularization method calledldquodropoutrdquo is used (ie drop6 drop7) [21] The ReLU activa-tion function is applied to each of the first seven layers (ierelu1 relu2 and so on) [45] In Figure 3 the notation119898times119898times119899in each convolutional layer represents the size of the feature

map for each layer 4096 represents the number of neurons ofthe first two fully connected layers The number of neuronsof the final fully connected layer was modified to 9 sincethe classification problem in this work has 9 categories Inaddition the size of input imagesmust be shaped to 227times227which meets the input pixel size requirement of AlexNet

332 GoogLeNet GoogLeNet is an inception architecture[22] which is the winner of ILSVRC 2014 and owns roughly68 million parameters The architecture of GoogLeNet ispresented in Figure 4 The inception module is inspired bythe network in network [46] and uses a parallel combinationof 1 times 1 3 times 3 and 5 times 5 convolutional layer along with3 times 3 max-pooling layer [45] the 1 times 1 convolutional layerbefore 3 times 3 and 5 times 5 convolutional layer reduces the spatialdimension and limits the size of GoogLeNet The wholearchitecture of GoogLeNet is stacked by inception moduleon top of each other (See Figure 4) which has nine inceptionmodules two convolutional layers four max-pooling layersone average pooling layer one fully connected layer and alinear layer with softmax function in the output GoogLeNetuses dropout regularization in the fully connected layerand applies the ReLU activation function in all of theconvolutional layers [29] In this work the last three layersof GoogLeNet were replaced by a fully connected layer asoftmax layer and a classification layer the fully connectedlayer was modified to 9 neurons which is equal to the

Advances in Multimedia 5

Table 2 The final tomato leaf dataset

Labels Label1 Label2 Label3 Label4 Label5 Label6 Label7 Label8 Label9 TotalTraining set 3933 2916 4626 5229 3456 5283 5184 3465 5859 39951Testing set 110 81 129 145 96 147 144 161 163 1176

Figure 3 The architecture of AlexNet in this work

categories in the tomato leaf disease identification problemThe size requested of input image of GoogLeNet is 224times 224333 ResNet The deep residual learning framework is pro-posed for addressing the degradation problem ResNet con-sists of many stacked residual units which won the first placein ILSVRC 2015 and COCO 2015 classification challenge witherror rate of 357 [26] Each unit can be expressed in thefollowing formulas [47]

119910119897 = ℎ (119909119897) + 119865 (119909119897119882119897) (4)

119909119897+1 = 119891 (119910119897) (5)

where 119909119897 and 119909119897+1 are input and output of the l-th unit and 119865is a residual function In [26]ℎ(119909119897) = 119909119897 is an identity mappingand 119891 is a ReLU function [42] A ldquobottleneckrdquo building blockis designed for ResNet (See Figure 5) and comprises two 1times1convolutions with a 3 times 3 convolution in between and adirect skip connection bypassing input and output The 1 times 1layers are responsible for changing in dimensions ResNetmodel has three types of layers with 50 101 and 152 Forsaving computing resources and training time we choose theResNet50 which also has high performance In this work atfirst the last three layers of ResNet were modified by a fullyconnected layer a softmax layer and a classification layerthe fully connected layer was replaced to 9 neurons which isequal to the categories of the tomato leaf disease We changedthe structure of ResNet subsequently The size of input imageof ResNet should satisfy 224 times 2244 Experiments and Results

In this section we reveal the experiments and discuss theexperimental results All the experiments were implementedin Matlab under Windows 10 using the GPU NVIDIAGTX1050 with 4G video memory or NVIDIA GTX1080Tiwith 11G video memory In this paper overall accuracy wasregarded as the evaluation metric in every experiment on

tomato leaf disease detection which means the percentage ofsamples that are correctly classified

accuracy = true positive + true negativepositive + negative (6)

where ldquotrue positiverdquo is the number of instances that are pos-itive and classified as positive ldquotrue negativerdquo is the numberof instances that are negative and classified as negative andthe denominator represents the total number of samples Inaddition the training time was regarded as an additionalperformance metric of the network structure experiment

41 Experiments on Optimization Methods The first exper-iment is designed for seeking the suitable optimizationmethod between SGD [30] and Adam [30 31] in identifyingtomato leaf diseases combining with the pretrained net-work AlexNet GoogLeNet and ResNet respectively In thisexperiment the hyperparameters were set as follows for eachnetwork the batch size was set to 32 the initial learning ratewas set to 0001 and dropped by a factor of 05 every 2 epochsand the max epoch was set to 5 ie the number of iterationsis 6240 So far as SGD optimization method the momentumwas set to 09 For Adam the gradient decay rate 1205731 was set to09 the squared gradient decay rate 1205732 was set to 0999 andthe denominator offset 120576 was set to 10minus8 [31] The accuracyof different networks is displayed in Table 3 In addition wechoose the better results in each deep neural network to showthe training loss against number of iterations during the fine-tuning process (See Figure 6) The words inside parenthesisindicate the corresponding optimization method

In Table 3 the ResNet with SGD optimization methodgets the highest test accuracy 9651 In identifying tomatoleaf diseases the performance of Adam optimization methodis inferior to the SGD optimization method especially incombining with AlexNet In the following paper AlexNet(SGD) GoogLeNet (SGD) and ResNet (SGD) are referred toas AlexNet GoogLeNet and ResNet respectively

As it can be seen in Figure 6 the training loss of ResNetdrops rapidly in the earlier iterations and tends to stable after3000 iterations Consistent with Table 3 the performance of

6 Advances in Multimedia

Inputimage

Conv1 Conv2

MaxPool MaxPool

MaxPool

MaxPool

Inception 3a

Inception 3b

Inception 4a

Inception 4b

Inception 4c

Inception 4d

Inception 4e

Inception 5a

Inception 5b

AvgPool

Dropout40

FC

somax

output

Figure 4 The architecture of GoogLeNet [22 45]

1times1 conv 64

3times3 conv 64

1times1 conv 256

relu

relu

relu

+

Figure 5 ResNet bottleneck residual building block [26]

Table 3 Model recognition accuracy

Model AccuracyAlexNet (SGD) 9583AlexNet (Adam) 1386GoogLeNet (SGD) 9566GoogLeNet (Adam) 9406ResNet (SGD) 9651ResNet (Adam) 9439

AlexNet and GoogLeNet is similar and both inferior to theResNet

42 Experiments on Batch Size and Number of IterationsFrom the experiment on optimization methods the ResNetobtains the highest classification accuracy Next we evaluatedthe effects of batch size and the number of iterations on theperformance of the ResNet The batch size was set to 16 32and 64 respectively Meanwhile the number of iterations wasset to 2496 4992 and 9984 The classification accuracy ofdifferent training scenarios is given in Table 4 At the sametime the classification accuracy of each labels representativeleaf disease category (See Table 1) is given In this experimentthe initial learning rate was set to 0001 and dropped by afactor of 05 every 2496 iterations

In Table 4 the best overall classification accuracy 9719is got by the ResNet combining with batch size 16 and

Trai

ning

Los

s

Number of Iterations

AlexNetGoogLeNetResNet

3

25

2

15

1

05

00 1000 2000 3000 4000 5000 6000

Figure 6 The training loss during the fine-tuning process

iterations 4992 As shown in Table 4 whether increasingthe number of iterations or batch size the performance ofcorresponding models has not been improved significantlyin identifying tomato leaf disease A small batch size witha medium number of iterations is quite effective in thiswork Moreover a larger batch size and number of iterationsincreases the training duration We have not tried higher orlower values for the attempted parameters since differentclassification task may have various suitable parameters andit is hard to give a certain rule in setting hyperparameters

43 Experiments on Full Training and Fine-Tuning of ResNetThis section is designed for exploring the performance ofCNN by changing the structure of the models In practicala deep CNN always owns a large size which means a largenumber of parameters Thus full training of a deep CNNrequires extensive computational resources and is time-consuming In addition full training of a deep CNN mayled to overfitting when the training data is limited So wecompared the performance of the pretrained CNN throughfull training and fine-tuning their structures

We changed the structure of ResNet and combination ofthe best parameters from the front experiments was utilized

Advances in Multimedia 7

Table 4 Classification accuracies with different parameters during fine-tuning of the ResNetThe numbers inside parenthesis indicate batchsize and number of iterations

Networks label1 label2 label3 label4 label5 label6 label7 label8 label9 overallResNet (162496) 9091 8889 100 100 9688 100 9028 8820 9755 9498ResNet (164992) 9818 9877 100 9862 9688 100 9653 8882 9877 9719ResNet (169984) 9818 9753 100 9862 9688 100 9722 8696 9877 9694ResNet (322496) 9727 9506 100 9793 9688 100 9514 8634 9939 9634ResNet (324992) 9727 9506 100 9724 9688 100 9653 8696 9939 9651ResNet (329984) 9636 9630 100 9931 9688 100 9444 8820 9877 9660ResNet (642496) 9364 9383 100 9724 9688 100 9444 8758 9939 9592ResNet (644992) 9455 9529 100 9655 9583 100 9583 8696 9939 9583ResNet (649984) 9545 9383 100 9793 9688 100 9444 8758 9939 9617

ResNet50 has 177 layers if the layers for each building blockand connection are calculated In this experiment the lastthree layers of ResNet were modified to a fully connectedlayer (denoted as ldquofcrdquo) a softmax layer and a classificationlayer and the fully connected layer owns 9 neurons Thestructure was changed by freezing the weights of a certainnumber of layers in the network by setting the learning ratein those layers to zero During training the parameters ofthe frozen layers are not updated Full training and fine-tuning are defined by the number of training layers ie fulltraining (1-ldquofcrdquo) fine-tuning (37-ldquofcrdquo 79-ldquofcrdquo 111-ldquofcrdquo 141-ldquofcrdquo 163-ldquofcrdquo) The accuracy and training time of differentnetwork structure are presented in Table 5 At first the batchsize and 4992 iterations were combined the initial learningrate was set to 0001 and dropped by a factor of 01 every2496 iterations In order to get more convincing conclusionsResNet (16 9984) which gets the second place in Table 4 wasalso used to execute the experiments

In Table 5 the accuracy and training time of differentnetwork structures are presented In two cases ie the 4992iterations and 9984 iterations of ResNet the accuracy of themodel from the 37 layer fine-tuning structure are higherthan that of the full training model In the case where thenumber of iterations is 4992 the accuracy of the model fromthe 79 layer fine-tuning structure is equal to that of the fulltraining modelThe final column of the Table 5 represents thetraining time of the corresponding network and it is clearthat the training time of the fine-tuning models is greatlylowered than the full training model Because the gradientsof the frozen layers do not need to be computed freezingthe weights of initial layers can speed up network trainingWe observe that the moderate fine-tuning models (37-ldquofcrdquo79-ldquofcrdquo 111-ldquofcrdquo) always led to a performance superior orapproximately equal to the full training models Thus wesuggest that for practical application the moderate fine-tuning models may be a good choice Especially for theresearcher who holds massive data the fine-tuning modelsmay achieve good performance while saving computationalresources and time

Moreover the features of the final fully connected layerof ResNet (16 4992 37-ldquofcrdquo) were examined by utilizingthe t-distributed Stochastic Neighbour Embedding (t-SNE)algorithm (see Figure 7) [48] 1176 test images were used to

label1 label2 label3 label4 label5

label6 label7 label8 label9

Figure 7 Two-dimensional scatter plot of high-dimensional fea-tures generated with t-SNE

extract the features In Figure 7 different colors representdifferent labels the corresponding disease categories of thelabels were listed in Table 1 As shown in Figure 7 9 differentcolor points are clearly separated which indicates that thefeatures learned from the ResNet with the optimal structurecan be used to classify the tomato leaf disease precisely

5 Conclusion

This paper concentrates on identifying tomato leaf diseaseusing deep convolutional neural networks by transfer learn-ing The utilized networks are based on the pretrained deeplearning models of AlexNet GoogLeNet and ResNet Firstwe compared the relative performance of these networksby using SGD and Adam optimization method revealingthat the ResNet with SGD optimization method obtainsthe highest result with the best accuracy 9651 Thenthe performance evaluation of batch size and number of

8 Advances in Multimedia

Table 5 Accuracies and training time in different network structuresThe values inside parenthesis denote batch size number of iterationsand training layers

Network topology Accuracy Time (minsec)ResNet (16 4992 1-ldquofcrdquo) 9643 59min 30secResNet (16 4992 37-ldquofcrdquo) 9728 44min 13secResNet (16 4992 79-ldquofcrdquo) 9643 37min 27secResNet (16 4992 111-ldquofcrdquo) 9575 30min 6secResNet (16 4992 141-ldquofcrdquo) 9532 24min 15secResNet (16 4992 163-ldquofcrdquo) 9269 19min 31secResNet (16 9984 1-ldquofcrdquo) 9694 118min 32secResNet (16 9984 37-ldquofcrdquo) 9702 92min 53secResNet (16 9984 79-ldquofcrdquo) 9677 72min 23secResNet (16 9984 111-ldquofcrdquo) 9626 58min 40secResNet (16 9984 141-ldquofcrdquo) 9575 47min 22secResNet (16 9984 163-ldquofcrdquo) 9396 39min 32sec

iterations affecting the transfer learning of the ResNet wasconducted A small batch size of 16 combining a moderatenumber of iterations of 4992 is the optimal choice in thiswork Our findings suggest that for a particular task neitherlarge batch size nor large number of iterations may improvethe accuracy of the target modelThe setting of batch size andnumber of iterations depends on your data set and the utilizednetwork Next the best combined model was used to fine-tune the structure Fine-tuning ResNet layers from 37 to ldquofcrdquoobtained the highest accuracy 9728 in identifying tomatoleaf disease Based on the amount of available data layer-wisefine-tuning may provide a practical way to achieve the bestperformance of the application at hand We believe that theresults obtained in this work will bring some inspiration toother similar visual recognition problems and the practicalstudy of this work can be easily extended to other plant leafdisease identification problems

Data Availability

Thetomato leaf data supporting this work are frompreviouslyreported studies which have been cited The processed dataare available from the corresponding author request

Conflicts of Interest

The authors declare that there are no conflicts of interestregarding the publication of this paper

Acknowledgments

This study was supported by the National Science and tech-nology support program (2014BAD12B01-1-3) PublicWelfareIndustry (Agriculture) Research Projects Level-2 (201503116-04-06) Postdoctoral Foundation of Heilongjiang Province(LBHZ15020) Harbin Applied Technology Research andDevelopment Program (2017RAQXJ096) and EconomicDecision Making and Early Warning of Soybean Industryin Technology Collaborative Innovation System of SoybeanIndustry in Heilongjiang Province (20170401)

References

[1] RChaerani andR EVoorrips ldquoTomato early blight (Alternariasolani) The pathogen genetics and breeding for resistancerdquoJournal of General Plant Pathology vol 72 no 6 pp 335ndash3472006

[2] A M Dickey L S Osborne and C L Mckenzie ldquoPapaya(Carica papaya Brassicales Caricaceae) is not a host plant oftomato yellow leaf curl virus (TYLCV family Geminiviridaegenus Begomovirus)rdquo Florida Entomologist vol 95 no 1 pp211ndash213 2012

[3] G Wei L Baoju S Yanxia and X Xuewen ldquoStudies on path-ogenicity differentiation of corynespora cassiicola isolatesagainst cucumber tomato and eggplantrdquo Acta HorticulturaeSinica vol 38 no 3 pp 465ndash470 2011

[4] P Lindhout W Korta M Cislik I Vos and T GerlaghldquoFurther identification of races of Cladosporium fulvum (Fulviafulva) on tomato originating from the Netherlands France andPolandrdquo Netherlands Journal of Plant Pathology vol 95 no 3pp 143ndash148 1989

[5] K Kubota S Tsuda A Tamai and T Meshi ldquoTomato mosaicvirus replication protein suppresses virus-targeted posttran-scriptional gene silencingrdquo Journal of Virology vol 77 no 20pp 11016ndash11026 2003

[6] M Tian B Benedetti and S Kamoun ldquoA second Kazal-like protease inhibitor from Phytophthora infestans inhibitsand interacts with the apoplastic pathogenesis-related proteaseP69B of tomatordquo Plant Physiology vol 138 no 3 pp 1785ndash17932005

[7] L E Blum ldquoReduction of incidence and severity of Septorialycopersici leaf spot of tomatowith bacteria and yeastsrdquoCienciaRural vol 30 no 5 pp 761ndash765 2000

[8] E A Chatzivasileiadis and M W Sabelis ldquoToxicity of methylketones from tomato trichomes to Tetranychus urticae KochrdquoExperimental and Applied Acarology vol 21 no 6-7 pp 473ndash484 1997

[9] M Anthimopoulos S Christodoulidis L Ebner A Christeand S Mougiakakou ldquoLung Pattern Classification for Inter-stitial Lung Diseases Using a Deep Convolutional NeuralNetworkrdquo IEEE Transactions on Medical Imaging vol 35 no5 pp 1207ndash1216 2016

[10] Y Tang and X Wu ldquoText-independent writer identification viaCNN features and joint Bayesianrdquo in Proceedings of the 15th

Advances in Multimedia 9

International Conference on Frontiers in Handwriting Recogni-tion ICFHR 2016 pp 566ndash571 Shenzhen China October 2016

[11] Y Tang and XWu ldquoSaliency Detection via Combining Region-Level and Pixel-Level Predictions with CNNsrdquo in Proceedings ofthe EuropeanConference onComputer Vision (ECCV) pp 1608ndash05186 2016

[12] Y Tang and X Wu ldquoSalient object detection with chainedmulti-scale fully convolutional networkrdquo ACM Multimedia(ACMMM) pp 618ndash626 2017

[13] Y Tang and X Wu ldquoScene text detection and segmentationbased on cascaded convolution neural networksrdquo IEEE Trans-actions on Image Processing vol 26 no 3 pp 1509ndash1520 2017

[14] Y Tang and X Wu ldquoScene Text Detection using Superpixelbased Stroke Feature Transform and Deep Learning basedRegion Classificationrdquo IEEE Transactions on Multimedia vol20 no 9 pp 2276ndash2288 2018

[15] Y Yao XWu Z Lei S Shan andW Zuo ldquoJoint Representationand Truncated Inference Learning for Correlation Filter basedTrackingrdquo in Proceedings of the European Conference on Com-puter Vision (ECCV) pp 1ndash14 2018

[16] L Zhang F Yang Y Daniel Zhang and Y J Zhu ldquoRoadcrack detection using deep convolutional neural networkrdquoin Proceedings of the 23rd IEEE International Conference onImage Processing ICIP 2016 pp 3708ndash3712 Phoenix AZ USASeptember 2016

[17] D Xie L Zhang andL Bai ldquoDeep learning in visual computingand signal processingrdquo Applied Computational Intelligence andSoft Computing vol 2017 Article ID 1320780 13 pages 2017

[18] Z Zhou J Shin L Zhang S Gurudu M Gotway and J LiangldquoFine-tuning convolutional neural networks for biomedicalimage analysis Actively and incrementallyrdquo inProceedings of the30th IEEE Conference on Computer Vision and Pattern Recog-nition CVPR 2017 pp 4761ndash4772 USA July 2017

[19] Z Liu P Luo X Wang and X Tang ldquoDeep learning face attri-butes in the wildrdquo in Proceedings of the 15th IEEE InternationalConference on Computer Vision ICCV 2015 pp 3730ndash3738Santiago Chile 2015

[20] W Ouyang and X Wang ldquoJoint deep learning for pedestriandetectionrdquo in Proceedings of the 14th IEEE International Con-ference on Computer Vision (ICCV rsquo13) pp 2056ndash2063 SydneyAustralia December 2013

[21] A Krizhevsky I Sutskever andG EHinton ldquoImagenet classifi-cation with deep convolutional neural networksrdquo in Proceedingsof the 26th Annual Conference on Neural Information ProcessingSystems (NIPS rsquo12) pp 1097ndash1105 Lake Tahoe Nev USADecember 2012

[22] C Szegedy W Liu Y Jia et al ldquoGoing deeper with convolu-tionsrdquo inProceedings of the IEEEConference onComputer Visionand Pattern Recognition (CVPR rsquo15) pp 1ndash9 IEEE BostonMass USA June 2015

[23] K Simonyan and A Zisserman ldquoVery deep convolutional net-works for large-scale image recognitionrdquo httpsarxivorgabs14091556 2015

[24] C Szegedy V Vanhoucke S Ioffe J Shlens and Z WojnaldquoRethinking the inception architecture for computer visionrdquohttpsarxivorgabs151200567 2015

[25] C Szegedy S Ioffe V Vanhoucke and A Alemi ldquoInception-v4 inception-ResNet and the impact of residual connections onlearningrdquo httpsarxivorgabs160207261 2016

[26] K He X Zhang S Ren and J Sun ldquoDeep residual learning forimage recognitionrdquo in Proceedings of the 2016 IEEE Conference

on Computer Vision and Pattern Recognition CVPR 2016 pp770ndash778 July 2016

[27] G Huang Z Liu K Q Weinberger and L van der MaatenldquoDensely connected convolutional networksrdquo in Proceedingsof the 2017 IEEE Conference on Computer Vision and PatternRecognition (CVPR) Honolulu HI USA 2016

[28] S J Pan and Q Yang ldquoA survey on transfer learningrdquo IEEETransactions on Knowledge and Data Engineering vol 22 no10 pp 1345ndash1359 2010

[29] M Mehdipour Ghazi B Yanikoglu and E Aptoula ldquoPlantidentification using deep neural networks via optimization oftransfer learning parametersrdquo Neurocomputing vol 235 pp228ndash235 2017

[30] S Ruder ldquoAn overview of gradient descent optimization algo-rithmsrdquo httpsarxivorgabs160904747 2017

[31] D P Kingma and J L Ba ldquoAdam a method for stochasticoptimizationrdquo httpsarxivorgabs14126980 2017

[32] S S Sannakki V S Rajpurohit V B Nargund R Arun Kumarand S Prema Yallur ldquoLeaf disease grading by machine visionand fuzzy logicrdquo International Journal of Computer Technologyand Applications vol 2 no 5 pp 1709ndash1716 2011

[33] D Samanta P P Chaudhury and A Ghosh ldquoScab diseasesdetection of potato using image processingrdquo InternationalJournal of Computer Trends and Technology vol 3 pp 109ndash1132012

[34] P J Herrera J Dorado and A Ribeiro ldquoA novel approach forweed type classification based on shape descriptors and a fuzzydecision-making methodrdquo Sensors vol 14 no 8 pp 15304ndash15324 2014

[35] B Cheng and E T Matson ldquoA feature-based machine learningagent for automatic rice andweed discriminationrdquo InternationalConference onArtificial Intelligence and SoftComputing pp 517ndash527 2015

[36] S Sankaran and R Ehsani ldquoComparison of visible-near infra-red and mid-infrared spectroscopy for classification of Huang-longbing and citrus canker infected leavesrdquo Agricultural Engi-neering International CIGR Journal vol 15 no 3 pp 75ndash792013

[37] X Cheng Y Zhang Y Chen Y Wu and Y Yue ldquoPest iden-tification via deep residual learning in complex backgroundrdquoComputers and Electronics in Agriculture vol 141 pp 351ndash3562017

[38] A dos Santos Ferreira D Matte Freitas G Goncalves daSilva H Pistori and M Theophilo Folhes ldquoWeed detection insoybean crops using ConvNetsrdquo Computers and Electronics inAgriculture vol 143 pp 314ndash324 2017

[39] S Sladojevic M Arsenovic A Anderla D Culibrk andD Stefanovic ldquoDeep Neural Networks Based Recognition ofPlant Diseases by Leaf Image Classificationrdquo ComputationalIntelligence and Neuroscience vol 2016 Article ID 3289801 11pages 2016

[40] S P Mohanty D P Hughes and M Salathe ldquoUsing deep learn-ing for image-based plant disease detectionrdquo Frontiers in PlantScience vol 7 article no 1419 2016

[41] I Sa Z Ge F Dayoub B Upcroft T Perez and C McCoolldquoDeepfruits A fruit detection system using deep neural net-worksrdquo Sensors vol 16 article no 1222 no 8 2016

[42] V Nair and G E Hinton ldquoRectified linear units improveRestricted Boltzmann machinesrdquo in Proceedings of the 27thInternational Conference on Machine Learning (ICML rsquo10) pp807ndash814 Haifa Israel June 2010

10 Advances in Multimedia

[43] D P Hughes and M Salathe ldquoAn open access repository ofimages on plant health to enable the development of mobiledisease diagnosticsrdquo httpsarxivorgabs151108060 2016

[44] D Ciresan U Meier J Masci and J Schmidhuber ldquoA com-mittee of neural networks for traffic sign classificationrdquo inProceedings of the 2011 International Joint Conference on NeuralNetworks (IJCNN 2011 - San Jose) San Jose CA USA July 2011

[45] P Pawara E Okafor O Surinta L Schomaker andMWieringldquoComparing Local Descriptors and Bags of Visual Words toDeep Convolutional Neural Networks for Plant Recognitionrdquoin Proceedings of the 6th International Conference on PatternRecognition Applications and Methods pp 479ndash486 PortoPortugal Feburary 2017

[46] M Lin ldquoNetwork inNnetworkrdquo httpsarxivorgabs131244002014

[47] K He X Zhang S Ren and J Sun ldquoIdentity mappings in deepresidual networksrdquo in Proceedings of the European Conferenceon Computer Vision pp 630ndash645 2016

[48] L van der Maaten and G Hinton ldquoVisualizing data using t-SNErdquo Journal of Machine Learning Research vol 9 pp 2579ndash2625 2008

International Journal of

AerospaceEngineeringHindawiwwwhindawicom Volume 2018

RoboticsJournal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Active and Passive Electronic Components

VLSI Design

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Shock and Vibration

Hindawiwwwhindawicom Volume 2018

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawiwwwhindawicom

Volume 2018

Hindawi Publishing Corporation httpwwwhindawicom Volume 2013Hindawiwwwhindawicom

The Scientific World Journal

Volume 2018

Control Scienceand Engineering

Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom

Journal ofEngineeringVolume 2018

SensorsJournal of

Hindawiwwwhindawicom Volume 2018

International Journal of

RotatingMachinery

Hindawiwwwhindawicom Volume 2018

Modelling ampSimulationin EngineeringHindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Navigation and Observation

International Journal of

Hindawi

wwwhindawicom Volume 2018

Advances in

Multimedia

Submit your manuscripts atwwwhindawicom

Page 5: ResearchArticle Can Deep Learning Identify Tomato Leaf

Advances in Multimedia 5

Table 2 The final tomato leaf dataset

Labels Label1 Label2 Label3 Label4 Label5 Label6 Label7 Label8 Label9 TotalTraining set 3933 2916 4626 5229 3456 5283 5184 3465 5859 39951Testing set 110 81 129 145 96 147 144 161 163 1176

Figure 3 The architecture of AlexNet in this work

categories in the tomato leaf disease identification problemThe size requested of input image of GoogLeNet is 224times 224333 ResNet The deep residual learning framework is pro-posed for addressing the degradation problem ResNet con-sists of many stacked residual units which won the first placein ILSVRC 2015 and COCO 2015 classification challenge witherror rate of 357 [26] Each unit can be expressed in thefollowing formulas [47]

119910119897 = ℎ (119909119897) + 119865 (119909119897119882119897) (4)

119909119897+1 = 119891 (119910119897) (5)

where 119909119897 and 119909119897+1 are input and output of the l-th unit and 119865is a residual function In [26]ℎ(119909119897) = 119909119897 is an identity mappingand 119891 is a ReLU function [42] A ldquobottleneckrdquo building blockis designed for ResNet (See Figure 5) and comprises two 1times1convolutions with a 3 times 3 convolution in between and adirect skip connection bypassing input and output The 1 times 1layers are responsible for changing in dimensions ResNetmodel has three types of layers with 50 101 and 152 Forsaving computing resources and training time we choose theResNet50 which also has high performance In this work atfirst the last three layers of ResNet were modified by a fullyconnected layer a softmax layer and a classification layerthe fully connected layer was replaced to 9 neurons which isequal to the categories of the tomato leaf disease We changedthe structure of ResNet subsequently The size of input imageof ResNet should satisfy 224 times 2244 Experiments and Results

In this section we reveal the experiments and discuss theexperimental results All the experiments were implementedin Matlab under Windows 10 using the GPU NVIDIAGTX1050 with 4G video memory or NVIDIA GTX1080Tiwith 11G video memory In this paper overall accuracy wasregarded as the evaluation metric in every experiment on

tomato leaf disease detection which means the percentage ofsamples that are correctly classified

accuracy = true positive + true negativepositive + negative (6)

where ldquotrue positiverdquo is the number of instances that are pos-itive and classified as positive ldquotrue negativerdquo is the numberof instances that are negative and classified as negative andthe denominator represents the total number of samples Inaddition the training time was regarded as an additionalperformance metric of the network structure experiment

41 Experiments on Optimization Methods The first exper-iment is designed for seeking the suitable optimizationmethod between SGD [30] and Adam [30 31] in identifyingtomato leaf diseases combining with the pretrained net-work AlexNet GoogLeNet and ResNet respectively In thisexperiment the hyperparameters were set as follows for eachnetwork the batch size was set to 32 the initial learning ratewas set to 0001 and dropped by a factor of 05 every 2 epochsand the max epoch was set to 5 ie the number of iterationsis 6240 So far as SGD optimization method the momentumwas set to 09 For Adam the gradient decay rate 1205731 was set to09 the squared gradient decay rate 1205732 was set to 0999 andthe denominator offset 120576 was set to 10minus8 [31] The accuracyof different networks is displayed in Table 3 In addition wechoose the better results in each deep neural network to showthe training loss against number of iterations during the fine-tuning process (See Figure 6) The words inside parenthesisindicate the corresponding optimization method

In Table 3 the ResNet with SGD optimization methodgets the highest test accuracy 9651 In identifying tomatoleaf diseases the performance of Adam optimization methodis inferior to the SGD optimization method especially incombining with AlexNet In the following paper AlexNet(SGD) GoogLeNet (SGD) and ResNet (SGD) are referred toas AlexNet GoogLeNet and ResNet respectively

As it can be seen in Figure 6 the training loss of ResNetdrops rapidly in the earlier iterations and tends to stable after3000 iterations Consistent with Table 3 the performance of

6 Advances in Multimedia

Inputimage

Conv1 Conv2

MaxPool MaxPool

MaxPool

MaxPool

Inception 3a

Inception 3b

Inception 4a

Inception 4b

Inception 4c

Inception 4d

Inception 4e

Inception 5a

Inception 5b

AvgPool

Dropout40

FC

somax

output

Figure 4 The architecture of GoogLeNet [22 45]

1times1 conv 64

3times3 conv 64

1times1 conv 256

relu

relu

relu

+

Figure 5 ResNet bottleneck residual building block [26]

Table 3 Model recognition accuracy

Model AccuracyAlexNet (SGD) 9583AlexNet (Adam) 1386GoogLeNet (SGD) 9566GoogLeNet (Adam) 9406ResNet (SGD) 9651ResNet (Adam) 9439

AlexNet and GoogLeNet is similar and both inferior to theResNet

42 Experiments on Batch Size and Number of IterationsFrom the experiment on optimization methods the ResNetobtains the highest classification accuracy Next we evaluatedthe effects of batch size and the number of iterations on theperformance of the ResNet The batch size was set to 16 32and 64 respectively Meanwhile the number of iterations wasset to 2496 4992 and 9984 The classification accuracy ofdifferent training scenarios is given in Table 4 At the sametime the classification accuracy of each labels representativeleaf disease category (See Table 1) is given In this experimentthe initial learning rate was set to 0001 and dropped by afactor of 05 every 2496 iterations

In Table 4 the best overall classification accuracy 9719is got by the ResNet combining with batch size 16 and

Trai

ning

Los

s

Number of Iterations

AlexNetGoogLeNetResNet

3

25

2

15

1

05

00 1000 2000 3000 4000 5000 6000

Figure 6 The training loss during the fine-tuning process

iterations 4992 As shown in Table 4 whether increasingthe number of iterations or batch size the performance ofcorresponding models has not been improved significantlyin identifying tomato leaf disease A small batch size witha medium number of iterations is quite effective in thiswork Moreover a larger batch size and number of iterationsincreases the training duration We have not tried higher orlower values for the attempted parameters since differentclassification task may have various suitable parameters andit is hard to give a certain rule in setting hyperparameters

43 Experiments on Full Training and Fine-Tuning of ResNetThis section is designed for exploring the performance ofCNN by changing the structure of the models In practicala deep CNN always owns a large size which means a largenumber of parameters Thus full training of a deep CNNrequires extensive computational resources and is time-consuming In addition full training of a deep CNN mayled to overfitting when the training data is limited So wecompared the performance of the pretrained CNN throughfull training and fine-tuning their structures

We changed the structure of ResNet and combination ofthe best parameters from the front experiments was utilized

Advances in Multimedia 7

Table 4 Classification accuracies with different parameters during fine-tuning of the ResNetThe numbers inside parenthesis indicate batchsize and number of iterations

Networks label1 label2 label3 label4 label5 label6 label7 label8 label9 overallResNet (162496) 9091 8889 100 100 9688 100 9028 8820 9755 9498ResNet (164992) 9818 9877 100 9862 9688 100 9653 8882 9877 9719ResNet (169984) 9818 9753 100 9862 9688 100 9722 8696 9877 9694ResNet (322496) 9727 9506 100 9793 9688 100 9514 8634 9939 9634ResNet (324992) 9727 9506 100 9724 9688 100 9653 8696 9939 9651ResNet (329984) 9636 9630 100 9931 9688 100 9444 8820 9877 9660ResNet (642496) 9364 9383 100 9724 9688 100 9444 8758 9939 9592ResNet (644992) 9455 9529 100 9655 9583 100 9583 8696 9939 9583ResNet (649984) 9545 9383 100 9793 9688 100 9444 8758 9939 9617

ResNet50 has 177 layers if the layers for each building blockand connection are calculated In this experiment the lastthree layers of ResNet were modified to a fully connectedlayer (denoted as ldquofcrdquo) a softmax layer and a classificationlayer and the fully connected layer owns 9 neurons Thestructure was changed by freezing the weights of a certainnumber of layers in the network by setting the learning ratein those layers to zero During training the parameters ofthe frozen layers are not updated Full training and fine-tuning are defined by the number of training layers ie fulltraining (1-ldquofcrdquo) fine-tuning (37-ldquofcrdquo 79-ldquofcrdquo 111-ldquofcrdquo 141-ldquofcrdquo 163-ldquofcrdquo) The accuracy and training time of differentnetwork structure are presented in Table 5 At first the batchsize and 4992 iterations were combined the initial learningrate was set to 0001 and dropped by a factor of 01 every2496 iterations In order to get more convincing conclusionsResNet (16 9984) which gets the second place in Table 4 wasalso used to execute the experiments

In Table 5 the accuracy and training time of differentnetwork structures are presented In two cases ie the 4992iterations and 9984 iterations of ResNet the accuracy of themodel from the 37 layer fine-tuning structure are higherthan that of the full training model In the case where thenumber of iterations is 4992 the accuracy of the model fromthe 79 layer fine-tuning structure is equal to that of the fulltraining modelThe final column of the Table 5 represents thetraining time of the corresponding network and it is clearthat the training time of the fine-tuning models is greatlylowered than the full training model Because the gradientsof the frozen layers do not need to be computed freezingthe weights of initial layers can speed up network trainingWe observe that the moderate fine-tuning models (37-ldquofcrdquo79-ldquofcrdquo 111-ldquofcrdquo) always led to a performance superior orapproximately equal to the full training models Thus wesuggest that for practical application the moderate fine-tuning models may be a good choice Especially for theresearcher who holds massive data the fine-tuning modelsmay achieve good performance while saving computationalresources and time

Moreover the features of the final fully connected layerof ResNet (16 4992 37-ldquofcrdquo) were examined by utilizingthe t-distributed Stochastic Neighbour Embedding (t-SNE)algorithm (see Figure 7) [48] 1176 test images were used to

label1 label2 label3 label4 label5

label6 label7 label8 label9

Figure 7 Two-dimensional scatter plot of high-dimensional fea-tures generated with t-SNE

extract the features In Figure 7 different colors representdifferent labels the corresponding disease categories of thelabels were listed in Table 1 As shown in Figure 7 9 differentcolor points are clearly separated which indicates that thefeatures learned from the ResNet with the optimal structurecan be used to classify the tomato leaf disease precisely

5 Conclusion

This paper concentrates on identifying tomato leaf diseaseusing deep convolutional neural networks by transfer learn-ing The utilized networks are based on the pretrained deeplearning models of AlexNet GoogLeNet and ResNet Firstwe compared the relative performance of these networksby using SGD and Adam optimization method revealingthat the ResNet with SGD optimization method obtainsthe highest result with the best accuracy 9651 Thenthe performance evaluation of batch size and number of

8 Advances in Multimedia

Table 5 Accuracies and training time in different network structuresThe values inside parenthesis denote batch size number of iterationsand training layers

Network topology Accuracy Time (minsec)ResNet (16 4992 1-ldquofcrdquo) 9643 59min 30secResNet (16 4992 37-ldquofcrdquo) 9728 44min 13secResNet (16 4992 79-ldquofcrdquo) 9643 37min 27secResNet (16 4992 111-ldquofcrdquo) 9575 30min 6secResNet (16 4992 141-ldquofcrdquo) 9532 24min 15secResNet (16 4992 163-ldquofcrdquo) 9269 19min 31secResNet (16 9984 1-ldquofcrdquo) 9694 118min 32secResNet (16 9984 37-ldquofcrdquo) 9702 92min 53secResNet (16 9984 79-ldquofcrdquo) 9677 72min 23secResNet (16 9984 111-ldquofcrdquo) 9626 58min 40secResNet (16 9984 141-ldquofcrdquo) 9575 47min 22secResNet (16 9984 163-ldquofcrdquo) 9396 39min 32sec

iterations affecting the transfer learning of the ResNet wasconducted A small batch size of 16 combining a moderatenumber of iterations of 4992 is the optimal choice in thiswork Our findings suggest that for a particular task neitherlarge batch size nor large number of iterations may improvethe accuracy of the target modelThe setting of batch size andnumber of iterations depends on your data set and the utilizednetwork Next the best combined model was used to fine-tune the structure Fine-tuning ResNet layers from 37 to ldquofcrdquoobtained the highest accuracy 9728 in identifying tomatoleaf disease Based on the amount of available data layer-wisefine-tuning may provide a practical way to achieve the bestperformance of the application at hand We believe that theresults obtained in this work will bring some inspiration toother similar visual recognition problems and the practicalstudy of this work can be easily extended to other plant leafdisease identification problems

Data Availability

Thetomato leaf data supporting this work are frompreviouslyreported studies which have been cited The processed dataare available from the corresponding author request

Conflicts of Interest

The authors declare that there are no conflicts of interestregarding the publication of this paper

Acknowledgments

This study was supported by the National Science and tech-nology support program (2014BAD12B01-1-3) PublicWelfareIndustry (Agriculture) Research Projects Level-2 (201503116-04-06) Postdoctoral Foundation of Heilongjiang Province(LBHZ15020) Harbin Applied Technology Research andDevelopment Program (2017RAQXJ096) and EconomicDecision Making and Early Warning of Soybean Industryin Technology Collaborative Innovation System of SoybeanIndustry in Heilongjiang Province (20170401)

References

[1] RChaerani andR EVoorrips ldquoTomato early blight (Alternariasolani) The pathogen genetics and breeding for resistancerdquoJournal of General Plant Pathology vol 72 no 6 pp 335ndash3472006

[2] A M Dickey L S Osborne and C L Mckenzie ldquoPapaya(Carica papaya Brassicales Caricaceae) is not a host plant oftomato yellow leaf curl virus (TYLCV family Geminiviridaegenus Begomovirus)rdquo Florida Entomologist vol 95 no 1 pp211ndash213 2012

[3] G Wei L Baoju S Yanxia and X Xuewen ldquoStudies on path-ogenicity differentiation of corynespora cassiicola isolatesagainst cucumber tomato and eggplantrdquo Acta HorticulturaeSinica vol 38 no 3 pp 465ndash470 2011

[4] P Lindhout W Korta M Cislik I Vos and T GerlaghldquoFurther identification of races of Cladosporium fulvum (Fulviafulva) on tomato originating from the Netherlands France andPolandrdquo Netherlands Journal of Plant Pathology vol 95 no 3pp 143ndash148 1989

[5] K Kubota S Tsuda A Tamai and T Meshi ldquoTomato mosaicvirus replication protein suppresses virus-targeted posttran-scriptional gene silencingrdquo Journal of Virology vol 77 no 20pp 11016ndash11026 2003

[6] M Tian B Benedetti and S Kamoun ldquoA second Kazal-like protease inhibitor from Phytophthora infestans inhibitsand interacts with the apoplastic pathogenesis-related proteaseP69B of tomatordquo Plant Physiology vol 138 no 3 pp 1785ndash17932005

[7] L E Blum ldquoReduction of incidence and severity of Septorialycopersici leaf spot of tomatowith bacteria and yeastsrdquoCienciaRural vol 30 no 5 pp 761ndash765 2000

[8] E A Chatzivasileiadis and M W Sabelis ldquoToxicity of methylketones from tomato trichomes to Tetranychus urticae KochrdquoExperimental and Applied Acarology vol 21 no 6-7 pp 473ndash484 1997

[9] M Anthimopoulos S Christodoulidis L Ebner A Christeand S Mougiakakou ldquoLung Pattern Classification for Inter-stitial Lung Diseases Using a Deep Convolutional NeuralNetworkrdquo IEEE Transactions on Medical Imaging vol 35 no5 pp 1207ndash1216 2016

[10] Y Tang and X Wu ldquoText-independent writer identification viaCNN features and joint Bayesianrdquo in Proceedings of the 15th

Advances in Multimedia 9

International Conference on Frontiers in Handwriting Recogni-tion ICFHR 2016 pp 566ndash571 Shenzhen China October 2016

[11] Y Tang and XWu ldquoSaliency Detection via Combining Region-Level and Pixel-Level Predictions with CNNsrdquo in Proceedings ofthe EuropeanConference onComputer Vision (ECCV) pp 1608ndash05186 2016

[12] Y Tang and X Wu ldquoSalient object detection with chainedmulti-scale fully convolutional networkrdquo ACM Multimedia(ACMMM) pp 618ndash626 2017

[13] Y Tang and X Wu ldquoScene text detection and segmentationbased on cascaded convolution neural networksrdquo IEEE Trans-actions on Image Processing vol 26 no 3 pp 1509ndash1520 2017

[14] Y Tang and X Wu ldquoScene Text Detection using Superpixelbased Stroke Feature Transform and Deep Learning basedRegion Classificationrdquo IEEE Transactions on Multimedia vol20 no 9 pp 2276ndash2288 2018

[15] Y Yao XWu Z Lei S Shan andW Zuo ldquoJoint Representationand Truncated Inference Learning for Correlation Filter basedTrackingrdquo in Proceedings of the European Conference on Com-puter Vision (ECCV) pp 1ndash14 2018

[16] L Zhang F Yang Y Daniel Zhang and Y J Zhu ldquoRoadcrack detection using deep convolutional neural networkrdquoin Proceedings of the 23rd IEEE International Conference onImage Processing ICIP 2016 pp 3708ndash3712 Phoenix AZ USASeptember 2016

[17] D Xie L Zhang andL Bai ldquoDeep learning in visual computingand signal processingrdquo Applied Computational Intelligence andSoft Computing vol 2017 Article ID 1320780 13 pages 2017

[18] Z Zhou J Shin L Zhang S Gurudu M Gotway and J LiangldquoFine-tuning convolutional neural networks for biomedicalimage analysis Actively and incrementallyrdquo inProceedings of the30th IEEE Conference on Computer Vision and Pattern Recog-nition CVPR 2017 pp 4761ndash4772 USA July 2017

[19] Z Liu P Luo X Wang and X Tang ldquoDeep learning face attri-butes in the wildrdquo in Proceedings of the 15th IEEE InternationalConference on Computer Vision ICCV 2015 pp 3730ndash3738Santiago Chile 2015

[20] W Ouyang and X Wang ldquoJoint deep learning for pedestriandetectionrdquo in Proceedings of the 14th IEEE International Con-ference on Computer Vision (ICCV rsquo13) pp 2056ndash2063 SydneyAustralia December 2013

[21] A Krizhevsky I Sutskever andG EHinton ldquoImagenet classifi-cation with deep convolutional neural networksrdquo in Proceedingsof the 26th Annual Conference on Neural Information ProcessingSystems (NIPS rsquo12) pp 1097ndash1105 Lake Tahoe Nev USADecember 2012

[22] C Szegedy W Liu Y Jia et al ldquoGoing deeper with convolu-tionsrdquo inProceedings of the IEEEConference onComputer Visionand Pattern Recognition (CVPR rsquo15) pp 1ndash9 IEEE BostonMass USA June 2015

[23] K Simonyan and A Zisserman ldquoVery deep convolutional net-works for large-scale image recognitionrdquo httpsarxivorgabs14091556 2015

[24] C Szegedy V Vanhoucke S Ioffe J Shlens and Z WojnaldquoRethinking the inception architecture for computer visionrdquohttpsarxivorgabs151200567 2015

[25] C Szegedy S Ioffe V Vanhoucke and A Alemi ldquoInception-v4 inception-ResNet and the impact of residual connections onlearningrdquo httpsarxivorgabs160207261 2016

[26] K He X Zhang S Ren and J Sun ldquoDeep residual learning forimage recognitionrdquo in Proceedings of the 2016 IEEE Conference

on Computer Vision and Pattern Recognition CVPR 2016 pp770ndash778 July 2016

[27] G Huang Z Liu K Q Weinberger and L van der MaatenldquoDensely connected convolutional networksrdquo in Proceedingsof the 2017 IEEE Conference on Computer Vision and PatternRecognition (CVPR) Honolulu HI USA 2016

[28] S J Pan and Q Yang ldquoA survey on transfer learningrdquo IEEETransactions on Knowledge and Data Engineering vol 22 no10 pp 1345ndash1359 2010

[29] M Mehdipour Ghazi B Yanikoglu and E Aptoula ldquoPlantidentification using deep neural networks via optimization oftransfer learning parametersrdquo Neurocomputing vol 235 pp228ndash235 2017

[30] S Ruder ldquoAn overview of gradient descent optimization algo-rithmsrdquo httpsarxivorgabs160904747 2017

[31] D P Kingma and J L Ba ldquoAdam a method for stochasticoptimizationrdquo httpsarxivorgabs14126980 2017

[32] S S Sannakki V S Rajpurohit V B Nargund R Arun Kumarand S Prema Yallur ldquoLeaf disease grading by machine visionand fuzzy logicrdquo International Journal of Computer Technologyand Applications vol 2 no 5 pp 1709ndash1716 2011

[33] D Samanta P P Chaudhury and A Ghosh ldquoScab diseasesdetection of potato using image processingrdquo InternationalJournal of Computer Trends and Technology vol 3 pp 109ndash1132012

[34] P J Herrera J Dorado and A Ribeiro ldquoA novel approach forweed type classification based on shape descriptors and a fuzzydecision-making methodrdquo Sensors vol 14 no 8 pp 15304ndash15324 2014

[35] B Cheng and E T Matson ldquoA feature-based machine learningagent for automatic rice andweed discriminationrdquo InternationalConference onArtificial Intelligence and SoftComputing pp 517ndash527 2015

[36] S Sankaran and R Ehsani ldquoComparison of visible-near infra-red and mid-infrared spectroscopy for classification of Huang-longbing and citrus canker infected leavesrdquo Agricultural Engi-neering International CIGR Journal vol 15 no 3 pp 75ndash792013

[37] X Cheng Y Zhang Y Chen Y Wu and Y Yue ldquoPest iden-tification via deep residual learning in complex backgroundrdquoComputers and Electronics in Agriculture vol 141 pp 351ndash3562017

[38] A dos Santos Ferreira D Matte Freitas G Goncalves daSilva H Pistori and M Theophilo Folhes ldquoWeed detection insoybean crops using ConvNetsrdquo Computers and Electronics inAgriculture vol 143 pp 314ndash324 2017

[39] S Sladojevic M Arsenovic A Anderla D Culibrk andD Stefanovic ldquoDeep Neural Networks Based Recognition ofPlant Diseases by Leaf Image Classificationrdquo ComputationalIntelligence and Neuroscience vol 2016 Article ID 3289801 11pages 2016

[40] S P Mohanty D P Hughes and M Salathe ldquoUsing deep learn-ing for image-based plant disease detectionrdquo Frontiers in PlantScience vol 7 article no 1419 2016

[41] I Sa Z Ge F Dayoub B Upcroft T Perez and C McCoolldquoDeepfruits A fruit detection system using deep neural net-worksrdquo Sensors vol 16 article no 1222 no 8 2016

[42] V Nair and G E Hinton ldquoRectified linear units improveRestricted Boltzmann machinesrdquo in Proceedings of the 27thInternational Conference on Machine Learning (ICML rsquo10) pp807ndash814 Haifa Israel June 2010

10 Advances in Multimedia

[43] D P Hughes and M Salathe ldquoAn open access repository ofimages on plant health to enable the development of mobiledisease diagnosticsrdquo httpsarxivorgabs151108060 2016

[44] D Ciresan U Meier J Masci and J Schmidhuber ldquoA com-mittee of neural networks for traffic sign classificationrdquo inProceedings of the 2011 International Joint Conference on NeuralNetworks (IJCNN 2011 - San Jose) San Jose CA USA July 2011

[45] P Pawara E Okafor O Surinta L Schomaker andMWieringldquoComparing Local Descriptors and Bags of Visual Words toDeep Convolutional Neural Networks for Plant Recognitionrdquoin Proceedings of the 6th International Conference on PatternRecognition Applications and Methods pp 479ndash486 PortoPortugal Feburary 2017

[46] M Lin ldquoNetwork inNnetworkrdquo httpsarxivorgabs131244002014

[47] K He X Zhang S Ren and J Sun ldquoIdentity mappings in deepresidual networksrdquo in Proceedings of the European Conferenceon Computer Vision pp 630ndash645 2016

[48] L van der Maaten and G Hinton ldquoVisualizing data using t-SNErdquo Journal of Machine Learning Research vol 9 pp 2579ndash2625 2008

International Journal of

AerospaceEngineeringHindawiwwwhindawicom Volume 2018

RoboticsJournal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Active and Passive Electronic Components

VLSI Design

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Shock and Vibration

Hindawiwwwhindawicom Volume 2018

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawiwwwhindawicom

Volume 2018

Hindawi Publishing Corporation httpwwwhindawicom Volume 2013Hindawiwwwhindawicom

The Scientific World Journal

Volume 2018

Control Scienceand Engineering

Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom

Journal ofEngineeringVolume 2018

SensorsJournal of

Hindawiwwwhindawicom Volume 2018

International Journal of

RotatingMachinery

Hindawiwwwhindawicom Volume 2018

Modelling ampSimulationin EngineeringHindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Navigation and Observation

International Journal of

Hindawi

wwwhindawicom Volume 2018

Advances in

Multimedia

Submit your manuscripts atwwwhindawicom

Page 6: ResearchArticle Can Deep Learning Identify Tomato Leaf

6 Advances in Multimedia

Inputimage

Conv1 Conv2

MaxPool MaxPool

MaxPool

MaxPool

Inception 3a

Inception 3b

Inception 4a

Inception 4b

Inception 4c

Inception 4d

Inception 4e

Inception 5a

Inception 5b

AvgPool

Dropout40

FC

somax

output

Figure 4 The architecture of GoogLeNet [22 45]

1times1 conv 64

3times3 conv 64

1times1 conv 256

relu

relu

relu

+

Figure 5 ResNet bottleneck residual building block [26]

Table 3 Model recognition accuracy

Model AccuracyAlexNet (SGD) 9583AlexNet (Adam) 1386GoogLeNet (SGD) 9566GoogLeNet (Adam) 9406ResNet (SGD) 9651ResNet (Adam) 9439

AlexNet and GoogLeNet is similar and both inferior to theResNet

42 Experiments on Batch Size and Number of IterationsFrom the experiment on optimization methods the ResNetobtains the highest classification accuracy Next we evaluatedthe effects of batch size and the number of iterations on theperformance of the ResNet The batch size was set to 16 32and 64 respectively Meanwhile the number of iterations wasset to 2496 4992 and 9984 The classification accuracy ofdifferent training scenarios is given in Table 4 At the sametime the classification accuracy of each labels representativeleaf disease category (See Table 1) is given In this experimentthe initial learning rate was set to 0001 and dropped by afactor of 05 every 2496 iterations

In Table 4 the best overall classification accuracy 9719is got by the ResNet combining with batch size 16 and

Trai

ning

Los

s

Number of Iterations

AlexNetGoogLeNetResNet

3

25

2

15

1

05

00 1000 2000 3000 4000 5000 6000

Figure 6 The training loss during the fine-tuning process

iterations 4992 As shown in Table 4 whether increasingthe number of iterations or batch size the performance ofcorresponding models has not been improved significantlyin identifying tomato leaf disease A small batch size witha medium number of iterations is quite effective in thiswork Moreover a larger batch size and number of iterationsincreases the training duration We have not tried higher orlower values for the attempted parameters since differentclassification task may have various suitable parameters andit is hard to give a certain rule in setting hyperparameters

43 Experiments on Full Training and Fine-Tuning of ResNetThis section is designed for exploring the performance ofCNN by changing the structure of the models In practicala deep CNN always owns a large size which means a largenumber of parameters Thus full training of a deep CNNrequires extensive computational resources and is time-consuming In addition full training of a deep CNN mayled to overfitting when the training data is limited So wecompared the performance of the pretrained CNN throughfull training and fine-tuning their structures

We changed the structure of ResNet and combination ofthe best parameters from the front experiments was utilized

Advances in Multimedia 7

Table 4 Classification accuracies with different parameters during fine-tuning of the ResNetThe numbers inside parenthesis indicate batchsize and number of iterations

Networks label1 label2 label3 label4 label5 label6 label7 label8 label9 overallResNet (162496) 9091 8889 100 100 9688 100 9028 8820 9755 9498ResNet (164992) 9818 9877 100 9862 9688 100 9653 8882 9877 9719ResNet (169984) 9818 9753 100 9862 9688 100 9722 8696 9877 9694ResNet (322496) 9727 9506 100 9793 9688 100 9514 8634 9939 9634ResNet (324992) 9727 9506 100 9724 9688 100 9653 8696 9939 9651ResNet (329984) 9636 9630 100 9931 9688 100 9444 8820 9877 9660ResNet (642496) 9364 9383 100 9724 9688 100 9444 8758 9939 9592ResNet (644992) 9455 9529 100 9655 9583 100 9583 8696 9939 9583ResNet (649984) 9545 9383 100 9793 9688 100 9444 8758 9939 9617

ResNet50 has 177 layers if the layers for each building blockand connection are calculated In this experiment the lastthree layers of ResNet were modified to a fully connectedlayer (denoted as ldquofcrdquo) a softmax layer and a classificationlayer and the fully connected layer owns 9 neurons Thestructure was changed by freezing the weights of a certainnumber of layers in the network by setting the learning ratein those layers to zero During training the parameters ofthe frozen layers are not updated Full training and fine-tuning are defined by the number of training layers ie fulltraining (1-ldquofcrdquo) fine-tuning (37-ldquofcrdquo 79-ldquofcrdquo 111-ldquofcrdquo 141-ldquofcrdquo 163-ldquofcrdquo) The accuracy and training time of differentnetwork structure are presented in Table 5 At first the batchsize and 4992 iterations were combined the initial learningrate was set to 0001 and dropped by a factor of 01 every2496 iterations In order to get more convincing conclusionsResNet (16 9984) which gets the second place in Table 4 wasalso used to execute the experiments

In Table 5 the accuracy and training time of differentnetwork structures are presented In two cases ie the 4992iterations and 9984 iterations of ResNet the accuracy of themodel from the 37 layer fine-tuning structure are higherthan that of the full training model In the case where thenumber of iterations is 4992 the accuracy of the model fromthe 79 layer fine-tuning structure is equal to that of the fulltraining modelThe final column of the Table 5 represents thetraining time of the corresponding network and it is clearthat the training time of the fine-tuning models is greatlylowered than the full training model Because the gradientsof the frozen layers do not need to be computed freezingthe weights of initial layers can speed up network trainingWe observe that the moderate fine-tuning models (37-ldquofcrdquo79-ldquofcrdquo 111-ldquofcrdquo) always led to a performance superior orapproximately equal to the full training models Thus wesuggest that for practical application the moderate fine-tuning models may be a good choice Especially for theresearcher who holds massive data the fine-tuning modelsmay achieve good performance while saving computationalresources and time

Moreover the features of the final fully connected layerof ResNet (16 4992 37-ldquofcrdquo) were examined by utilizingthe t-distributed Stochastic Neighbour Embedding (t-SNE)algorithm (see Figure 7) [48] 1176 test images were used to

label1 label2 label3 label4 label5

label6 label7 label8 label9

Figure 7 Two-dimensional scatter plot of high-dimensional fea-tures generated with t-SNE

extract the features In Figure 7 different colors representdifferent labels the corresponding disease categories of thelabels were listed in Table 1 As shown in Figure 7 9 differentcolor points are clearly separated which indicates that thefeatures learned from the ResNet with the optimal structurecan be used to classify the tomato leaf disease precisely

5 Conclusion

This paper concentrates on identifying tomato leaf diseaseusing deep convolutional neural networks by transfer learn-ing The utilized networks are based on the pretrained deeplearning models of AlexNet GoogLeNet and ResNet Firstwe compared the relative performance of these networksby using SGD and Adam optimization method revealingthat the ResNet with SGD optimization method obtainsthe highest result with the best accuracy 9651 Thenthe performance evaluation of batch size and number of

8 Advances in Multimedia

Table 5 Accuracies and training time in different network structuresThe values inside parenthesis denote batch size number of iterationsand training layers

Network topology Accuracy Time (minsec)ResNet (16 4992 1-ldquofcrdquo) 9643 59min 30secResNet (16 4992 37-ldquofcrdquo) 9728 44min 13secResNet (16 4992 79-ldquofcrdquo) 9643 37min 27secResNet (16 4992 111-ldquofcrdquo) 9575 30min 6secResNet (16 4992 141-ldquofcrdquo) 9532 24min 15secResNet (16 4992 163-ldquofcrdquo) 9269 19min 31secResNet (16 9984 1-ldquofcrdquo) 9694 118min 32secResNet (16 9984 37-ldquofcrdquo) 9702 92min 53secResNet (16 9984 79-ldquofcrdquo) 9677 72min 23secResNet (16 9984 111-ldquofcrdquo) 9626 58min 40secResNet (16 9984 141-ldquofcrdquo) 9575 47min 22secResNet (16 9984 163-ldquofcrdquo) 9396 39min 32sec

iterations affecting the transfer learning of the ResNet wasconducted A small batch size of 16 combining a moderatenumber of iterations of 4992 is the optimal choice in thiswork Our findings suggest that for a particular task neitherlarge batch size nor large number of iterations may improvethe accuracy of the target modelThe setting of batch size andnumber of iterations depends on your data set and the utilizednetwork Next the best combined model was used to fine-tune the structure Fine-tuning ResNet layers from 37 to ldquofcrdquoobtained the highest accuracy 9728 in identifying tomatoleaf disease Based on the amount of available data layer-wisefine-tuning may provide a practical way to achieve the bestperformance of the application at hand We believe that theresults obtained in this work will bring some inspiration toother similar visual recognition problems and the practicalstudy of this work can be easily extended to other plant leafdisease identification problems

Data Availability

Thetomato leaf data supporting this work are frompreviouslyreported studies which have been cited The processed dataare available from the corresponding author request

Conflicts of Interest

The authors declare that there are no conflicts of interestregarding the publication of this paper

Acknowledgments

This study was supported by the National Science and tech-nology support program (2014BAD12B01-1-3) PublicWelfareIndustry (Agriculture) Research Projects Level-2 (201503116-04-06) Postdoctoral Foundation of Heilongjiang Province(LBHZ15020) Harbin Applied Technology Research andDevelopment Program (2017RAQXJ096) and EconomicDecision Making and Early Warning of Soybean Industryin Technology Collaborative Innovation System of SoybeanIndustry in Heilongjiang Province (20170401)

References

[1] RChaerani andR EVoorrips ldquoTomato early blight (Alternariasolani) The pathogen genetics and breeding for resistancerdquoJournal of General Plant Pathology vol 72 no 6 pp 335ndash3472006

[2] A M Dickey L S Osborne and C L Mckenzie ldquoPapaya(Carica papaya Brassicales Caricaceae) is not a host plant oftomato yellow leaf curl virus (TYLCV family Geminiviridaegenus Begomovirus)rdquo Florida Entomologist vol 95 no 1 pp211ndash213 2012

[3] G Wei L Baoju S Yanxia and X Xuewen ldquoStudies on path-ogenicity differentiation of corynespora cassiicola isolatesagainst cucumber tomato and eggplantrdquo Acta HorticulturaeSinica vol 38 no 3 pp 465ndash470 2011

[4] P Lindhout W Korta M Cislik I Vos and T GerlaghldquoFurther identification of races of Cladosporium fulvum (Fulviafulva) on tomato originating from the Netherlands France andPolandrdquo Netherlands Journal of Plant Pathology vol 95 no 3pp 143ndash148 1989

[5] K Kubota S Tsuda A Tamai and T Meshi ldquoTomato mosaicvirus replication protein suppresses virus-targeted posttran-scriptional gene silencingrdquo Journal of Virology vol 77 no 20pp 11016ndash11026 2003

[6] M Tian B Benedetti and S Kamoun ldquoA second Kazal-like protease inhibitor from Phytophthora infestans inhibitsand interacts with the apoplastic pathogenesis-related proteaseP69B of tomatordquo Plant Physiology vol 138 no 3 pp 1785ndash17932005

[7] L E Blum ldquoReduction of incidence and severity of Septorialycopersici leaf spot of tomatowith bacteria and yeastsrdquoCienciaRural vol 30 no 5 pp 761ndash765 2000

[8] E A Chatzivasileiadis and M W Sabelis ldquoToxicity of methylketones from tomato trichomes to Tetranychus urticae KochrdquoExperimental and Applied Acarology vol 21 no 6-7 pp 473ndash484 1997

[9] M Anthimopoulos S Christodoulidis L Ebner A Christeand S Mougiakakou ldquoLung Pattern Classification for Inter-stitial Lung Diseases Using a Deep Convolutional NeuralNetworkrdquo IEEE Transactions on Medical Imaging vol 35 no5 pp 1207ndash1216 2016

[10] Y Tang and X Wu ldquoText-independent writer identification viaCNN features and joint Bayesianrdquo in Proceedings of the 15th

Advances in Multimedia 9

International Conference on Frontiers in Handwriting Recogni-tion ICFHR 2016 pp 566ndash571 Shenzhen China October 2016

[11] Y Tang and XWu ldquoSaliency Detection via Combining Region-Level and Pixel-Level Predictions with CNNsrdquo in Proceedings ofthe EuropeanConference onComputer Vision (ECCV) pp 1608ndash05186 2016

[12] Y Tang and X Wu ldquoSalient object detection with chainedmulti-scale fully convolutional networkrdquo ACM Multimedia(ACMMM) pp 618ndash626 2017

[13] Y Tang and X Wu ldquoScene text detection and segmentationbased on cascaded convolution neural networksrdquo IEEE Trans-actions on Image Processing vol 26 no 3 pp 1509ndash1520 2017

[14] Y Tang and X Wu ldquoScene Text Detection using Superpixelbased Stroke Feature Transform and Deep Learning basedRegion Classificationrdquo IEEE Transactions on Multimedia vol20 no 9 pp 2276ndash2288 2018

[15] Y Yao XWu Z Lei S Shan andW Zuo ldquoJoint Representationand Truncated Inference Learning for Correlation Filter basedTrackingrdquo in Proceedings of the European Conference on Com-puter Vision (ECCV) pp 1ndash14 2018

[16] L Zhang F Yang Y Daniel Zhang and Y J Zhu ldquoRoadcrack detection using deep convolutional neural networkrdquoin Proceedings of the 23rd IEEE International Conference onImage Processing ICIP 2016 pp 3708ndash3712 Phoenix AZ USASeptember 2016

[17] D Xie L Zhang andL Bai ldquoDeep learning in visual computingand signal processingrdquo Applied Computational Intelligence andSoft Computing vol 2017 Article ID 1320780 13 pages 2017

[18] Z Zhou J Shin L Zhang S Gurudu M Gotway and J LiangldquoFine-tuning convolutional neural networks for biomedicalimage analysis Actively and incrementallyrdquo inProceedings of the30th IEEE Conference on Computer Vision and Pattern Recog-nition CVPR 2017 pp 4761ndash4772 USA July 2017

[19] Z Liu P Luo X Wang and X Tang ldquoDeep learning face attri-butes in the wildrdquo in Proceedings of the 15th IEEE InternationalConference on Computer Vision ICCV 2015 pp 3730ndash3738Santiago Chile 2015

[20] W Ouyang and X Wang ldquoJoint deep learning for pedestriandetectionrdquo in Proceedings of the 14th IEEE International Con-ference on Computer Vision (ICCV rsquo13) pp 2056ndash2063 SydneyAustralia December 2013

[21] A Krizhevsky I Sutskever andG EHinton ldquoImagenet classifi-cation with deep convolutional neural networksrdquo in Proceedingsof the 26th Annual Conference on Neural Information ProcessingSystems (NIPS rsquo12) pp 1097ndash1105 Lake Tahoe Nev USADecember 2012

[22] C Szegedy W Liu Y Jia et al ldquoGoing deeper with convolu-tionsrdquo inProceedings of the IEEEConference onComputer Visionand Pattern Recognition (CVPR rsquo15) pp 1ndash9 IEEE BostonMass USA June 2015

[23] K Simonyan and A Zisserman ldquoVery deep convolutional net-works for large-scale image recognitionrdquo httpsarxivorgabs14091556 2015

[24] C Szegedy V Vanhoucke S Ioffe J Shlens and Z WojnaldquoRethinking the inception architecture for computer visionrdquohttpsarxivorgabs151200567 2015

[25] C Szegedy S Ioffe V Vanhoucke and A Alemi ldquoInception-v4 inception-ResNet and the impact of residual connections onlearningrdquo httpsarxivorgabs160207261 2016

[26] K He X Zhang S Ren and J Sun ldquoDeep residual learning forimage recognitionrdquo in Proceedings of the 2016 IEEE Conference

on Computer Vision and Pattern Recognition CVPR 2016 pp770ndash778 July 2016

[27] G Huang Z Liu K Q Weinberger and L van der MaatenldquoDensely connected convolutional networksrdquo in Proceedingsof the 2017 IEEE Conference on Computer Vision and PatternRecognition (CVPR) Honolulu HI USA 2016

[28] S J Pan and Q Yang ldquoA survey on transfer learningrdquo IEEETransactions on Knowledge and Data Engineering vol 22 no10 pp 1345ndash1359 2010

[29] M Mehdipour Ghazi B Yanikoglu and E Aptoula ldquoPlantidentification using deep neural networks via optimization oftransfer learning parametersrdquo Neurocomputing vol 235 pp228ndash235 2017

[30] S Ruder ldquoAn overview of gradient descent optimization algo-rithmsrdquo httpsarxivorgabs160904747 2017

[31] D P Kingma and J L Ba ldquoAdam a method for stochasticoptimizationrdquo httpsarxivorgabs14126980 2017

[32] S S Sannakki V S Rajpurohit V B Nargund R Arun Kumarand S Prema Yallur ldquoLeaf disease grading by machine visionand fuzzy logicrdquo International Journal of Computer Technologyand Applications vol 2 no 5 pp 1709ndash1716 2011

[33] D Samanta P P Chaudhury and A Ghosh ldquoScab diseasesdetection of potato using image processingrdquo InternationalJournal of Computer Trends and Technology vol 3 pp 109ndash1132012

[34] P J Herrera J Dorado and A Ribeiro ldquoA novel approach forweed type classification based on shape descriptors and a fuzzydecision-making methodrdquo Sensors vol 14 no 8 pp 15304ndash15324 2014

[35] B Cheng and E T Matson ldquoA feature-based machine learningagent for automatic rice andweed discriminationrdquo InternationalConference onArtificial Intelligence and SoftComputing pp 517ndash527 2015

[36] S Sankaran and R Ehsani ldquoComparison of visible-near infra-red and mid-infrared spectroscopy for classification of Huang-longbing and citrus canker infected leavesrdquo Agricultural Engi-neering International CIGR Journal vol 15 no 3 pp 75ndash792013

[37] X Cheng Y Zhang Y Chen Y Wu and Y Yue ldquoPest iden-tification via deep residual learning in complex backgroundrdquoComputers and Electronics in Agriculture vol 141 pp 351ndash3562017

[38] A dos Santos Ferreira D Matte Freitas G Goncalves daSilva H Pistori and M Theophilo Folhes ldquoWeed detection insoybean crops using ConvNetsrdquo Computers and Electronics inAgriculture vol 143 pp 314ndash324 2017

[39] S Sladojevic M Arsenovic A Anderla D Culibrk andD Stefanovic ldquoDeep Neural Networks Based Recognition ofPlant Diseases by Leaf Image Classificationrdquo ComputationalIntelligence and Neuroscience vol 2016 Article ID 3289801 11pages 2016

[40] S P Mohanty D P Hughes and M Salathe ldquoUsing deep learn-ing for image-based plant disease detectionrdquo Frontiers in PlantScience vol 7 article no 1419 2016

[41] I Sa Z Ge F Dayoub B Upcroft T Perez and C McCoolldquoDeepfruits A fruit detection system using deep neural net-worksrdquo Sensors vol 16 article no 1222 no 8 2016

[42] V Nair and G E Hinton ldquoRectified linear units improveRestricted Boltzmann machinesrdquo in Proceedings of the 27thInternational Conference on Machine Learning (ICML rsquo10) pp807ndash814 Haifa Israel June 2010

10 Advances in Multimedia

[43] D P Hughes and M Salathe ldquoAn open access repository ofimages on plant health to enable the development of mobiledisease diagnosticsrdquo httpsarxivorgabs151108060 2016

[44] D Ciresan U Meier J Masci and J Schmidhuber ldquoA com-mittee of neural networks for traffic sign classificationrdquo inProceedings of the 2011 International Joint Conference on NeuralNetworks (IJCNN 2011 - San Jose) San Jose CA USA July 2011

[45] P Pawara E Okafor O Surinta L Schomaker andMWieringldquoComparing Local Descriptors and Bags of Visual Words toDeep Convolutional Neural Networks for Plant Recognitionrdquoin Proceedings of the 6th International Conference on PatternRecognition Applications and Methods pp 479ndash486 PortoPortugal Feburary 2017

[46] M Lin ldquoNetwork inNnetworkrdquo httpsarxivorgabs131244002014

[47] K He X Zhang S Ren and J Sun ldquoIdentity mappings in deepresidual networksrdquo in Proceedings of the European Conferenceon Computer Vision pp 630ndash645 2016

[48] L van der Maaten and G Hinton ldquoVisualizing data using t-SNErdquo Journal of Machine Learning Research vol 9 pp 2579ndash2625 2008

International Journal of

AerospaceEngineeringHindawiwwwhindawicom Volume 2018

RoboticsJournal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Active and Passive Electronic Components

VLSI Design

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Shock and Vibration

Hindawiwwwhindawicom Volume 2018

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawiwwwhindawicom

Volume 2018

Hindawi Publishing Corporation httpwwwhindawicom Volume 2013Hindawiwwwhindawicom

The Scientific World Journal

Volume 2018

Control Scienceand Engineering

Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom

Journal ofEngineeringVolume 2018

SensorsJournal of

Hindawiwwwhindawicom Volume 2018

International Journal of

RotatingMachinery

Hindawiwwwhindawicom Volume 2018

Modelling ampSimulationin EngineeringHindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Navigation and Observation

International Journal of

Hindawi

wwwhindawicom Volume 2018

Advances in

Multimedia

Submit your manuscripts atwwwhindawicom

Page 7: ResearchArticle Can Deep Learning Identify Tomato Leaf

Advances in Multimedia 7

Table 4 Classification accuracies with different parameters during fine-tuning of the ResNetThe numbers inside parenthesis indicate batchsize and number of iterations

Networks label1 label2 label3 label4 label5 label6 label7 label8 label9 overallResNet (162496) 9091 8889 100 100 9688 100 9028 8820 9755 9498ResNet (164992) 9818 9877 100 9862 9688 100 9653 8882 9877 9719ResNet (169984) 9818 9753 100 9862 9688 100 9722 8696 9877 9694ResNet (322496) 9727 9506 100 9793 9688 100 9514 8634 9939 9634ResNet (324992) 9727 9506 100 9724 9688 100 9653 8696 9939 9651ResNet (329984) 9636 9630 100 9931 9688 100 9444 8820 9877 9660ResNet (642496) 9364 9383 100 9724 9688 100 9444 8758 9939 9592ResNet (644992) 9455 9529 100 9655 9583 100 9583 8696 9939 9583ResNet (649984) 9545 9383 100 9793 9688 100 9444 8758 9939 9617

ResNet50 has 177 layers if the layers for each building blockand connection are calculated In this experiment the lastthree layers of ResNet were modified to a fully connectedlayer (denoted as ldquofcrdquo) a softmax layer and a classificationlayer and the fully connected layer owns 9 neurons Thestructure was changed by freezing the weights of a certainnumber of layers in the network by setting the learning ratein those layers to zero During training the parameters ofthe frozen layers are not updated Full training and fine-tuning are defined by the number of training layers ie fulltraining (1-ldquofcrdquo) fine-tuning (37-ldquofcrdquo 79-ldquofcrdquo 111-ldquofcrdquo 141-ldquofcrdquo 163-ldquofcrdquo) The accuracy and training time of differentnetwork structure are presented in Table 5 At first the batchsize and 4992 iterations were combined the initial learningrate was set to 0001 and dropped by a factor of 01 every2496 iterations In order to get more convincing conclusionsResNet (16 9984) which gets the second place in Table 4 wasalso used to execute the experiments

In Table 5 the accuracy and training time of differentnetwork structures are presented In two cases ie the 4992iterations and 9984 iterations of ResNet the accuracy of themodel from the 37 layer fine-tuning structure are higherthan that of the full training model In the case where thenumber of iterations is 4992 the accuracy of the model fromthe 79 layer fine-tuning structure is equal to that of the fulltraining modelThe final column of the Table 5 represents thetraining time of the corresponding network and it is clearthat the training time of the fine-tuning models is greatlylowered than the full training model Because the gradientsof the frozen layers do not need to be computed freezingthe weights of initial layers can speed up network trainingWe observe that the moderate fine-tuning models (37-ldquofcrdquo79-ldquofcrdquo 111-ldquofcrdquo) always led to a performance superior orapproximately equal to the full training models Thus wesuggest that for practical application the moderate fine-tuning models may be a good choice Especially for theresearcher who holds massive data the fine-tuning modelsmay achieve good performance while saving computationalresources and time

Moreover the features of the final fully connected layerof ResNet (16 4992 37-ldquofcrdquo) were examined by utilizingthe t-distributed Stochastic Neighbour Embedding (t-SNE)algorithm (see Figure 7) [48] 1176 test images were used to

label1 label2 label3 label4 label5

label6 label7 label8 label9

Figure 7 Two-dimensional scatter plot of high-dimensional fea-tures generated with t-SNE

extract the features In Figure 7 different colors representdifferent labels the corresponding disease categories of thelabels were listed in Table 1 As shown in Figure 7 9 differentcolor points are clearly separated which indicates that thefeatures learned from the ResNet with the optimal structurecan be used to classify the tomato leaf disease precisely

5 Conclusion

This paper concentrates on identifying tomato leaf diseaseusing deep convolutional neural networks by transfer learn-ing The utilized networks are based on the pretrained deeplearning models of AlexNet GoogLeNet and ResNet Firstwe compared the relative performance of these networksby using SGD and Adam optimization method revealingthat the ResNet with SGD optimization method obtainsthe highest result with the best accuracy 9651 Thenthe performance evaluation of batch size and number of

8 Advances in Multimedia

Table 5 Accuracies and training time in different network structuresThe values inside parenthesis denote batch size number of iterationsand training layers

Network topology Accuracy Time (minsec)ResNet (16 4992 1-ldquofcrdquo) 9643 59min 30secResNet (16 4992 37-ldquofcrdquo) 9728 44min 13secResNet (16 4992 79-ldquofcrdquo) 9643 37min 27secResNet (16 4992 111-ldquofcrdquo) 9575 30min 6secResNet (16 4992 141-ldquofcrdquo) 9532 24min 15secResNet (16 4992 163-ldquofcrdquo) 9269 19min 31secResNet (16 9984 1-ldquofcrdquo) 9694 118min 32secResNet (16 9984 37-ldquofcrdquo) 9702 92min 53secResNet (16 9984 79-ldquofcrdquo) 9677 72min 23secResNet (16 9984 111-ldquofcrdquo) 9626 58min 40secResNet (16 9984 141-ldquofcrdquo) 9575 47min 22secResNet (16 9984 163-ldquofcrdquo) 9396 39min 32sec

iterations affecting the transfer learning of the ResNet wasconducted A small batch size of 16 combining a moderatenumber of iterations of 4992 is the optimal choice in thiswork Our findings suggest that for a particular task neitherlarge batch size nor large number of iterations may improvethe accuracy of the target modelThe setting of batch size andnumber of iterations depends on your data set and the utilizednetwork Next the best combined model was used to fine-tune the structure Fine-tuning ResNet layers from 37 to ldquofcrdquoobtained the highest accuracy 9728 in identifying tomatoleaf disease Based on the amount of available data layer-wisefine-tuning may provide a practical way to achieve the bestperformance of the application at hand We believe that theresults obtained in this work will bring some inspiration toother similar visual recognition problems and the practicalstudy of this work can be easily extended to other plant leafdisease identification problems

Data Availability

Thetomato leaf data supporting this work are frompreviouslyreported studies which have been cited The processed dataare available from the corresponding author request

Conflicts of Interest

The authors declare that there are no conflicts of interestregarding the publication of this paper

Acknowledgments

This study was supported by the National Science and tech-nology support program (2014BAD12B01-1-3) PublicWelfareIndustry (Agriculture) Research Projects Level-2 (201503116-04-06) Postdoctoral Foundation of Heilongjiang Province(LBHZ15020) Harbin Applied Technology Research andDevelopment Program (2017RAQXJ096) and EconomicDecision Making and Early Warning of Soybean Industryin Technology Collaborative Innovation System of SoybeanIndustry in Heilongjiang Province (20170401)

References

[1] RChaerani andR EVoorrips ldquoTomato early blight (Alternariasolani) The pathogen genetics and breeding for resistancerdquoJournal of General Plant Pathology vol 72 no 6 pp 335ndash3472006

[2] A M Dickey L S Osborne and C L Mckenzie ldquoPapaya(Carica papaya Brassicales Caricaceae) is not a host plant oftomato yellow leaf curl virus (TYLCV family Geminiviridaegenus Begomovirus)rdquo Florida Entomologist vol 95 no 1 pp211ndash213 2012

[3] G Wei L Baoju S Yanxia and X Xuewen ldquoStudies on path-ogenicity differentiation of corynespora cassiicola isolatesagainst cucumber tomato and eggplantrdquo Acta HorticulturaeSinica vol 38 no 3 pp 465ndash470 2011

[4] P Lindhout W Korta M Cislik I Vos and T GerlaghldquoFurther identification of races of Cladosporium fulvum (Fulviafulva) on tomato originating from the Netherlands France andPolandrdquo Netherlands Journal of Plant Pathology vol 95 no 3pp 143ndash148 1989

[5] K Kubota S Tsuda A Tamai and T Meshi ldquoTomato mosaicvirus replication protein suppresses virus-targeted posttran-scriptional gene silencingrdquo Journal of Virology vol 77 no 20pp 11016ndash11026 2003

[6] M Tian B Benedetti and S Kamoun ldquoA second Kazal-like protease inhibitor from Phytophthora infestans inhibitsand interacts with the apoplastic pathogenesis-related proteaseP69B of tomatordquo Plant Physiology vol 138 no 3 pp 1785ndash17932005

[7] L E Blum ldquoReduction of incidence and severity of Septorialycopersici leaf spot of tomatowith bacteria and yeastsrdquoCienciaRural vol 30 no 5 pp 761ndash765 2000

[8] E A Chatzivasileiadis and M W Sabelis ldquoToxicity of methylketones from tomato trichomes to Tetranychus urticae KochrdquoExperimental and Applied Acarology vol 21 no 6-7 pp 473ndash484 1997

[9] M Anthimopoulos S Christodoulidis L Ebner A Christeand S Mougiakakou ldquoLung Pattern Classification for Inter-stitial Lung Diseases Using a Deep Convolutional NeuralNetworkrdquo IEEE Transactions on Medical Imaging vol 35 no5 pp 1207ndash1216 2016

[10] Y Tang and X Wu ldquoText-independent writer identification viaCNN features and joint Bayesianrdquo in Proceedings of the 15th

Advances in Multimedia 9

International Conference on Frontiers in Handwriting Recogni-tion ICFHR 2016 pp 566ndash571 Shenzhen China October 2016

[11] Y Tang and XWu ldquoSaliency Detection via Combining Region-Level and Pixel-Level Predictions with CNNsrdquo in Proceedings ofthe EuropeanConference onComputer Vision (ECCV) pp 1608ndash05186 2016

[12] Y Tang and X Wu ldquoSalient object detection with chainedmulti-scale fully convolutional networkrdquo ACM Multimedia(ACMMM) pp 618ndash626 2017

[13] Y Tang and X Wu ldquoScene text detection and segmentationbased on cascaded convolution neural networksrdquo IEEE Trans-actions on Image Processing vol 26 no 3 pp 1509ndash1520 2017

[14] Y Tang and X Wu ldquoScene Text Detection using Superpixelbased Stroke Feature Transform and Deep Learning basedRegion Classificationrdquo IEEE Transactions on Multimedia vol20 no 9 pp 2276ndash2288 2018

[15] Y Yao XWu Z Lei S Shan andW Zuo ldquoJoint Representationand Truncated Inference Learning for Correlation Filter basedTrackingrdquo in Proceedings of the European Conference on Com-puter Vision (ECCV) pp 1ndash14 2018

[16] L Zhang F Yang Y Daniel Zhang and Y J Zhu ldquoRoadcrack detection using deep convolutional neural networkrdquoin Proceedings of the 23rd IEEE International Conference onImage Processing ICIP 2016 pp 3708ndash3712 Phoenix AZ USASeptember 2016

[17] D Xie L Zhang andL Bai ldquoDeep learning in visual computingand signal processingrdquo Applied Computational Intelligence andSoft Computing vol 2017 Article ID 1320780 13 pages 2017

[18] Z Zhou J Shin L Zhang S Gurudu M Gotway and J LiangldquoFine-tuning convolutional neural networks for biomedicalimage analysis Actively and incrementallyrdquo inProceedings of the30th IEEE Conference on Computer Vision and Pattern Recog-nition CVPR 2017 pp 4761ndash4772 USA July 2017

[19] Z Liu P Luo X Wang and X Tang ldquoDeep learning face attri-butes in the wildrdquo in Proceedings of the 15th IEEE InternationalConference on Computer Vision ICCV 2015 pp 3730ndash3738Santiago Chile 2015

[20] W Ouyang and X Wang ldquoJoint deep learning for pedestriandetectionrdquo in Proceedings of the 14th IEEE International Con-ference on Computer Vision (ICCV rsquo13) pp 2056ndash2063 SydneyAustralia December 2013

[21] A Krizhevsky I Sutskever andG EHinton ldquoImagenet classifi-cation with deep convolutional neural networksrdquo in Proceedingsof the 26th Annual Conference on Neural Information ProcessingSystems (NIPS rsquo12) pp 1097ndash1105 Lake Tahoe Nev USADecember 2012

[22] C Szegedy W Liu Y Jia et al ldquoGoing deeper with convolu-tionsrdquo inProceedings of the IEEEConference onComputer Visionand Pattern Recognition (CVPR rsquo15) pp 1ndash9 IEEE BostonMass USA June 2015

[23] K Simonyan and A Zisserman ldquoVery deep convolutional net-works for large-scale image recognitionrdquo httpsarxivorgabs14091556 2015

[24] C Szegedy V Vanhoucke S Ioffe J Shlens and Z WojnaldquoRethinking the inception architecture for computer visionrdquohttpsarxivorgabs151200567 2015

[25] C Szegedy S Ioffe V Vanhoucke and A Alemi ldquoInception-v4 inception-ResNet and the impact of residual connections onlearningrdquo httpsarxivorgabs160207261 2016

[26] K He X Zhang S Ren and J Sun ldquoDeep residual learning forimage recognitionrdquo in Proceedings of the 2016 IEEE Conference

on Computer Vision and Pattern Recognition CVPR 2016 pp770ndash778 July 2016

[27] G Huang Z Liu K Q Weinberger and L van der MaatenldquoDensely connected convolutional networksrdquo in Proceedingsof the 2017 IEEE Conference on Computer Vision and PatternRecognition (CVPR) Honolulu HI USA 2016

[28] S J Pan and Q Yang ldquoA survey on transfer learningrdquo IEEETransactions on Knowledge and Data Engineering vol 22 no10 pp 1345ndash1359 2010

[29] M Mehdipour Ghazi B Yanikoglu and E Aptoula ldquoPlantidentification using deep neural networks via optimization oftransfer learning parametersrdquo Neurocomputing vol 235 pp228ndash235 2017

[30] S Ruder ldquoAn overview of gradient descent optimization algo-rithmsrdquo httpsarxivorgabs160904747 2017

[31] D P Kingma and J L Ba ldquoAdam a method for stochasticoptimizationrdquo httpsarxivorgabs14126980 2017

[32] S S Sannakki V S Rajpurohit V B Nargund R Arun Kumarand S Prema Yallur ldquoLeaf disease grading by machine visionand fuzzy logicrdquo International Journal of Computer Technologyand Applications vol 2 no 5 pp 1709ndash1716 2011

[33] D Samanta P P Chaudhury and A Ghosh ldquoScab diseasesdetection of potato using image processingrdquo InternationalJournal of Computer Trends and Technology vol 3 pp 109ndash1132012

[34] P J Herrera J Dorado and A Ribeiro ldquoA novel approach forweed type classification based on shape descriptors and a fuzzydecision-making methodrdquo Sensors vol 14 no 8 pp 15304ndash15324 2014

[35] B Cheng and E T Matson ldquoA feature-based machine learningagent for automatic rice andweed discriminationrdquo InternationalConference onArtificial Intelligence and SoftComputing pp 517ndash527 2015

[36] S Sankaran and R Ehsani ldquoComparison of visible-near infra-red and mid-infrared spectroscopy for classification of Huang-longbing and citrus canker infected leavesrdquo Agricultural Engi-neering International CIGR Journal vol 15 no 3 pp 75ndash792013

[37] X Cheng Y Zhang Y Chen Y Wu and Y Yue ldquoPest iden-tification via deep residual learning in complex backgroundrdquoComputers and Electronics in Agriculture vol 141 pp 351ndash3562017

[38] A dos Santos Ferreira D Matte Freitas G Goncalves daSilva H Pistori and M Theophilo Folhes ldquoWeed detection insoybean crops using ConvNetsrdquo Computers and Electronics inAgriculture vol 143 pp 314ndash324 2017

[39] S Sladojevic M Arsenovic A Anderla D Culibrk andD Stefanovic ldquoDeep Neural Networks Based Recognition ofPlant Diseases by Leaf Image Classificationrdquo ComputationalIntelligence and Neuroscience vol 2016 Article ID 3289801 11pages 2016

[40] S P Mohanty D P Hughes and M Salathe ldquoUsing deep learn-ing for image-based plant disease detectionrdquo Frontiers in PlantScience vol 7 article no 1419 2016

[41] I Sa Z Ge F Dayoub B Upcroft T Perez and C McCoolldquoDeepfruits A fruit detection system using deep neural net-worksrdquo Sensors vol 16 article no 1222 no 8 2016

[42] V Nair and G E Hinton ldquoRectified linear units improveRestricted Boltzmann machinesrdquo in Proceedings of the 27thInternational Conference on Machine Learning (ICML rsquo10) pp807ndash814 Haifa Israel June 2010

10 Advances in Multimedia

[43] D P Hughes and M Salathe ldquoAn open access repository ofimages on plant health to enable the development of mobiledisease diagnosticsrdquo httpsarxivorgabs151108060 2016

[44] D Ciresan U Meier J Masci and J Schmidhuber ldquoA com-mittee of neural networks for traffic sign classificationrdquo inProceedings of the 2011 International Joint Conference on NeuralNetworks (IJCNN 2011 - San Jose) San Jose CA USA July 2011

[45] P Pawara E Okafor O Surinta L Schomaker andMWieringldquoComparing Local Descriptors and Bags of Visual Words toDeep Convolutional Neural Networks for Plant Recognitionrdquoin Proceedings of the 6th International Conference on PatternRecognition Applications and Methods pp 479ndash486 PortoPortugal Feburary 2017

[46] M Lin ldquoNetwork inNnetworkrdquo httpsarxivorgabs131244002014

[47] K He X Zhang S Ren and J Sun ldquoIdentity mappings in deepresidual networksrdquo in Proceedings of the European Conferenceon Computer Vision pp 630ndash645 2016

[48] L van der Maaten and G Hinton ldquoVisualizing data using t-SNErdquo Journal of Machine Learning Research vol 9 pp 2579ndash2625 2008

International Journal of

AerospaceEngineeringHindawiwwwhindawicom Volume 2018

RoboticsJournal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Active and Passive Electronic Components

VLSI Design

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Shock and Vibration

Hindawiwwwhindawicom Volume 2018

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawiwwwhindawicom

Volume 2018

Hindawi Publishing Corporation httpwwwhindawicom Volume 2013Hindawiwwwhindawicom

The Scientific World Journal

Volume 2018

Control Scienceand Engineering

Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom

Journal ofEngineeringVolume 2018

SensorsJournal of

Hindawiwwwhindawicom Volume 2018

International Journal of

RotatingMachinery

Hindawiwwwhindawicom Volume 2018

Modelling ampSimulationin EngineeringHindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Navigation and Observation

International Journal of

Hindawi

wwwhindawicom Volume 2018

Advances in

Multimedia

Submit your manuscripts atwwwhindawicom

Page 8: ResearchArticle Can Deep Learning Identify Tomato Leaf

8 Advances in Multimedia

Table 5 Accuracies and training time in different network structuresThe values inside parenthesis denote batch size number of iterationsand training layers

Network topology Accuracy Time (minsec)ResNet (16 4992 1-ldquofcrdquo) 9643 59min 30secResNet (16 4992 37-ldquofcrdquo) 9728 44min 13secResNet (16 4992 79-ldquofcrdquo) 9643 37min 27secResNet (16 4992 111-ldquofcrdquo) 9575 30min 6secResNet (16 4992 141-ldquofcrdquo) 9532 24min 15secResNet (16 4992 163-ldquofcrdquo) 9269 19min 31secResNet (16 9984 1-ldquofcrdquo) 9694 118min 32secResNet (16 9984 37-ldquofcrdquo) 9702 92min 53secResNet (16 9984 79-ldquofcrdquo) 9677 72min 23secResNet (16 9984 111-ldquofcrdquo) 9626 58min 40secResNet (16 9984 141-ldquofcrdquo) 9575 47min 22secResNet (16 9984 163-ldquofcrdquo) 9396 39min 32sec

iterations affecting the transfer learning of the ResNet wasconducted A small batch size of 16 combining a moderatenumber of iterations of 4992 is the optimal choice in thiswork Our findings suggest that for a particular task neitherlarge batch size nor large number of iterations may improvethe accuracy of the target modelThe setting of batch size andnumber of iterations depends on your data set and the utilizednetwork Next the best combined model was used to fine-tune the structure Fine-tuning ResNet layers from 37 to ldquofcrdquoobtained the highest accuracy 9728 in identifying tomatoleaf disease Based on the amount of available data layer-wisefine-tuning may provide a practical way to achieve the bestperformance of the application at hand We believe that theresults obtained in this work will bring some inspiration toother similar visual recognition problems and the practicalstudy of this work can be easily extended to other plant leafdisease identification problems

Data Availability

Thetomato leaf data supporting this work are frompreviouslyreported studies which have been cited The processed dataare available from the corresponding author request

Conflicts of Interest

The authors declare that there are no conflicts of interestregarding the publication of this paper

Acknowledgments

This study was supported by the National Science and tech-nology support program (2014BAD12B01-1-3) PublicWelfareIndustry (Agriculture) Research Projects Level-2 (201503116-04-06) Postdoctoral Foundation of Heilongjiang Province(LBHZ15020) Harbin Applied Technology Research andDevelopment Program (2017RAQXJ096) and EconomicDecision Making and Early Warning of Soybean Industryin Technology Collaborative Innovation System of SoybeanIndustry in Heilongjiang Province (20170401)

References

[1] RChaerani andR EVoorrips ldquoTomato early blight (Alternariasolani) The pathogen genetics and breeding for resistancerdquoJournal of General Plant Pathology vol 72 no 6 pp 335ndash3472006

[2] A M Dickey L S Osborne and C L Mckenzie ldquoPapaya(Carica papaya Brassicales Caricaceae) is not a host plant oftomato yellow leaf curl virus (TYLCV family Geminiviridaegenus Begomovirus)rdquo Florida Entomologist vol 95 no 1 pp211ndash213 2012

[3] G Wei L Baoju S Yanxia and X Xuewen ldquoStudies on path-ogenicity differentiation of corynespora cassiicola isolatesagainst cucumber tomato and eggplantrdquo Acta HorticulturaeSinica vol 38 no 3 pp 465ndash470 2011

[4] P Lindhout W Korta M Cislik I Vos and T GerlaghldquoFurther identification of races of Cladosporium fulvum (Fulviafulva) on tomato originating from the Netherlands France andPolandrdquo Netherlands Journal of Plant Pathology vol 95 no 3pp 143ndash148 1989

[5] K Kubota S Tsuda A Tamai and T Meshi ldquoTomato mosaicvirus replication protein suppresses virus-targeted posttran-scriptional gene silencingrdquo Journal of Virology vol 77 no 20pp 11016ndash11026 2003

[6] M Tian B Benedetti and S Kamoun ldquoA second Kazal-like protease inhibitor from Phytophthora infestans inhibitsand interacts with the apoplastic pathogenesis-related proteaseP69B of tomatordquo Plant Physiology vol 138 no 3 pp 1785ndash17932005

[7] L E Blum ldquoReduction of incidence and severity of Septorialycopersici leaf spot of tomatowith bacteria and yeastsrdquoCienciaRural vol 30 no 5 pp 761ndash765 2000

[8] E A Chatzivasileiadis and M W Sabelis ldquoToxicity of methylketones from tomato trichomes to Tetranychus urticae KochrdquoExperimental and Applied Acarology vol 21 no 6-7 pp 473ndash484 1997

[9] M Anthimopoulos S Christodoulidis L Ebner A Christeand S Mougiakakou ldquoLung Pattern Classification for Inter-stitial Lung Diseases Using a Deep Convolutional NeuralNetworkrdquo IEEE Transactions on Medical Imaging vol 35 no5 pp 1207ndash1216 2016

[10] Y Tang and X Wu ldquoText-independent writer identification viaCNN features and joint Bayesianrdquo in Proceedings of the 15th

Advances in Multimedia 9

International Conference on Frontiers in Handwriting Recogni-tion ICFHR 2016 pp 566ndash571 Shenzhen China October 2016

[11] Y Tang and XWu ldquoSaliency Detection via Combining Region-Level and Pixel-Level Predictions with CNNsrdquo in Proceedings ofthe EuropeanConference onComputer Vision (ECCV) pp 1608ndash05186 2016

[12] Y Tang and X Wu ldquoSalient object detection with chainedmulti-scale fully convolutional networkrdquo ACM Multimedia(ACMMM) pp 618ndash626 2017

[13] Y Tang and X Wu ldquoScene text detection and segmentationbased on cascaded convolution neural networksrdquo IEEE Trans-actions on Image Processing vol 26 no 3 pp 1509ndash1520 2017

[14] Y Tang and X Wu ldquoScene Text Detection using Superpixelbased Stroke Feature Transform and Deep Learning basedRegion Classificationrdquo IEEE Transactions on Multimedia vol20 no 9 pp 2276ndash2288 2018

[15] Y Yao XWu Z Lei S Shan andW Zuo ldquoJoint Representationand Truncated Inference Learning for Correlation Filter basedTrackingrdquo in Proceedings of the European Conference on Com-puter Vision (ECCV) pp 1ndash14 2018

[16] L Zhang F Yang Y Daniel Zhang and Y J Zhu ldquoRoadcrack detection using deep convolutional neural networkrdquoin Proceedings of the 23rd IEEE International Conference onImage Processing ICIP 2016 pp 3708ndash3712 Phoenix AZ USASeptember 2016

[17] D Xie L Zhang andL Bai ldquoDeep learning in visual computingand signal processingrdquo Applied Computational Intelligence andSoft Computing vol 2017 Article ID 1320780 13 pages 2017

[18] Z Zhou J Shin L Zhang S Gurudu M Gotway and J LiangldquoFine-tuning convolutional neural networks for biomedicalimage analysis Actively and incrementallyrdquo inProceedings of the30th IEEE Conference on Computer Vision and Pattern Recog-nition CVPR 2017 pp 4761ndash4772 USA July 2017

[19] Z Liu P Luo X Wang and X Tang ldquoDeep learning face attri-butes in the wildrdquo in Proceedings of the 15th IEEE InternationalConference on Computer Vision ICCV 2015 pp 3730ndash3738Santiago Chile 2015

[20] W Ouyang and X Wang ldquoJoint deep learning for pedestriandetectionrdquo in Proceedings of the 14th IEEE International Con-ference on Computer Vision (ICCV rsquo13) pp 2056ndash2063 SydneyAustralia December 2013

[21] A Krizhevsky I Sutskever andG EHinton ldquoImagenet classifi-cation with deep convolutional neural networksrdquo in Proceedingsof the 26th Annual Conference on Neural Information ProcessingSystems (NIPS rsquo12) pp 1097ndash1105 Lake Tahoe Nev USADecember 2012

[22] C Szegedy W Liu Y Jia et al ldquoGoing deeper with convolu-tionsrdquo inProceedings of the IEEEConference onComputer Visionand Pattern Recognition (CVPR rsquo15) pp 1ndash9 IEEE BostonMass USA June 2015

[23] K Simonyan and A Zisserman ldquoVery deep convolutional net-works for large-scale image recognitionrdquo httpsarxivorgabs14091556 2015

[24] C Szegedy V Vanhoucke S Ioffe J Shlens and Z WojnaldquoRethinking the inception architecture for computer visionrdquohttpsarxivorgabs151200567 2015

[25] C Szegedy S Ioffe V Vanhoucke and A Alemi ldquoInception-v4 inception-ResNet and the impact of residual connections onlearningrdquo httpsarxivorgabs160207261 2016

[26] K He X Zhang S Ren and J Sun ldquoDeep residual learning forimage recognitionrdquo in Proceedings of the 2016 IEEE Conference

on Computer Vision and Pattern Recognition CVPR 2016 pp770ndash778 July 2016

[27] G Huang Z Liu K Q Weinberger and L van der MaatenldquoDensely connected convolutional networksrdquo in Proceedingsof the 2017 IEEE Conference on Computer Vision and PatternRecognition (CVPR) Honolulu HI USA 2016

[28] S J Pan and Q Yang ldquoA survey on transfer learningrdquo IEEETransactions on Knowledge and Data Engineering vol 22 no10 pp 1345ndash1359 2010

[29] M Mehdipour Ghazi B Yanikoglu and E Aptoula ldquoPlantidentification using deep neural networks via optimization oftransfer learning parametersrdquo Neurocomputing vol 235 pp228ndash235 2017

[30] S Ruder ldquoAn overview of gradient descent optimization algo-rithmsrdquo httpsarxivorgabs160904747 2017

[31] D P Kingma and J L Ba ldquoAdam a method for stochasticoptimizationrdquo httpsarxivorgabs14126980 2017

[32] S S Sannakki V S Rajpurohit V B Nargund R Arun Kumarand S Prema Yallur ldquoLeaf disease grading by machine visionand fuzzy logicrdquo International Journal of Computer Technologyand Applications vol 2 no 5 pp 1709ndash1716 2011

[33] D Samanta P P Chaudhury and A Ghosh ldquoScab diseasesdetection of potato using image processingrdquo InternationalJournal of Computer Trends and Technology vol 3 pp 109ndash1132012

[34] P J Herrera J Dorado and A Ribeiro ldquoA novel approach forweed type classification based on shape descriptors and a fuzzydecision-making methodrdquo Sensors vol 14 no 8 pp 15304ndash15324 2014

[35] B Cheng and E T Matson ldquoA feature-based machine learningagent for automatic rice andweed discriminationrdquo InternationalConference onArtificial Intelligence and SoftComputing pp 517ndash527 2015

[36] S Sankaran and R Ehsani ldquoComparison of visible-near infra-red and mid-infrared spectroscopy for classification of Huang-longbing and citrus canker infected leavesrdquo Agricultural Engi-neering International CIGR Journal vol 15 no 3 pp 75ndash792013

[37] X Cheng Y Zhang Y Chen Y Wu and Y Yue ldquoPest iden-tification via deep residual learning in complex backgroundrdquoComputers and Electronics in Agriculture vol 141 pp 351ndash3562017

[38] A dos Santos Ferreira D Matte Freitas G Goncalves daSilva H Pistori and M Theophilo Folhes ldquoWeed detection insoybean crops using ConvNetsrdquo Computers and Electronics inAgriculture vol 143 pp 314ndash324 2017

[39] S Sladojevic M Arsenovic A Anderla D Culibrk andD Stefanovic ldquoDeep Neural Networks Based Recognition ofPlant Diseases by Leaf Image Classificationrdquo ComputationalIntelligence and Neuroscience vol 2016 Article ID 3289801 11pages 2016

[40] S P Mohanty D P Hughes and M Salathe ldquoUsing deep learn-ing for image-based plant disease detectionrdquo Frontiers in PlantScience vol 7 article no 1419 2016

[41] I Sa Z Ge F Dayoub B Upcroft T Perez and C McCoolldquoDeepfruits A fruit detection system using deep neural net-worksrdquo Sensors vol 16 article no 1222 no 8 2016

[42] V Nair and G E Hinton ldquoRectified linear units improveRestricted Boltzmann machinesrdquo in Proceedings of the 27thInternational Conference on Machine Learning (ICML rsquo10) pp807ndash814 Haifa Israel June 2010

10 Advances in Multimedia

[43] D P Hughes and M Salathe ldquoAn open access repository ofimages on plant health to enable the development of mobiledisease diagnosticsrdquo httpsarxivorgabs151108060 2016

[44] D Ciresan U Meier J Masci and J Schmidhuber ldquoA com-mittee of neural networks for traffic sign classificationrdquo inProceedings of the 2011 International Joint Conference on NeuralNetworks (IJCNN 2011 - San Jose) San Jose CA USA July 2011

[45] P Pawara E Okafor O Surinta L Schomaker andMWieringldquoComparing Local Descriptors and Bags of Visual Words toDeep Convolutional Neural Networks for Plant Recognitionrdquoin Proceedings of the 6th International Conference on PatternRecognition Applications and Methods pp 479ndash486 PortoPortugal Feburary 2017

[46] M Lin ldquoNetwork inNnetworkrdquo httpsarxivorgabs131244002014

[47] K He X Zhang S Ren and J Sun ldquoIdentity mappings in deepresidual networksrdquo in Proceedings of the European Conferenceon Computer Vision pp 630ndash645 2016

[48] L van der Maaten and G Hinton ldquoVisualizing data using t-SNErdquo Journal of Machine Learning Research vol 9 pp 2579ndash2625 2008

International Journal of

AerospaceEngineeringHindawiwwwhindawicom Volume 2018

RoboticsJournal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Active and Passive Electronic Components

VLSI Design

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Shock and Vibration

Hindawiwwwhindawicom Volume 2018

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawiwwwhindawicom

Volume 2018

Hindawi Publishing Corporation httpwwwhindawicom Volume 2013Hindawiwwwhindawicom

The Scientific World Journal

Volume 2018

Control Scienceand Engineering

Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom

Journal ofEngineeringVolume 2018

SensorsJournal of

Hindawiwwwhindawicom Volume 2018

International Journal of

RotatingMachinery

Hindawiwwwhindawicom Volume 2018

Modelling ampSimulationin EngineeringHindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Navigation and Observation

International Journal of

Hindawi

wwwhindawicom Volume 2018

Advances in

Multimedia

Submit your manuscripts atwwwhindawicom

Page 9: ResearchArticle Can Deep Learning Identify Tomato Leaf

Advances in Multimedia 9

International Conference on Frontiers in Handwriting Recogni-tion ICFHR 2016 pp 566ndash571 Shenzhen China October 2016

[11] Y Tang and XWu ldquoSaliency Detection via Combining Region-Level and Pixel-Level Predictions with CNNsrdquo in Proceedings ofthe EuropeanConference onComputer Vision (ECCV) pp 1608ndash05186 2016

[12] Y Tang and X Wu ldquoSalient object detection with chainedmulti-scale fully convolutional networkrdquo ACM Multimedia(ACMMM) pp 618ndash626 2017

[13] Y Tang and X Wu ldquoScene text detection and segmentationbased on cascaded convolution neural networksrdquo IEEE Trans-actions on Image Processing vol 26 no 3 pp 1509ndash1520 2017

[14] Y Tang and X Wu ldquoScene Text Detection using Superpixelbased Stroke Feature Transform and Deep Learning basedRegion Classificationrdquo IEEE Transactions on Multimedia vol20 no 9 pp 2276ndash2288 2018

[15] Y Yao XWu Z Lei S Shan andW Zuo ldquoJoint Representationand Truncated Inference Learning for Correlation Filter basedTrackingrdquo in Proceedings of the European Conference on Com-puter Vision (ECCV) pp 1ndash14 2018

[16] L Zhang F Yang Y Daniel Zhang and Y J Zhu ldquoRoadcrack detection using deep convolutional neural networkrdquoin Proceedings of the 23rd IEEE International Conference onImage Processing ICIP 2016 pp 3708ndash3712 Phoenix AZ USASeptember 2016

[17] D Xie L Zhang andL Bai ldquoDeep learning in visual computingand signal processingrdquo Applied Computational Intelligence andSoft Computing vol 2017 Article ID 1320780 13 pages 2017

[18] Z Zhou J Shin L Zhang S Gurudu M Gotway and J LiangldquoFine-tuning convolutional neural networks for biomedicalimage analysis Actively and incrementallyrdquo inProceedings of the30th IEEE Conference on Computer Vision and Pattern Recog-nition CVPR 2017 pp 4761ndash4772 USA July 2017

[19] Z Liu P Luo X Wang and X Tang ldquoDeep learning face attri-butes in the wildrdquo in Proceedings of the 15th IEEE InternationalConference on Computer Vision ICCV 2015 pp 3730ndash3738Santiago Chile 2015

[20] W Ouyang and X Wang ldquoJoint deep learning for pedestriandetectionrdquo in Proceedings of the 14th IEEE International Con-ference on Computer Vision (ICCV rsquo13) pp 2056ndash2063 SydneyAustralia December 2013

[21] A Krizhevsky I Sutskever andG EHinton ldquoImagenet classifi-cation with deep convolutional neural networksrdquo in Proceedingsof the 26th Annual Conference on Neural Information ProcessingSystems (NIPS rsquo12) pp 1097ndash1105 Lake Tahoe Nev USADecember 2012

[22] C Szegedy W Liu Y Jia et al ldquoGoing deeper with convolu-tionsrdquo inProceedings of the IEEEConference onComputer Visionand Pattern Recognition (CVPR rsquo15) pp 1ndash9 IEEE BostonMass USA June 2015

[23] K Simonyan and A Zisserman ldquoVery deep convolutional net-works for large-scale image recognitionrdquo httpsarxivorgabs14091556 2015

[24] C Szegedy V Vanhoucke S Ioffe J Shlens and Z WojnaldquoRethinking the inception architecture for computer visionrdquohttpsarxivorgabs151200567 2015

[25] C Szegedy S Ioffe V Vanhoucke and A Alemi ldquoInception-v4 inception-ResNet and the impact of residual connections onlearningrdquo httpsarxivorgabs160207261 2016

[26] K He X Zhang S Ren and J Sun ldquoDeep residual learning forimage recognitionrdquo in Proceedings of the 2016 IEEE Conference

on Computer Vision and Pattern Recognition CVPR 2016 pp770ndash778 July 2016

[27] G Huang Z Liu K Q Weinberger and L van der MaatenldquoDensely connected convolutional networksrdquo in Proceedingsof the 2017 IEEE Conference on Computer Vision and PatternRecognition (CVPR) Honolulu HI USA 2016

[28] S J Pan and Q Yang ldquoA survey on transfer learningrdquo IEEETransactions on Knowledge and Data Engineering vol 22 no10 pp 1345ndash1359 2010

[29] M Mehdipour Ghazi B Yanikoglu and E Aptoula ldquoPlantidentification using deep neural networks via optimization oftransfer learning parametersrdquo Neurocomputing vol 235 pp228ndash235 2017

[30] S Ruder ldquoAn overview of gradient descent optimization algo-rithmsrdquo httpsarxivorgabs160904747 2017

[31] D P Kingma and J L Ba ldquoAdam a method for stochasticoptimizationrdquo httpsarxivorgabs14126980 2017

[32] S S Sannakki V S Rajpurohit V B Nargund R Arun Kumarand S Prema Yallur ldquoLeaf disease grading by machine visionand fuzzy logicrdquo International Journal of Computer Technologyand Applications vol 2 no 5 pp 1709ndash1716 2011

[33] D Samanta P P Chaudhury and A Ghosh ldquoScab diseasesdetection of potato using image processingrdquo InternationalJournal of Computer Trends and Technology vol 3 pp 109ndash1132012

[34] P J Herrera J Dorado and A Ribeiro ldquoA novel approach forweed type classification based on shape descriptors and a fuzzydecision-making methodrdquo Sensors vol 14 no 8 pp 15304ndash15324 2014

[35] B Cheng and E T Matson ldquoA feature-based machine learningagent for automatic rice andweed discriminationrdquo InternationalConference onArtificial Intelligence and SoftComputing pp 517ndash527 2015

[36] S Sankaran and R Ehsani ldquoComparison of visible-near infra-red and mid-infrared spectroscopy for classification of Huang-longbing and citrus canker infected leavesrdquo Agricultural Engi-neering International CIGR Journal vol 15 no 3 pp 75ndash792013

[37] X Cheng Y Zhang Y Chen Y Wu and Y Yue ldquoPest iden-tification via deep residual learning in complex backgroundrdquoComputers and Electronics in Agriculture vol 141 pp 351ndash3562017

[38] A dos Santos Ferreira D Matte Freitas G Goncalves daSilva H Pistori and M Theophilo Folhes ldquoWeed detection insoybean crops using ConvNetsrdquo Computers and Electronics inAgriculture vol 143 pp 314ndash324 2017

[39] S Sladojevic M Arsenovic A Anderla D Culibrk andD Stefanovic ldquoDeep Neural Networks Based Recognition ofPlant Diseases by Leaf Image Classificationrdquo ComputationalIntelligence and Neuroscience vol 2016 Article ID 3289801 11pages 2016

[40] S P Mohanty D P Hughes and M Salathe ldquoUsing deep learn-ing for image-based plant disease detectionrdquo Frontiers in PlantScience vol 7 article no 1419 2016

[41] I Sa Z Ge F Dayoub B Upcroft T Perez and C McCoolldquoDeepfruits A fruit detection system using deep neural net-worksrdquo Sensors vol 16 article no 1222 no 8 2016

[42] V Nair and G E Hinton ldquoRectified linear units improveRestricted Boltzmann machinesrdquo in Proceedings of the 27thInternational Conference on Machine Learning (ICML rsquo10) pp807ndash814 Haifa Israel June 2010

10 Advances in Multimedia

[43] D P Hughes and M Salathe ldquoAn open access repository ofimages on plant health to enable the development of mobiledisease diagnosticsrdquo httpsarxivorgabs151108060 2016

[44] D Ciresan U Meier J Masci and J Schmidhuber ldquoA com-mittee of neural networks for traffic sign classificationrdquo inProceedings of the 2011 International Joint Conference on NeuralNetworks (IJCNN 2011 - San Jose) San Jose CA USA July 2011

[45] P Pawara E Okafor O Surinta L Schomaker andMWieringldquoComparing Local Descriptors and Bags of Visual Words toDeep Convolutional Neural Networks for Plant Recognitionrdquoin Proceedings of the 6th International Conference on PatternRecognition Applications and Methods pp 479ndash486 PortoPortugal Feburary 2017

[46] M Lin ldquoNetwork inNnetworkrdquo httpsarxivorgabs131244002014

[47] K He X Zhang S Ren and J Sun ldquoIdentity mappings in deepresidual networksrdquo in Proceedings of the European Conferenceon Computer Vision pp 630ndash645 2016

[48] L van der Maaten and G Hinton ldquoVisualizing data using t-SNErdquo Journal of Machine Learning Research vol 9 pp 2579ndash2625 2008

International Journal of

AerospaceEngineeringHindawiwwwhindawicom Volume 2018

RoboticsJournal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Active and Passive Electronic Components

VLSI Design

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Shock and Vibration

Hindawiwwwhindawicom Volume 2018

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawiwwwhindawicom

Volume 2018

Hindawi Publishing Corporation httpwwwhindawicom Volume 2013Hindawiwwwhindawicom

The Scientific World Journal

Volume 2018

Control Scienceand Engineering

Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom

Journal ofEngineeringVolume 2018

SensorsJournal of

Hindawiwwwhindawicom Volume 2018

International Journal of

RotatingMachinery

Hindawiwwwhindawicom Volume 2018

Modelling ampSimulationin EngineeringHindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Navigation and Observation

International Journal of

Hindawi

wwwhindawicom Volume 2018

Advances in

Multimedia

Submit your manuscripts atwwwhindawicom

Page 10: ResearchArticle Can Deep Learning Identify Tomato Leaf

10 Advances in Multimedia

[43] D P Hughes and M Salathe ldquoAn open access repository ofimages on plant health to enable the development of mobiledisease diagnosticsrdquo httpsarxivorgabs151108060 2016

[44] D Ciresan U Meier J Masci and J Schmidhuber ldquoA com-mittee of neural networks for traffic sign classificationrdquo inProceedings of the 2011 International Joint Conference on NeuralNetworks (IJCNN 2011 - San Jose) San Jose CA USA July 2011

[45] P Pawara E Okafor O Surinta L Schomaker andMWieringldquoComparing Local Descriptors and Bags of Visual Words toDeep Convolutional Neural Networks for Plant Recognitionrdquoin Proceedings of the 6th International Conference on PatternRecognition Applications and Methods pp 479ndash486 PortoPortugal Feburary 2017

[46] M Lin ldquoNetwork inNnetworkrdquo httpsarxivorgabs131244002014

[47] K He X Zhang S Ren and J Sun ldquoIdentity mappings in deepresidual networksrdquo in Proceedings of the European Conferenceon Computer Vision pp 630ndash645 2016

[48] L van der Maaten and G Hinton ldquoVisualizing data using t-SNErdquo Journal of Machine Learning Research vol 9 pp 2579ndash2625 2008

International Journal of

AerospaceEngineeringHindawiwwwhindawicom Volume 2018

RoboticsJournal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Active and Passive Electronic Components

VLSI Design

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Shock and Vibration

Hindawiwwwhindawicom Volume 2018

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawiwwwhindawicom

Volume 2018

Hindawi Publishing Corporation httpwwwhindawicom Volume 2013Hindawiwwwhindawicom

The Scientific World Journal

Volume 2018

Control Scienceand Engineering

Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom

Journal ofEngineeringVolume 2018

SensorsJournal of

Hindawiwwwhindawicom Volume 2018

International Journal of

RotatingMachinery

Hindawiwwwhindawicom Volume 2018

Modelling ampSimulationin EngineeringHindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Navigation and Observation

International Journal of

Hindawi

wwwhindawicom Volume 2018

Advances in

Multimedia

Submit your manuscripts atwwwhindawicom

Page 11: ResearchArticle Can Deep Learning Identify Tomato Leaf

International Journal of

AerospaceEngineeringHindawiwwwhindawicom Volume 2018

RoboticsJournal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Active and Passive Electronic Components

VLSI Design

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Shock and Vibration

Hindawiwwwhindawicom Volume 2018

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawiwwwhindawicom

Volume 2018

Hindawi Publishing Corporation httpwwwhindawicom Volume 2013Hindawiwwwhindawicom

The Scientific World Journal

Volume 2018

Control Scienceand Engineering

Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom

Journal ofEngineeringVolume 2018

SensorsJournal of

Hindawiwwwhindawicom Volume 2018

International Journal of

RotatingMachinery

Hindawiwwwhindawicom Volume 2018

Modelling ampSimulationin EngineeringHindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawiwwwhindawicom Volume 2018

Hindawiwwwhindawicom Volume 2018

Navigation and Observation

International Journal of

Hindawi

wwwhindawicom Volume 2018

Advances in

Multimedia

Submit your manuscripts atwwwhindawicom