Fine Tuning Akurasi 8 Persen
Fine Tuning Akurasi 8 Persen
Fine Tuning Akurasi 8 Persen
1. Introduction
Crop diseases are one of the leading causes of the decline in agricultural production. In recent years, various
crop diseases have increased and affected the agricultural, economic, and health sectors. The disease generally
generates noticeable marks or lesions on the crop leaves. Therefore, it is crucial to detect the presence of the disease
at an early stage. However, traditional approaches are commonly used by most farms and plantations to recognize the
disease. Consequently, it is time-consuming and leads to high costs and misdetection [1].
Hence, researchers have designed the Convolutional Neural Network (CNN) models as a computer-aided
disease diagnosis. CNN models can overcome the problem of object recognition, and classification has significantly
improved in the last decade [2], [3]. In recent years, CNN models have emerged as the most extensively used model in
crop disease detection [3]–[10]. The CNN model can be applied from scratch or by using pre-trained designs.
Promisingly, pre-trained CNN models outperformed the scratch designs. AlexNet was the first proposed pre-trained
CNN model by [11] on the 1.2 million high-resolution images within the 1.000 different classes in the ImageNet LSVRC-
2010 contest. Furthermore, many CNN models were offered in the next contest ImageNet LSVRC.
Recently, several pre-trained CNN architectures have been suggested, for example, GoogleNet [12], VGGNet
[13], ResNet [14], InceptionV3 [15], DenseNet [16], Xception [17], MobileNet [18], InceptionResNetV2 [19], NasNet [20],
and EfficientNet [21]. In prior research, researchers used two methods in crop disease recognition using CNN models
[22]. Some use transfer learning to fine-tune well-known designs. Others use well-known architectures, such as
modifying the convolutional layers, changing the order of layers, adding attention layers before the classifier layer, and
ensembling more than two pre-trained CNN models. They aim to improve the model performance.
The most considered dataset is the Plant Village dataset. In this dataset, [23] corresponded with the results of
AlexNet and GoogLeNet, in detecting 26 crop diseases in laboratory leave images. They employed transfer learning to
increase categorization performance and achieve an accuracy of 99.35%. CaffeNet was proposed by [24] on 15 class
leaf images and achieved an accuracy of 96.0%. Cucumber leaves were considered by [25]–[27] using Deep
Convolution Neural Network (Scratch designs), modified Deep Convolution Neural Network, and combined
DeepLabV3+ and U-Net, respectively. Moreover, the detection of citrus diseases conducted by [22], [28], [29] used
modified MobileNetV2 and an Ensemble AlexNet, VGG16, ResNet50, and InceptionResNetV2. Both cucumber and
citrus crops achieved an accuracy of over 98%.
In addition to the entire categories of the Plant Village dataset, researchers considered multiclass classification
with 38 classes proposed by [30]–[34]. Pre-trained MobileNet, EfficientNet, DensNet, and ensemble MobileNet with
DenseNets were applied. All suggested model performances achieved an accuracy of over 92%, and 100% for the
ensemble model was proposed by [34].
Cite: D. Sutaji and H. Rosyid, “Convolutional Neural Network (CNN) Models for Crop Diseases Classification”, KINETIK, vol. 7, no. 2, May.
2022. https://doi.org/10.22219/kinetik.v7i2.1443
188 Kinetik: Game Technology, Information System, Computer Network, Computing, Electronics, and Control
Other datasets such as iBean, Citrus, Rice, and Turk-Plants were also used to detect crop diseases. Several pre-
trained models were used, such as modified and ensembled by researchers, such as [35]–[40]. The model's average
performance is above 99%.
However, various datasets have been used in the literature, making it difficult to compare directly. Several
detection systems work only on two classes, while the others detect crop diseases in over 38 classes. Considering this
term, we suggested investigating the classification of crop disease by using ten pre-trained CNN algorithms to consider
the entire Plant Village dataset. This dataset consists of 54.305 RGB leaf images with 256 x 256 pixels. The dataset
subsists of 38 classes from 14 crop species: strawberries, apples, grapes, cherries, corn, oranges, blueberries,
potatoes, soybeans, pumpkins, peppers, raspberries, peaches, and tomatoes.
Another problematic aspect of comparing which CNN models can perform better is the various configurations
and modifications of the pre-trained CNN models to detect crop disease. As a simple investigation to compare, we
proposed regular and unfroze the final convolutional blocks aforementioned pre-trained CNN models. It aims to
generate high-dimensional extracted features to provide a more helpful characteristic of an object. Our main
contributions to this research are:
1. Comparing the performance of 10 pre-trained CNN models to detect crop disease.
2. Improving the performance of 10 pre-trained CNN models by unfroze the final convolution blocks.
This research is organized as follows: Section 1 introduces crop diseases, state-of-the-art methods for classifying
crop diseases, explains the problem of statements, and gives solutions as a contribution to the research. Section 2
discusses research methods. Section 3 presents the experimental results and performance comparison. Eventually,
Section 4 explains the conclusion of this study.
2. Research Method
2.1 Method
This study considered 10 pre-trained CNN architectures, such as AlexNet, GoogleNet, VGG16, VGG19,
ResNet50V2, InceptionV3, Xception, MobileNetV2, DenseNet121, and InceptionResnetV2 to classify crop diseases.
These CNN architectures were developed with a transfer learning approach and implemented fine-tuning by unfreezing
the last block of convolutional layers. It aims to enhance the high-dimensional extracted features. Updating the weight
parameter impacts capture more details of high-dimensional extracted features. Figure 1 illustrates our research
method, and Figure 2 describes our suggested transfer learning and fine-tuning design.
2.2 Dataset
Plant Village dataset consists of 38 classes and includes 54,305 images with 256 x 256 size in the RGB channel.
The dataset consists of 14 plant species: apples, blueberries, cherries, corn, grapes, oranges, peppers, raspberries,
potatoes, pumpkins, peaches, raspberries, soybeans, strawberries, and tomatoes. The diseases from the 14 plant
© 2022 The Authors. Published by Universitas Muhammadiyah Malang
This is an open access article under the CC BY SA license. (https://creativecommons.org/licenses/by-sa/4.0/)
Kinetik: Game Technology, Information System, Computer Network, Computing, Electronics, and Control 189
species are distributed as 17 fungal diseases, 4 bacterial diseases, 2 fungal diseases, 2 viral diseases, and the last
disease caused by mites has 1 class. Meanwhile, there are 12 plant species in the healthy category of plant leaves [23].
The considered dataset was collected from https://www.kaggle.com/abdallahalidev/plantvillage-dataset. The example
of each class image is shown in Figure 3.
We pre-processed the collected dataset in four steps. The first step is resizing all images into 224 x 224 pixels.
The second step is to split and distribute the dataset into a training set, a validation set, and a test set with a composition
of 70%, 10%, and 20%, respectively. The detail of distribution is described in Figure 3. The third step is enforced image
normalization, dividing the intensity value of each pixel by 255. The reason is to reduce computational costs because
the intensity value of each pixel is between 0 and 1 [41].
Furthermore, the last step is image augmentation, completing the image in the data series more varied since the
acquired image is subject to different conditions (angle, illumination, and background). Image augmentation is only
performed on the training data because this model can accept different conditions of the leaf image under test. This
step's configuration includes image rotation with a rotation angle of 30 degrees, zooming image with a percentage of
0.3, cropping with width_shift_range, height_shift_range, shear_range, and zoom_range settings with values of 0.3.
The last configuration is horizontal_flip with the setting True.
(a1) (a2) (a3) (a4) (b1) (ce1) (ce2) (co1) (co2) (co3)
(co4) (g1) (g2) (g3) (g4) (o1) (p1) (p2) (pp1) (pp2)
(po1) (po2) (po3) (r1) (s1) (sq1) (st1) (st2) (t1) (t2)
Cite: D. Sutaji and H. Rosyid, “Convolutional Neural Network (CNN) Models for Crop Diseases Classification”, KINETIK, vol. 7, no. 2, May.
2022. https://doi.org/10.22219/kinetik.v7i2.1443
190 Kinetik: Game Technology, Information System, Computer Network, Computing, Electronics, and Control
3. Results and Discussion
Determining the hyperparameters for model training is critical work. We accept Adam as the optimizer, including
a learning rate of 0.0001 as the best value. We initially tried different learning rate values starting from 0.1 to 0.0001 in
0.1 additions. We also tested other optimizers, such as SGD and RMSprop. Therefore, Adam optimizer produces
excellent training, validation, and testing performances. Another parameter is an epoch, which we set at 100 and
consider to have an early stop value of 10. Training can be more efficient, and the process does not waste time in 100
epochs. We also employed cross-entropy to track each epoch's decreasing error rate and accuracy. Finally, a batch
size of 25, 18, and 18 for training, validation, and testing data were presented in Table 1.
The highest accuracy obtained by the DenseNet121 model of 98.97% is quite promising for crop disease
detection. It will significantly assist farmers or agricultural experts in detecting crop diseases.
Cite: D. Sutaji and H. Rosyid, “Convolutional Neural Network (CNN) Models for Crop Diseases Classification”, KINETIK, vol. 7, no. 2, May.
2022. https://doi.org/10.22219/kinetik.v7i2.1443
192 Kinetik: Game Technology, Information System, Computer Network, Computing, Electronics, and Control
AlexNet GoogleNet
VGG16 VGG19
InceptionV3 Xception
DenseNet121 ResNet50
Figure 4. Training Accuracy and Loss Curve of the Second Scenario Design
In addition to improving accuracy, unfreezing the final block of the convolution layer speeds up the training
process. The average reduction in time consumption in each model is almost 50%, or 2.20 hours, with the most
significant reduction being 8 hours.
However, our research did not ultimately outperform the earlier research models. We accomplished just simple
fine-tuning by unfreezing the final convolution block layers. For example, the CNN AlexNet and GoogleNet models
proposed by [23] get an accuracy value of 99.35%, while our proposed model is only 98.34%. If we refer to the model
performance graph in Figure 4, our proposed model is overfitting, and the VGG19 model followed this condition. While
the DenseNet121 model, which is our best model, does not outperform the model proposed by [31]. Our DenseNet121
model only gets 98.97% accuracy and 99.75% for [31] model.
On the other hand, our model outperforms the model proposed by [43], which was only implemented on citrus
crops. Their models get 97%, 96.2%, and 97% accuracy with the VGG16, VGG19, and InceptionV3 models. Meanwhile,
our proposed model achieved an accuracy of 98.27%, 97.19%, and 97.30%. This positive result also outperforms the
Model proposed by [37], which obtained an accuracy of 92% on rice plants. Likewise, the InceptionResNetV2
model we propose has an accuracy of 98.76%, higher than the model proposed by [22] of 98.64%. This positive result
is followed by the Xception model we propose to get almost the same results as the Xception model proposed by [46],
which are 97.58% and 97.60%, respectively. Unfortunately, our MobileNetV2 model did not follow these positive results.
Our MobileNetV2 model is inferior to some of the previously proposed models, which applied modification to an
ensemble of MobileNet proposed by [29], [30], [36] and obtains more than 99% accuracy. Their models extract features
better than our model. Moreover, their models added several convolutional layers before the classification layer. Our
Cite: D. Sutaji and H. Rosyid, “Convolutional Neural Network (CNN) Models for Crop Diseases Classification”, KINETIK, vol. 7, no. 2, May.
2022. https://doi.org/10.22219/kinetik.v7i2.1443
194 Kinetik: Game Technology, Information System, Computer Network, Computing, Electronics, and Control
MobileNetV2 model provides 98.95% of accuracy in the second scenario. Table 5 describes the performance
comparison of our model with earlier studies.
In completing the discussion of this section, the fine-tuning model that we propose is relatively simple compared
to the model proposed by previous studies. Some of them applied architectural modifications, ensembles, and other
advance works. Fortunately, our model is not entirely inferior to the previously proposed model. Some models like
VGG16, VGG19, InceptionV3, and InceptionResNetV2 outperform models by [43]. The best models we
proposeDenseNet121 and MobileNetV2, promise to be improved in terms of accuracy and time consumption. Moreover,
MobileNetV2 has fewer parameters than DenseNet121, which means more affordable to embed in the mobile device
for live detection.
Although the proposed models in this article have achieved good results for the entire Plant Village dataset, there
are several works to improve the performance:
1. Adding several convolution layers before the classifier layer.
2. Model modifications by unfreezing the initial convolution layers to update low-dimensional weight or ensembling
models.
4. Conclusion
Fine-tuning by unfreezing the final convolution layer blocks can significantly improve the performance of pre-
trained 10 CNN models. From the results of our experiments, the best performance was obtained by DenseNet121 and
MobileNetV2 models of 98.97% and 98.95%, respectively. This result increases the average performance of CNN
models by 8.85%. Meanwhile, there was a drastic decrease of 50% for each model in time consumption, or 2.20 hours
on average. Further research can be developed with architectural modifications by adding several convolution layers
before the classifier layer, unfreezing the initial convolution layers to update low-dimensional weight, and ensembling
models.
Acknowledgment
Thank you to Lembaga Penelitian dan Pengabdian Masyarakat. (LPPM) University of Muhammadiyah Gresik
(UMG). This reseach article supported by LPM UMG Nomor: 066/MoU.In/II.3.UMG/DPPM/F/2021 for Internal Research
Funded.
References
[1] M. Turkoglu, B. Yanikoğlu, and D. Hanbay, “PlantDiseaseNet: convolutional neural network ensemble for plant disease and pest detection,”
Signal, Image and Video Processing, 2021. https://doi.org/10.1007/s11760-021-01909-2
[2] N. Ahmad, S. Asghar, and S. A. Gillani, “Transfer learning-assisted multi-resolution breast cancer histopathological images classification,”
Visual Computer, 2021. https://doi.org/10.1007/s00371-021-02153-y
[3] S. Kaur, S. Pandey, and S. Goel, “Plants Disease Identification and Classification Through Leaf Images: A Survey,” Archives of Computational
Methods in Engineering, vol. 26, no. 2, pp. 507–530, Apr. 2019. https://doi.org/10.1007/s11831-018-9255-6
[4] S. Sachar and A. Kumar, “Survey of feature extraction and classification techniques to identify plant through leaves,” Expert Systems with
Applications, vol. 167. Elsevier Ltd, Apr. 01, 2021. https://doi.org/10.1016/j.eswa.2020.114181
[5] V. Singh, N. Sharma, and S. Singh, “A review of imaging techniques for plant disease detection,” Artificial Intelligence in Agriculture, vol. 4, pp.
229–242, 2020. https://doi.org/10.1016/j.aiia.2020.10.002
[6] J. Lu, L. Tan, and H. Jiang, “Review on convolutional neural network (CNN) applied to plant leaf disease classification,” Agriculture
(Switzerland), vol. 11, no. 8. MDPI AG, Aug. 01, 2021. https://doi.org/10.3390/agriculture11080707
[7] A. Abade, P. A. Ferreira, and F. de Barros Vidal, “Plant diseases recognition on images using convolutional neural networks: A systematic
review,” Computers and Electronics in Agriculture, vol. 185. Elsevier B.V., Jun. 01, 2021. https://doi.org/10.1016/j.compag.2021.106125
[8] R. Manavalan, “Automatic identification of diseases in grains crops through computational approaches: A review,” Computers and Electronics
in Agriculture, vol. 178. Elsevier B.V., Nov. 01, 2020. https://doi.org/10.1016/j.compag.2020.105802
[9] Z. Iqbal, M. A. Khan, M. Sharif, J. H. Shah, M. H. ur Rehman, and K. Javed, “An automated detection and classification of citrus plant diseases
using image processing techniques: A review,” Computers and Electronics in Agriculture, vol. 153. Elsevier B.V., pp. 12–32, Oct. 01, 2018.
https://doi.org/10.1016/j.compag.2018.07.032
[10] L. C. Ngugi, M. Abelwahab, and M. Abo-Zahhad, “Recent advances in image processing techniques for automated leaf pest and disease
recognition – A review,” Information Processing in Agriculture, vol. 8, no. 1. China Agricultural University, pp. 27–51, Mar. 01, 2021.
https://doi.org/10.1016/j.inpa.2020.04.004
[11] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet Classification with Deep Convolutional Neural Networks,” in Advances in Neural
Information Processing Systems 25, 2012, pp. 1–9.
[12] C. Szegedy et al., “Going Deeper with Convolutions,” Sep. 2014.
[13] K. Simonyan and A. Zisserman, “Very Deep Convolutional Networks for Large-Scale Image Recognition,” Sep. 2014.
[14] K. He, X. Zhang, S. Ren, and J. Sun, “Deep Residual Learning for Image Recognition,” Dec. 2015.
[15] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking the Inception Architecture for Computer Vision,” Dec. 2015.
[16] G. Huang, Z. Liu, L. van der Maaten, and K. Q. Weinberger, “Densely Connected Convolutional Networks,” Aug. 2016.
[17] F. Chollet, “Xception: Deep Learning with Depthwise Separable Convolutions,” 2017.
[18] M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, and L.-C. Chen, “MobileNetV2: Inverted Residuals and Linear Bottlenecks,” Jan. 2018.
[19] C. Szegedy, S. Ioffe, V. Vanhoucke, and A. Alemi, “Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning,” Feb.
2016.
[20] B. Zoph, V. Vasudevan, J. Shlens, and Q. v. Le, “Learning Transferable Architectures for Scalable Image Recognition,” Jul. 2017.
[21] M. Tan and Q. v Le, “EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks,” 2019.
© 2022 The Authors. Published by Universitas Muhammadiyah Malang
This is an open access article under the CC BY SA license. (https://creativecommons.org/licenses/by-sa/4.0/)
Kinetik: Game Technology, Information System, Computer Network, Computing, Electronics, and Control 195
[22] M. Khanramaki, E. Askari Asli-Ardeh, and E. Kozegar, “Citrus pests classification using an ensemble of deep learning models,” Computers and
Electronics in Agriculture, vol. 186, Jul. 202. https://doi.org/10.1016/j.compag.2021.106192
[23] S. P. Mohanty, D. P. Hughes, and M. Salathé, “Using deep learning for image-based plant disease detection,” Frontiers in Plant Science, vol.
7, no. September, Sep. 2016. https://doi.org/10.3389/fpls.2016.01419
[24] S. Sladojevic, M. Arsenovic, A. Anderla, D. Culibrk, and D. Stefanovic, “Deep Neural Networks Based Recognition of Plant Diseases by Leaf
Image Classification,” Computational Intelligence and Neuroscience, vol. 2016, 2016. https://doi.org/10.1155/2016/3289801
[25] J. Ma, K. Du, F. Zheng, L. Zhang, Z. Gong, and Z. Sun, “A recognition method for cucumber diseases using leaf symptom images based on
deep convolutional neural network,” Computers and Electronics in Agriculture, vol. 154, pp. 18–24, Nov. 2018.
https://doi.org/10.1016/j.compag.2018.08.048
[26] S. Zhang, S. Zhang, C. Zhang, X. Wang, and Y. Shi, “Cucumber leaf disease identification with global pooling dilated convolutional neural
network,” Computers and Electronics in Agriculture, vol. 162, pp. 422–430, Jul. 2019. https://doi.org/10.1016/j.compag.2019.03.012
[27] C. Wang, P. Du, H. Wu, J. Li, C. Zhao, and H. Zhu, “A cucumber leaf disease severity classification method based on the fusion of DeepLabV3+
and U-Net,” Computers and Electronics in Agriculture, vol. 189, Oct. 2021. https://doi.org/10.1016/j.compag.2021.106373
[28] H. T. Rauf, B. A. Saleem, M. I. U. Lali, M. A. Khan, M. Sharif, and S. A. C. Bukhari, “A citrus fruits and leaves dataset for detection and
classification of citrus diseases through machine learning,” Data in Brief, vol. 26, Oct. 2019. https://doi.org/10.1016/j.dib.2019.104340
[29] U. Barman, R. D. Choudhury, D. Sahu, and G. G. Barman, “Comparison of convolution neural networks for smartphone image based real time
classification of citrus leaf disease,” Computers and Electronics in Agriculture, vol. 177, Oct. 2020.
https://doi.org/10.1016/j.compag.2020.105661
[30] R. Gandhi, S. Nimbalkar, N. Yelamanchili, and S. Ponkshe, “Plant Disease Detection Using CNNs and GANs as an Augmenting Approach,”
May 2018.
[31] E. C. Too, L. Yujian, S. Njuki, and L. Yingchun, “A comparative study of fine-tuning deep learning models for plant disease identification,”
Computers and Electronics in Agriculture, vol. 161, pp. 272–279, Jun. 2019. https://doi.org/10.1016/j.compag.2018.03.032
[32] G. Geetharamani, P. J. Arun, M. Agarwalc, and S. K. Gupta, “Identification of plant leaf diseases using a nine-layer deep convolutional neural
network,” Computers and Electrical Engineering, vol. 76, pp. 323–338, Jun. 2019. https://doi.org/10.1016/j.compeleceng.2019.04.011
[33] Ü. Atila, M. Uçar, K. Akyol, and E. Uçar, “Plant leaf disease classification using EfficientNet deep learning model,” Ecological Informatics, vol.
61, Mar. 2021. https://doi.org/10.1016/j.ecoinf.2020.101182
[34] S. Vallabhajosyula, V. Sistla, and V. K. K. Kolli, “Transfer learning-based deep ensemble neural network for plant leaf disease detection,”
Journal of Plant Diseases and Protection, 2021. https://doi.org/10.1007/s41348-021-00465-8
[35] I. Ahmad, M. Hamid, S. Yousaf, S. T. Shah, and M. O. Ahmad, “Optimizing pretrained convolutional neural networks for tomato leaf disease
detection,” Complexity, vol. 2020, 2020. https://doi.org/10.1155/2020/8812019
[36] J. Chen, D. Zhang, and Y. A. Nanehkaran, “Identifying plant diseases using deep transfer learning and enhanced lightweight network,”
Multimedia Tools and Applications, vol. 79, no. 41–42, pp. 31497–31515, Nov. 2020. https://doi.org/10.1007/s11042-020-09669-w
[37] J. Chen, J. Chen, D. Zhang, Y. Sun, and Y. A. Nanehkaran, “Using deep transfer learning for image-based plant disease identification,”
Computers and Electronics in Agriculture, vol. 173, Jun. 2020. https://doi.org/10.1016/j.compag.2020.105393
[38] V. Tiwari, R. C. Joshi, and M. K. Dutta, “Dense convolutional neural networks based multiclass plant disease detection and classification using
leaf images,” Ecological Informatics, vol. 63, Jul. 2021. https://doi.org/10.1016/j.ecoinf.2021.101289
[39] R. Gajjar, N. Gajjar, V. J. Thakor, N. P. Patel, and S. Ruparelia, “Real-time detection and identification of plant leaf diseases using convolutional
neural networks on an embedded platform,” Visual Computer, 2021. https://doi.org/10.1007/s00371-021-02164-9
[40] J. Chen, D. Zhang, M. Suzauddola, and A. Zeb, “Identifying crop diseases using attention embedded MobileNet-V2 model,” Applied Soft
Computing, vol. 113, Dec. 2021. https://doi.org/10.1016/j.asoc.2021.107901
[41] E. Prasetyo, N. Suciati, and C. Fatichah, “Multi-level residual network VGGNet for fish species classification,” Journal of King Saud University
- Computer and Information Sciences, 2021. https://doi.org/10.1016/j.jksuci.2021.05.015
[42] V. Kumar, H. Arora, Harsh, and J. Sisodia, “ResNet-based approach for detection classification of plant leaf diseases,” in Proceedings of the
International Conference on Electronics and Sustainable Communication Systems (ICESC 2020), 2020, pp. 495–502.
https://doi.org/10.1109/ICESC48915.2020.9155585
[43] R. Sujatha, J. M. Chatterjee, N. Z. Jhanjhi, and S. N. Brohi, “Performance of deep learning vs machine learning in plant leaf disease detection,”
Microprocessors and Microsystems, vol. 80, Feb. 2021. https://doi.org/10.1016/j.micpro.2020.103615
[44] Ü. Atila, M. Uçar, K. Akyol, and E. Uçar, “Plant leaf disease classification using EfficientNet deep learning model,” Ecological Informatics, vol.
61, Mar. 2021. https://doi.org/10.1016/j.ecoinf.2020.101182
[45] V. Tiwari, R. C. Joshi, and M. K. Dutta, “Dense convolutional neural networks based multiclass plant disease detection and classification using
leaf images,” Ecological Informatics, vol. 63, Jul. 2021. https://doi.org/10.1016/j.ecoinf.2021.101289
[46] M. A. Moid and M. A. Chaurasia, “Transfer Learning-based Plant Disease Detection and Diagnosis System using Xception,” in Proceedings of
the 5th International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud), I-SMAC 2021, 2021, pp. 451–455.
https://doi.org/10.1109/I-SMAC52330.2021.9640694
Cite: D. Sutaji and H. Rosyid, “Convolutional Neural Network (CNN) Models for Crop Diseases Classification”, KINETIK, vol. 7, no. 2, May.
2022. https://doi.org/10.22219/kinetik.v7i2.1443
196 Kinetik: Game Technology, Information System, Computer Network, Computing, Electronics, and Control