Crop Leaf Disease Identification Using Deep Transfer Learning

Changjian Zhou , Yutong Zhang and Wenzhong Zhao

Abstract

Abstract: Traditional manual identification of crop leaf diseases is challenging. Owing to the limitations in manpower and resources, it is challenging to explore crop diseases on a large scale. The emergence of artificial intelligence technologies, particularly the extensive application of deep learning technologies, is expected to overcome these challenges and greatly improve the accuracy and efficiency of crop disease identification. Crop leaf disease identification models have been designed and trained using large-scale training data, enabling them to predict different categories of diseases from unlabeled crop leaves. However, these models, which possess strong feature representation capabilities, require substantial training data, and there is often a shortage of such datasets in practical farming scenarios. To address this issue and improve the feature learning abilities of models, this study proposes a deep transfer learning adaptation strategy. The novel proposed method aims to transfer the weights and parameters from pre-trained models in similar large-scale training datasets, such as ImageNet. ImageNet pre-trained weights are adopted and fine-tuned with the features of crop leaf diseases to improve prediction ability. In this study, we collected 16,060 crop leaf disease images, spanning 12 categories, for training. The experimental results demonstrate that an impressive accuracy of 98% is achieved using the proposed method on the transferred ResNet-50 model, thereby confirming the effectiveness of our transfer learning approach.

Keywords: Agricultural Artificial Intelligence , Crop Leaf Disease Identification , Plant Protection , Transfer Learning

1. Introduction

Experts and researchers have made significant efforts to improve grain production. The accurate identification of crop diseases is a crucial research topic in the field of plant protection. With the development of information technology, especially with the application of artificial intelligence (AI) in the agricultural field, the efficiency and accuracy of plant disease identification have greatly improved. AI-based disease identification models are typically divided into three stages: data acquisition, data analysis and model selection, and data testing and model evaluation. The remainder of this paper is organized to follow these three stages.

In recent years, plant disease identification has been divided into two branches. One uses conventional machine learning algorithms, such as support vector machine (SVM), K-means, and random forest. Bhimte and Thool [1] proposed an SVM-based cotton blade disease recognition model in India, in which they first used image processing methods to select cotton disease features and then extracted features, including texture, color, and shape, via a data augmentation method. Finally, the features were integrated into an SVM classifier to identify cotton diseases. Padilla et al. [2] focused on sugarcane leaf disease identification. They hypothesized that the yellow spot of sugarcane can be detected by machine learning methods, and based on their analysis, they proposed an SVM-based mode that can classify infected yellow spot leaves. Sharma et al. [3] analyzed the main factors affecting India’s agricultural situation and proposed a plant leaf disease detection method. This approach, which could quickly classify leaf diseases and aimed to increase grain yield, consisted of four steps: image collection, image preprocessing, image segmentation, and image classification.

These methods have achieved advanced performance in crop disease recognition. However, these methods are based on traditional machine learning methods that require manually extracted features as inputs. This limitation restricts the comprehensiveness of feature extraction, particularly in extensive planting environments. The development of deep learning technology has mitigated this problem. Deep learning models can handle large-scale data, which can be expressed by the automatic learning characteristics of the given data [4]. In recent years, deep-learning models have played an increasingly important role in crop disease identification. Hasan et al. [5] focused on yellow mosaic disease identification using deep convolutional neural network (CNN) methods. They used a dataset containing 600 yellow mosaic leaf images and classified them into two groups: one containing training data and one that was used for testing. This method achieved 96% accuracy without using any feature extraction methods. Barbedo [6] consider deep learning to be an important tool that can provide good solutions to plant disease classification problems. They analyzed the main factors affecting the models and created a leaf image database that could be used for academic purposes. Ashok et al. [7] proposed a deep architecture approach for plant leaf disease detection tasks. They first employed an image segmentation and clustering method and inputted the extracted features into the proposed models. A tomato leaf disease dataset was adopted for training, and a reliable, safe, and accurate system was developed based on the proposed method. Zhou et al. [8] developed a tomato image disease recognition approach using a reconstructed deep residual-dense network that achieved a satisfactory performance.

Hence, conventional machine-learning algorithms have been widely used in plant leaf disease recognition. Compared to manual observation methods, machine learning and deep learning-based methods have significantly improved accuracy and efficiency. However, these technologies have some limitations:

· Lack of sufficient data for training. Limited training data leads to deep learning models that cannot effectively exert feature representation ability and often suffer from overfitting.

· Some existing models have insufficient generalization abilities. When transferring a new crop category from pre-trained models, it is difficult to achieve satisfactory performance.

Furthermore, the lack of training data sharply degrades the diagnostic performance in new environments [9]. To mitigate these limitations, this study proposes a crop disease identification model that uses transfer learning algorithms. Specifically, pre-trained deep learning models used on large-scale datasets, such as ImageNet, were implemented, and the pre-trained models were used to perform downstream tasks. The ImageNet dataset contains approximately 14,197,122 images with 21,841 categories; therefore, the models trained on ImageNet can provide satisfactory performance. However, in practical agricultural settings, training with extensive datasets for downstream tasks is not feasible. Some deep learning models have access to only a few hundred training images, making it challenging to achieve the desired results.

In this study, transfer learning was implemented to transfer parameters and weights to new models. After fine-tuning and retraining on a specific dataset, the proposed method exhibited a stronger feature representation ability for identifying crop diseases. The ImageNet dataset is similar to crop leaf images. Therefore, it is feasible to train crop leaf disease models with the help of pre-trained parameters and weights trained on large-scale datasets (such as ImageNet). The main contributions of this study are as follows.

· A transfer learning adaptation strategy was proposed as the stem stage for selecting pre-trained models.

· The transfer learning method and deep learning models were combined to compensate for a lack of training data.

· Hyperparameters and weights were fine-tuned and optimized using the proposed method to improve the generalization ability.

2. Materials and Methods

2.1 Data Acquisition

To improve the model generalization ability, a dataset includes 16,060 images with three categories of potato leaf disease images and nine categories of tomato disease images are collected. Some of the plant leaf disease images were downloaded from AI challenger and part of them were manual shooting, all the images were resized as the [TeX:] $$196 \times 196$$ pixels. There are 12 categories of leaf disease images and parts of leaf disease images are illustrated in Fig. 1.

Fig. 1.
(a–l) Part of leaf images with 12 categories of leaf disease.
2.2 Related Work and Proposed Method

In this work, the transfer learning method is combined with the improved deep learning models. Since the ImageNet dataset is similar to the dataset used in crop leaf disease identification, consequently, the ImageNet based pre-training deep learning models were employed and fine-tuned in this study.

2.2.1 Transfer learning

Deep learning models exhibit strong capabilities of feature representation, but they require a largescale dataset and long training time, which are often not met in most cases. Fortunately, the transfer learning approaches can make up these deficiencies to the greatest extent. Importantly, transfer learning is a particularly method for image classification in deep CNN context. The pre-trained models can learn parameters and weights from the original data and applied in new related fields to solve a new problem, furthermore, it can achieve a wonderful performance when retraining and fine-tuning a new classifier [10].

The requirement of transfer learning is that it must find the feature relationships between the target and source domains, and find the connection of the two domains with common features [11]. In addition, the different features of the target and source domains should be used to improve model architectures to complete the new task of the target domain and realize the transferred knowledge. Specifically, transfer learning can be expressed by formula (1):

(1)
[TeX:] $$D=\{\chi, P(X)\},$$

where D denotes the source domain, χ is feature space, P(X) denotes marginal distribution, [TeX:] $$X=\left\{\chi_1, \chi_2, \ldots, \chi_n\right\} \in \chi.$$ The target task is shown in (2):

(2)
[TeX:] $$T=\{y, f(\cdot)\},$$

where y donotes label space [TeX:] $$f(\cdot).$$ denotes the target prediction function. The category and structure of transfer learning can be shown in Fig. 2.

Fig. 2.
The category and structure of transfer learning.

From Fig. 2 can be found that only when [TeX:] $$P_s \neq P_t$$ and [TeX:] $$P_s(y \mid x)=P_t(y \mid x),$$ the transfer learning can be adapted in the target domain. In crop leaf disease identification tasks, the marginal distribution of ImageNet and crop leaf diseases is different, which can be shown in (3):

(3)
[TeX:] $$P_{ImageNet } \neq P_{CropLeafDisease }$$

Obviously, the feature space and label space are the same distribution between ImageNet and crop leaf diseases, which is shown in (4).

(4)
[TeX:] $$\begin{aligned} & X_{ImageNet }=X_{CropLeafDisease } \\ & Y_{ImageNet }=Y_{CropLeafDisease } . \end{aligned}$$

From (4), the dataset on crop leaf diseases and ImageNet are isomorphic, consequently, the transfer learning is appropriate in this study. From (3) and (4), the property of domain adaptation in deep learning models between ImageNet and crop leaf diseases can work well in theory.

2.2.2 Deep learning architectures

In this study, several deep learning approaches were introduced and fine-tuned with transfer learning approach, such AlexNet [12], VGG-16 [13], Inception-ResNet-v2 [14], ResNet [15], DenseNet-121 [16], and so on, this section briefly introduced these models.

AlexNet had proved its powerful ability in 2012 ImageNet Challenge [17]. As the first used deep learning model this year, AlexNet had achieved more than 80% improvement than the other models, it also won the COCO competition [18], and the 2014 ImageNet image location competition. This model replaced all convolution kernels with [TeX:] $$3 \times 3$$ and [TeX:] $$2 \times 2$$ pooling cores, which made it achieve a stronger ability. VGG is the first-choice algorithm to extract features from images, however, the disadvantage is that the number of parameters as much as 140M, so it needs more storage space. Incepetion-Res-v2, inherited and developed from VGG, adopted two [TeX:] $$3 \times 3$$ convolutions instead of [TeX:] $$5 \times 5$$, and incorporated batch normalization, leading to unparalleled results. ResNet emerged as the champion in the ILSVRC 2015 competition. The primary enhancement in ResNet was the introduction of residual blocks, and the use of residual connections resolved the issue of gradient degradation, elevating it to one of the most widely adopted deep networks. DenseNet won the best paper in CVPR 2017. It is improved by introducing a better way of shortcut named density connection, this method can make the ability more robust and get a faster convergence speed. In addition, DenseNet has fewer parameters and faster training speed than ResNet.

2.2.3 Proposed method

The proposed transfer learning-based approach in this study is outlined as follows. Firstly, a pre-trained CNN model with large-scale datasets such as ImageNet was employed. Subsequently a knowledge transfer process was executed on the pre-trained models, transferring the parameters and weights to a new training dataset. During this knowledge transfer, adjustments were made to the overall structure and original convolutional layers of the model based on relevant parameters, to align with the requirements of the new learning models.

In this work, a transfer learning adaptation strategy was proposed. Firstly, the stem stage was presented for selecting pre-trained models, the data distribution and conditional probability between target task and source domains need to be investigated. Only when the feature space and label space are isomorphic and the marginal distribution between the two data distribution is different can the domain adaptation of transformer learning works well. Secondly, when selected the target pre-trained model, a series of parameter fine-tuning to adapt the new tasks is necessary. In this work, the size of input images, number of label categories, selection of activation function and loss function, learning rate, optimization function were fine-tuned. Lastly, a novel fully connected layer was designed to address crop leaf diseases identification tasks. The pre-trained deep learning models on ImageNet were trained using crop leaf diseases datasets. After training and parameter fine-tuning until the models converged to the desired results, the test data was employed for final evaluation. To improve the generalization ability of the proposed models, a normalization layer was incorporated into each model. Additionally, dropout layers were introduced to mitigate overfitting, and fully connected layers were integrated into the models to equip them with recognition capabilities. The structure of improved deep transfer learning models can be demonstrated in Fig. 3.

Fig. 3.
The structure of improved deep transfer deep learning models.

3. Experimental Results and Analysis

3.1 Evaluation Metrics

A series of evaluation metrics such as accuracy, precision, recall and F1-score were adopted in this work. Accuracy represents the proportion of the entire sample correctly classified by the classifier, while precision denotes the proportion of the normal examples correctly identified among those classified as normal by the classifiers. Recall is the proportion of the predicted positive examples to the total positive examples, and the F1-score is utilized as a degree of assignment average of accuracy and recall.

3.2 Implementation Details

The HPC platform was the computing environment while the operating system was Cent OS 7.5 with [TeX:] $$8 \times$$ Intel 4216 CPU, 32G RAM, [TeX:] $$2 \times$$ 2080Ti GPUs, and the Python with TensorFlow framework was employed. The same dataset and experiment environment were implemented on the models of AlexNet, VGG-16, Inception-ResNet-v2, ResNet-50, and DenseNet-121, all of them were trained by transfer learning pre-trained methods without transfer learning for comparison. The dataset was divided into training/testing/validation by 0.6/0.2/0.2, respectively. All deep learning models utilized the categorical cross-entropy loss function with an initial learning rate of 0.0001 and were trained for 200 epochs.

3.3 Results Analysis and Discussion

The training and validation accuracy in this experiment are detailed in Fig. 4. It can be found that all models utilizing transfer learning outperform those trained without transfer learning, providing strong evidence for the effectiveness of transfer learning in enhancing identification capabilities. Table 1 shows the identification results on the test set. It can be demonstrated that whether on validation set or test set, the transfer learning models outperform those without transfer learning models. The maximum improvement exceeds 10% in accuracy for ResNet-50, with DenseNet-121 achieving the highest accuracy of 98%. This provides compelling evidence that inheriting hyperparameters and weights from pre-trained models enhances the feature representation capability in the target domain. Different from training models from scratch, the transfer learning mechanism learns prior knowledge shared from pre-trained models through three approaches: instances, features, and parameters. In this work, the correlation between the source and target domains was investigated based on the instance, feature, and parameter branches, and a novel transfer learning adaptation strategy was presented with the purpose of selecting the most suitable pretrained models for target datasets. The experimental results demonstrate that transfer learning technology significantly enhances the performance of the original models, and offers a solution for addressing limited samples in recognition tasks. Furthermore, an amazing harvest is that the transfer learning approach can greatly shorten training time, and improve training efficiency. As can be seen from Table 1, the VGG-16 and AlexNet with less training time than the other models, that is because these models have a simple model architecture, make them reduce many complex calculation processes. Together, the ResNet-50 architecture achieved a perfect balance between identification accuracy and efficiency.

Fig. 4.
Training and validation results of Inception-ResNet-v2, ResNet-50, and DenseNet-121 between transfer learning (left column) and without transfer learning (right column).
Table 1.
Experimental results of transfer learning and without transfer learning models

It is worth noting that the proposed method introduces a stem stage for model selection compared to training from scratch. This stage involves analyzing the data distribution between the source and target domains and may require additional time. Nevertheless, it does not diminish the advantages offered by the proposed approach.

5. Conclusion

Timely identifying crop diseases and taking corresponding measures to prevent crops from being intruded is good practice for plant protection. It is anticipated that the silicon-based approaches will be widely used in crop disease identification as artificial intelligence technologies play an important role in more and more fields. Due to the limited availability of training data for crop diseases, deep learningbased models for crop disease identification frequently encounter issues of overfitting and unsatisfactory performance. To address these pitfalls, a transfer learning adaptation strategy was proposed in this study, the target and source domains were analyzed in detail, and it was found that the problems can be solved in this study by adapting ImageNet for pre-training. Therefore, the deep transfer learning-based models with pre-trained on ImageNet were employed to solve the low crop leaf disease identification accuracy problem. Specifically, the parameters and weights were transferred from the pre-trained models, and a full connection layer was added to solve crop leaf disease identification tasks. In addition, a dropout layer was added to mitigate overfitting. To enhance the generalization capabilities of the proposed models, a normalization layer was introduced to shape the network architecture. Experimental results demonstrated that transfer learning-based approaches significantly improved the performance compared to training from scratch across 5 evaluation metrics. The presented method achieved satisfactory results and outperformed existing models, especially in the case of ResNet-50, where accuracy improved from 0.86 to 0.97 with the proposed method. Importantly, this model provides solutions for further control of plant diseases and preventing crops from disease threat. Nonetheless, the use of transfer learning does not come without some less conspicuous advantages, as evidenced by a slight decline in the recall of InceptionResNet-v2 from 0.83 to 0.82. This suggests that transfer learning may not comprehensively enhance performance. One possible reason could be incomplete adaptation of pre-trained data to the target dataset, a matter to be investigated in future work.

Biography

Changjian Zhou
https://orcid.org/0000-0002-2094-6405

He received M.S. degree in Harbin Engineering University in 2012, China. He is as a director in Department of Data and Computing, Northeast Agricultural University. His current research interests include agricultural artificial intelligence and computer vision.

Biography

Yutong Zhang
https://orcid.org/0000-0003-1012-3621

She is currently pursuing the M.S. degree in School of Computer and Software, Hohai University. Her research interests include image processing and agricultural artificial intelligence.

Biography

Wenzhong Zhao
https://orcid.org/0000-0002-9644-8592

He received a M.S. degree in Northeast Agricultural University in 2008, China. He is currently an associate researcher now. His research interests include agricultural informatization and data analysis.

References

  • 1 N. R. Bhimte and V . R. Thool, "Diseases detection of cotton leaf spot using image processing and SVM classifier," in Proceedings of 2018 2nd International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 2018, pp. 340-344. https://doi.org/10.1109/ICCONS.2018.8662906doi:[[[10.1109/ICCONS.2018.8662906]]]
  • 2 D. A. Padilla, G. V . Magwili, A. L. A. Marohom, C. M. G. Co, J. C. C. Gano, and J. M. U. Tuazon, "Portable yellow spot disease identifier on sugarcane leaf via image processing using support vector machine," in Proceedings of 2019 5th International Conference on Control, Automation and Robotics (ICCAR), Beijing, China, 2019, pp. 901-905. https://doi.org/10.1109/ICCAR.2019.8813495doi:[[[10.1109/ICCAR.2019.8813495]]]
  • 3 P. Sharma, P. Hans, and S. C. Gupta, "Classification of plant leaf diseases using machine learning and image preprocessing techniques," in Proceedings of 2020 10th International Conference on Cloud Computing, Data Science & Engineering (Confluence), Noida, India, 2020, pp. 480-484. https://doi.org/10.1109/Confluence47617.2020.9057889doi:[[[10.1109/Confluence47617.2020.9057889]]]
  • 4 C. Zhou, J. Song, S. Zhou, Z. Zhang, and J. Xing, "COVID-19 detection based on image regrouping and ResNet-SVM using chest X-ray images," IEEE Access, vol. 9, pp. 81902-81912, 2021. https://doi.org/10.1109/ACCESS.2021.3086229doi:[[[10.1109/ACCESS.2021.3086229]]]
  • 5 M. Z. Hasan, M. S. Ahamed, A. Rakshit, and K. Z. Hasan, "Recognition of jute diseases by leaf image classification using convolutional neural network," in Proceedings of 2019 10th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kanpur, India, 2019, pp. 1-5. https://doi.org/10.1109/ICCCNT45670.2019.8944907doi:[[[10.1109/ICCCNT45670.2019.8944907]]]
  • 6 J. G. Barbedo, "Factors influencing the use of deep learning for plant disease recognition," Biosystems Engineering, vol. 172, pp. 84-91, 2018. https://doi.org/10.1016/j.biosystemseng.2018.05.013doi:[[[10.1016/j.biosystemseng.2018.05.013]]]
  • 7 S. Ashok, G. Kishore, V. Rajesh, S. Suchitra, S. G. Sophia, and B. Pavithra, "Tomato leaf disease detection using deep learning techniques," in Proceedings of 2020 5th International Conference on Communication and Electronics Systems (ICCES), Coimbatore, India, 2020, pp. 979-983. https://doi.org/10.1109/ICCES48766.2020.9137986doi:[[[10.1109/ICCES48766.2020.9137986]]]
  • 8 C. Zhou, S. Zhou, J. Xing, and J. Song, "Tomato leaf disease identification by restructured deep residual dense network," IEEE Access, vol. 9, pp. 28822-28831, 2021. https://doi.org/10.1109/ACCESS.2021.3058947doi:[[[10.1109/ACCESS.2021.3058947]]]
  • 9 Q. H. Cap, H. Uga, S. Kagiwada, and H. Iyatomi, "LeafGAN: an effective data augmentation method for practical plant disease diagnosis," IEEE Transactions on Automation Science and Engineering, vol. 19, no. 2, pp. 1258-1267, 2022. https://doi.org/10.1109/TASE.2020.3041499doi:[[[10.1109/TASE.2020.3041499]]]
  • 10 P . Zhao, T. Wu, S. Zhao, and H. Liu, "Robust transfer learning based on geometric mean metric learning," Knowledge-Based Systems, vol. 227, article no. 107227, 2021. https://doi.org/10.1016/j.knosys.2021.107227doi:[[[10.1016/j.knosys.2021.107227]]]
  • 11 S. Coulibaly, B. Kamsu-Foguem, D. Kamissoko, and D. Traore, "Deep neural networks with transfer learning in millet crop images," Computers in Industry, vol. 108, pp. 115-120, 2019. https://doi.org/10.1016/j.compind.2019.02.003doi:[[[10.1016/j.compind.2019.02.003]]]
  • 12 A. Abd Almisreb, N. Jamil, and N. M. Din, "Utilizing AlexNet deep transfer learning for ear recognition," in Proceedings of 2018 4th International Conference on Information Retrieval and Knowledge Management (CAMP), Kota Kinabalu, Malaysia, 2018, pp. 1-5. https://doi.org/10.1109/INFRKM.2018.8464769doi:[[[10.1109/INFRKM.2018.8464769]]]
  • 13 X. Zhang, J. Zou, K. He, and J. Sun, "Accelerating very deep convolutional networks for classification and detection," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38, no. 10, pp. 1943-1955, 2016. https://doi.org/10.1109/TPAMI.2015.2502579doi:[[[10.1109/TPAMI.2015.2502579]]]
  • 14 T. Emara, H. M. Afify, F. H. Ismail, and A. E. Hassanien, "A modified Inception-v4 for imbalanced skin cancer classification dataset," in Proceedings of 2019 14th International Conference on Computer Engineering and Systems (ICCES), Cairo, Egypt, 2019, pp. 28-33. https://doi.org/10.1109/ICCES48960.2019.9068110doi:[[[10.1109/ICCES48960.2019.9068110]]]
  • 15 K. He, X. Zhang, S. Ren, and J. Sun, "Deep residual learning for image recognition," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las V egas, NV , USA, 2016, pp. 770778. https://doi.org/10.1109/CVPR.2016.90doi:[[[10.1109/CVPR.2016.90]]]
  • 16 G. Huang, Z. Liu, L. V an Der Maaten, and K. G. Weinberger, "Densely connected convolutional networks," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 2017, pp. 2261-2269. https://doi.org/10.1109/CVPR.2017.243doi:[[[10.1109/CVPR.2017.243]]]
  • 17 Stanford Vision Lab, "ImageNet large scale visual recognition challenge 2012 (ILSVRC2012)," 2020 (Online). Available: https://image-net.org/challenges/LSVRC/2012/.custom:[[[https://image-net.org/challenges/LSVRC/2012/]]]
  • 18 J. Long, E. Shelhamer, and T. Darrell, "Fully convolutional networks for semantic segmentation," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 4, pp. 640-651, 2017. https://doi.org/10.1109/TPAMI.2016.2572683doi:[[[10.1109/TPAMI.2016.2572683]]]

Table 1.

Experimental results of transfer learning and without transfer learning models
Models Precision Recall F1-score Accuracy Training time (hr)
AlexNet Without transfer learning 0.87368s42 0.76146784 0.8137254 0.8611 8.5
Transfer learning 0.8631578 0.85416665 0.8586387 0.8715 6
ResNet-50 Without transfer learning 0.9052631 0.83495142 0.8686868 0.8618 9.2
Transfer learning 0.9559255 0.92983091 0.9403476 0.9738 6.5
DenseNet-121 Without transfer learning 0.9166443 0.86148396 0.8787736 0.9343 11
Transfer learning 0.9638319 0.90897447 0.9212991 0.9785 8.4
VGG-16 Without transfer learning 0.8171744 0.77256984 0.7811751 0.8925 9
Transfer learning 0.9690077 0.94551516 0.95466834 0.9275 5.5
Inception-ResNet-v2 Without transfer learning 0.89529735 0.8304236 0.8457864 0.9240 10.5
Transfer learning 0.85986096 0.82081676 0.8255682 0.9592 9
(a–l) Part of leaf images with 12 categories of leaf disease.
The category and structure of transfer learning.
The structure of improved deep transfer deep learning models.
Training and validation results of Inception-ResNet-v2, ResNet-50, and DenseNet-121 between transfer learning (left column) and without transfer learning (right column).