Original Articles

Journal of Bio-Environment Control. 31 January 2024. 1-11
https://doi.org/10.12791/KSBEC.2024.33.1.001

ABSTRACT


MAIN

  • Introduction

  • Materials and Methods

  •   1. Data set

  •   2. Orange 3 software

  •   3. Selection of ideal model

  • Results and Discussion

  •   1. Data set

  •   2. Generating the workflow

  •   3. Training of models

  •   4. Selection of better performing AI model

  • Conclusion

Introduction

Tea, one of the most common and widely consumed beverages worldwide, is derived from the leaves of Camellia sinensis. Similar to other agricultural crops, tea leaves are susceptible to various diseases (Nanehkaran et al., 2020). Proper diagnosis of the disease is crucial to control and manage tea leaf diseases (Chakruno et al., 2022). However, distinguishing the symptoms of a particular disease is challenging as many diseases share similar characteristics and, require considerable time. As a solution, ongoing research in plant pathology and advances in disease identification techniques have provided new tools for the fast and precise detection of tea leaf diseases (Keith et al., 2006).

Deep learning, a subcategory of machine learning, is one of the most accurate and efficient methods that has revolutionized many industries, including agriculture, by providing advanced techniques for various tasks, such as disease detection and diagnosis (Hu et al., 2019). Recently, deep learning has gained considerable attention in tea cultivation for the detection of leaf diseases (Shruthi et al., 2019). Tea leaf disease recognition using deep learning generally uses a well-marked dataset of tea leaf images to train a deep learning model for tea leaf disease recognition. The model automatically distinguishes between healthy and diseased leaves, providing a precise disease identification method (Sladojevic et al., 2016).

Implementation of deep learning for tea leaf disease detection involves many challenges, including the need for large high-quality datasets, access to computational resources, and model deployment in real-world farming environments. The study of Barbedo (2016) highlights the limitations of disease identification through images, including busy backgrounds, unclear borders, and variations in capture conditions. However, data augmentation techniques can overcome these issues (Mikołajczyk and Grochowski, 2018). Furthermore, the high complexity and non-linearity of deep neural networks make it difficult to provide meaningful explanations of the predictions (Shrestha and Mahmood, 2019).

Orange 3 is an open-source data visualization and analysis software that includes a comprehensive set of machine learning algorithms designed for data mining, machine learning, and data science tasks (Demšar et al., 2013). Orange 3 provides a visual workflow that enables the users to deal with data interactively, perform data analysis, and create machine learning models without any advanced programming knowledge (Demšar and Zupan, 2012). New add-ons can be added to the orange 3 canvas as required (Mohapatra and Swarnkar, 2021). The image analytics add-on was used in this study includes import images, image viewer, image embedding, image grid, and save images widgets for image analysis.

However, it is crucial to identify the suitability of Orange 3 for scientific research and papers. Ratra and Gulia (2020), has evaluated the open-source data mining tools and suggested that Orange 3 is one of the effective software for data mining. Furthermore, they have suggested that results can be different depending on datasets or algorithms. Vaishnav and Rao (2018) utilized Orange 3 for fruit classification, highlighting its versatility in combining model training, testing, data preprocessing, and visualization in a single software package for scientific research. Furthermore, Ishak et al. (2020) have used orange 3 for their scientific studies.

In this study, the five AI models were taken, namely, Inception v3, SqueezeNet (local), VGG-16, Painters, and DeepLoc. Google’s Inception v3 model was trained using ImageNet. Information in the Inception-v3 [4] model can recognize 1000 classes in the ImageNet (Tiwari et al., 2022). Squeeznet model is a quick and small image recognition model built with ImageNet. It takes less bandwidth for this model to export new models to the cloud, and its tiny architecture makes it simpler to use on FPGA devices and other systems with memory constraints (Hidayatuloh et al., 2018). Two deep neural networks, VGG16 and VGG19, were developed by the Visual Geometry Group at the University of Oxford to help with image identification that was trained using the ImageNet dataset. Another embedder in Image Embedding is Painters, the winner of Kaggle’s Painter by Numbers competition. It was trained using 1,584 artists’ works from 79,433 pictures. A neural network called DeepLoc was trained with 21,882 single-cell images, each of which was uniquely labeled as belonging to one of 15 localization compartments (Tiwari et al., 2022). However, neural network models may not have the same accuracy level. Therefore, the performance metrics; F1, precision, recall, and confusion matrix can be used to evaluate models, determining accuracy and efficiency (Vaishnav and Rao, 2018).

In this study, we aimed to identify the most accurate deep learning model among the aforementioned AI models for the diagnosis of tea leaf diseases using Orange 3. For that, F1, precision, and recall values are taken. Further, the confusion matrix of the most ideal model is selected for further interpretation of images where the correctly classified and misclassified incidences.

Materials and Methods

1. Data set

In this study, we used tea (Camellia sinensis) leaf disease images. Kaggle website (Kaggle Data Science Company, 2017) was used to obtain the dataset and from there “tea sickness dataset” (Kimutai and Förster, 2022) was obtained for this study. Information on the number of images in each category of tea leaf disease is presented in Table 1. With a total of 885 images, the dataset included eight categories of tea leaf diseases: healthy, algal leaf spot, anthracnose, bird’s eye spot, brown blight, gray blight, red leaf spot, and white spot. The classes with the highest number of images were the red leaf spot and white spot, with 143 and 141 images, respectively. The class with the lowest number of images was healthy, containing only 74 images. To assess the implementation of the machine learning model, the dataset was split into two parts, with 700 images used for training and 180 images used for validation. Images of the input dataset are shown in Fig. 1.

https://static.apub.kr/journalsite/sites/phpf/2024-033-01/N0090330101/images/phpf_33_01_01_F1.jpg
Fig. 1.

Input Dataset after classifying images into eight categories of tea leaves diseases: Gray blight, Healthy, Red leaf spot, White leaf spot, Algal leaf spot, Anthracnose, Bird’s eye spot, and Brown blight. Three images are shown in each category.

Table 1.

Number of images in each category of the input dataset.

Disease Number of Images
Algal leaf spot 110
Anthracnose 100
Bird’s eye spot 100
Brown blight 112
Gray blight 100
Healthy 74
Red leaf spot 143
White spot 141

Total images in 8 class: 880

Images used for training: 700

Images used for validation: 180

2. Orange 3 software

The basic settings of the orange 3 program were adjusted according to the requirements of this experiment before image analysis. All models were trained using the Adam solver, rectified linear unit (ReLU) activation function, 100 neurons in hidden layers, 200 maximum iterations, in the neural network, and 0.0001 regularizations. 700 total number of images were used to train the models. An image analytics add-on was also used in this study. For the training model, the import image, image embedding, neural network, test and score, and confusion matrix widgets were used, whereas for the predictions, the import images, image embedding, predictions, and image viewer widgets were used (Fig. 2). The precision, F1, and recall of the five deep learning models (Inception v3, SqueezeNet [local], VGG-16, Painters, and DeepLoc) were calculated to evaluate their accuracy of detection.

https://static.apub.kr/journalsite/sites/phpf/2024-033-01/N0090330101/images/phpf_33_01_01_F2.jpg
Fig. 2.

Work flow of Orange 3 was generated by connecting widgets after installing the image analytics add-on. Widgets were connected for training and for testing the image dataset.

3. Selection of ideal model

To select the ideal model and estimate its performance, three performance evaluation metrics (precision, recall, and F1 score) were recorded (Table 2). The formulas of these metrics are provided in (1), (2), and (3), respectively, as previously reported (Tripathi, 2021).

(1)
Precision=TPTP+FP
(2)
Recall=TPTP+FN

(TP: True Positive, FP: False Positive, FN: False Negative)

(3)
F1Score=2×Precision×RecallPrecison+Recall
Table 2.

Performance matrix (F1, precision, and recall) of each model in Orange 3.

Model F1 Precision Recall
Inception v3 0.806 0.807 0.807
SqueezeNet (local) 0.901 0.901 0.901
VGG-16 0.779 0.780 0.780
Painters 0.799 0.800 0.799
DeepLoc 0.770 0.771 0.771

The number of true positive labels when positive labels are offered is referred to as precision. Recall refers to the number of instances that are accurately classified as positive. The overall performance of the model was assessed using the F1 score, which is an evaluation metric that considers both precision and recall. A false positive refers to a situation in which a false response is msitaken to be true, whereas a true positive refers to a situation in which a false answer is mistaken as true. Conversely, a true negative reflects situations in which a false response is predicted to be true. A false negative represents a situation in which the correct answer is predicted as false.

In this experiment most ideal model was selected based on the model’s F1, precision, and recall values which have a value closer to one and also the highest value compared to all other models. Further, the ideal AI model which has the best performance matrix was selected based on the TP, TN, FP, and FN number of images.

Results and Discussion

1. Data set

In this study five different AI models, namely, Inception v3, SqueezeNet (local), VGG-16, Painters, and DeepLoc were used to detect different types of tea leaves diseases from the dataset as these models are provided from the orange 3 software. In the dataset, there were eight categories of tea leaf diseases which are characterized by different types of symptoms as shown in Fig. 1. Different numbers of images in each class were used for this study as shown in Table 1. However, it is crucial to provide similar conditions for each AI model to avoid biased results. Therefore, before implementing image classification, data augmentation is a technique for addressing issues like insufficient training data or unequal class distribution within datasets (Mikołajczyk and Grochowski, 2018). Also, it is a commonly used method in machine learning and computer vision research for creating new iterations of an existing dataset and is especially helpful when the model needs to learn various iterations of the same object or when the initial dataset is small (Singh et al., 2020). By making numerous changes to the original photos, data augmentation can assist in improving the model’s robustness and generalization and produce better results. Moreover, it can aid in balancing the dataset (Shi et al., 2022). In this study, we followed a procedure that is mostly similar to the study conducted by Singh et al. (2023), and they also used different numbers of tea leaf images in each class and data augmentation was performed to their dataset to balance the dataset which is consistent with our study.

2. Generating the workflow

The workflow of Orange 3 can be generated by using widgets as required. In this study, the image analytics add-on was installed for image classification, as previously described (Demšar et al., 2013). As shown in Fig. 2, the workflow consists of two sections for dataset training and testing. The import images widget was used to upload the images, and the image viewer was used to display the uploaded images. The image embedding widget was used for extraction by uploading the images from the dataset to the server using a particular model to detect the features of each image and provide the feature vectors as a confusion matrix, which extracts features from the given dataset along with size, height, and width as key features. Moreover, these main features are used by the algorithms to learn other specific features in each image (Ishak et al., 2020). Furthermore, Adam solver, ReLU activation function, 100 neurons in hidden layers, 200 maximum iterations in the neural network, and 0.0001 regularizations were adjusted before implementing the training of models in this study.

3. Training of models

All models were trained using a training dataset which consists of well-marked images. To train each model, the number of neurons in hidden layers in the neural network model is one of the important factors for better functioning of AI models. In this study, under the ‘neural network’ widget, the number of neurons was adjusted to 100 in the hidden layers. Generally, a simple neural network comprises an input layer, an output layer, and one or more hidden layers. Each node in the input layer is connected to a node in the hidden layer, and each node is connected to a node in the output layer (Guo et al., 2017). The convolutional neural network (CNN) model, which consists of one input layer, four convolution layers, and two fully connected layers has been proposed by Latha et al. (2021). The input layer receives the image, and the output layer divides it into different classes. Convolution layers primarily extract features from the input images in the dataset. The performance of the model can be further enhanced by varying the number of layers, learning rate parameters, and optimizers used. Datta and Gupta (2023) proposed the use of multiple hidden layers in the CNN model for the accurate identification of disease classes. The deep learning models performed exceedingly well mainly because the hidden layers have a profound understanding of the immense amount of data (Kansara and Sawant, 2020). However, the accuracy of the model can differ with the number of hidden layers in the neural network. Raut and Dani (2020) have found that the number of hidden layers in the neural network has a significant effect on the model’s accuracy. They show that the accuracy of the model rises gradually for a particular number of layers and then decreases sharply once saturation is reached. Therefore, it is crucial to determine the proper number of neurons for the initial hidden layer, as it negates the need for additional hidden layers. Similarly, increasing the number of neurons or hidden layers results in making the network more complex (Shafi et al., 2006). In this study, we used the maximum available number of neurons (100) which is consistent with Uzair and Jamil (2020) study to avoid inaccuracies in complex situations due to improperly trained neural networks. However, it is needed to evaluate the optimum number of neurons for the neural network model to detect tea leaf diseases in future studies. Moreover, it is currently uncertain how to calculate the number of hidden layers and the number of neurons in each hidden layer to achieve high accuracy (Raut and Dani, 2020).

The complexity or smoothness of the model is penalized by regularization, allowing good generalization to unobserved data, even when training on a limited training set or with insufficient iterations. Therefore, regularization is a technique intended to enhance the generalization ability of a model by reducing overfitting (Tian and Zhang, 2022). Better performance on the training dataset and better predictions on new data indicate good generalization ability of the model (Neyshabur et al., 2017). In this study, regularization was adjusted to 0.0001 for better functioning of AI models. Though the regularization techniques might effectively address overfitting and underfitting issues; nonetheless, neural network models continued to experience issues during training despite the application of certain regularization values. Therefore, the problems are difficult to overcome, and more future research work is needed to overcome these issues (Nusrat and Jang, 2018).

An activation function is required for the hidden layer of a neural network to introduce non-linearity. Without an activation function, the neural network will be similar to plain perception, and the effect will not be powerful with linear functions. Linear, threshold, and sigmoid functions can be used as activation functions. Because it combines nearly linear, curvilinear, and nearly constant behaviors depending on the input value, the sigmoid activation function is frequently used for hidden layers (Mirza, 2018). SUM is a collection of hidden layer output nodes that have been multiplied by connection weights, summed to produce a single value, and passed through the activation function (Fig. 3). The output of a sigmoid function can only be a number between 0 and 1, whereas the input can be any value between negative and positive infinity (Sharma et al., 2020). Since the activation range is limited, blowing up during the activations is easy to avoid. Therefore, this is the most commonly used activation function. However, according to some research studies, the sigmoid function is not ideal because approaching either end of the function’s range, Y values have a tendency to respond slightly to changes in X values (Szandała, 2021). Compared with equivalents with saturating non-linearity, deep neural networks using ReLU train substantially more quickly. Large models trained on large datasets perform significantly better with quick learning. ReLU simply retains the positive component while eliminating all negative values and replacing them with 0. As the derivative of the positive part of this non-linear function is constant, it is superior to saturating non-linearity. Therefore, ReLU is not affected by vanishing gradients (Jiang et al., 2018). In this study, the ReLU activation function was used. However, Sibi et al., 2013 have shown that the performances of different activation functions are not significantly different and have approximately the same effect on a successfully trained network.

https://static.apub.kr/journalsite/sites/phpf/2024-033-01/N0090330101/images/phpf_33_01_01_F3.jpg
Fig. 3.

The Flow of signal through activation function created based on the study of Sibi et al., 2013.

In order to calculate the Precision score and identify the type of tea leaf disease, we employed cross-validation for the predictive performance evaluation of the six AI models. When utilizing supervised learning for categorization, the cross-validation approach is recommended (Vaishnav and Rao, 2018). Supervised learning is used if fewer data points with well-marked training data are available. It deals with the training of models utilizing previously collected, prior known data (Mahesh, 2020). In this study well-marked training data set and test data set were used to determine the most accurate model to detect tea leaf diseases. Unsupervised learning often yields superior outcomes for large datasets (Mahesh, 2020).

4. Selection of better performing AI model

To select the best model, different values obtained were considered. For instance, precision, recall, F1 value, and the confusion matrix’s performance were evaluated in this study. From of five AI models, the precisions of the neural networks of Inception v3, SqueezeNet (local), VGG-16, Painters, and DeepLoc were 0.807, 0.901, 0.780, 0.800, and 0.771, respectively. F1 score and recall values are listed in Table 2. F1, precision, and recall values are taken based on the number of true positive, true negative, false positive, and false negative images taken from the confusion matrix. The number of true positive labels when positive labels are offered is referred to as precision whereas recall refers to the number of instances that are accurately classified as positive. The overall performance of the model was assessed using the F1 score, which is an evaluation metric that considers both precision and recall. Therefore it is important to select a model that shows the highest F1, recall, and precision value (Tiwari et al., 2022). From the overall results, the SqueezeNet (local) model was selected as the optimal AI model for tea disease detection using tea leaf images owing to its high precision, F1, and recall values.

A confusion matrix was obtained in this study to determine the ideal model for tea leaf disease classification. The number/proportion of instances between the predicted and actual classes is shown in the confusion matrix which provides information on the functioning of AI models. Fig. 4 is the confusion matrix of the SqueezeNet model after performing cross-validation for clear understanding. Moreover, the confusion matrix can be used to calculate the prediction errors of the model. Off-diagonal matrix elements define the prediction error, whereas diagonal elements determine the accuracy of the prediction (Demšar et al., 2013). The confusion matrix of an effective model with better performance typically has high diagonal values. In this study, the predictions of the SqueezeNet model were successful, as indicated by the high diagonal values. In cases where the model made incorrect predictions, algal leaf spots were frequently misclassified as white spots, anthracnose as gray blight, bird’s eye spots as anthracnose, brown blight as white spots, gray blight as anthracnose, healthy leaves as red leaf spots, red leaf spots as white spots, and white spots as brown blights. Moreover, misclassified predictions were observed in off-diagonal cells. For instance, in this confusion matrix, of the 95 actual algal leaf spot images, the system predicted that 9 were brown blight disease, 1 image was Gray blight disease, 8 were red leaf spot disease, 13 were white spot disease, and of the 95 anthracnose images, it predicted that one was algal leaf spot. Further, 5 images of brown blight, 1 image of Gray blight, 1 image of healthy leaves, 6 images of red leaf spot, and 10 images of white spot disease were incorrectly predicted as algal leaf spot. However, these misclassifications were considerably low in the SqueezeNet model compared to other AI models. Therefore, the SqueezeNet model was selected as an ideal model for the classification of tea leaf diseases as it has good performance throughout the confusion matrix in this study (Fig. 4).

https://static.apub.kr/journalsite/sites/phpf/2024-033-01/N0090330101/images/phpf_33_01_01_F4.jpg
Fig. 4.

Confusion matrix of the most accurate model: SqueezNet for the detection of tea leaf diseases. The diagonal elements show the number of accurately predicted images while off-diagonal matrix elements define the number of incorrectly predicted images.

The number of True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) images can be calculated from the confusion matrix for further interpretation (Fig. 5). Also, using those values, the classification accuracy of each image category can be calculated by a formula i.e. (TP + TN) / (P + N) where the “P” represents the actual positive (TP + FN) and N represents the actual negative (FP + TN) (Patro and Patra, 2014). However, in this study, accuracy was not calculated as the results were taken only from the software-analyzed data which is precision, F1, recall, and confusion matrix data. In future studies, it is more appropriate to calculate the accuracy of each model as well. In this study, the number of TP, TN, FP, and FN images was calculated only for the algal leaf spot images from the data obtained from the most ideal model, the SqueezeNet model’s confusion matrix, and shown in Fig. 6 for more clarification.

https://static.apub.kr/journalsite/sites/phpf/2024-033-01/N0090330101/images/phpf_33_01_01_F5.jpg
Fig. 5.

The way of obtaining images relevant to True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) for a particular selected disease category (in this Figure, D3 has selected) from a confusion matrix. D1 to D8 represents each image category taken for this study (D1 - algal leaf spot, D2 - anthracnose, D3 - bird’s eye spot, D4 - brown blight, D5 - Gray blight, D6 - Healthy, D7 - red leaf spot, and D8 - white spot). This Figure was designed based on the study of Patro and Patra, 2014.

https://static.apub.kr/journalsite/sites/phpf/2024-033-01/N0090330101/images/phpf_33_01_01_F6.jpg
Fig. 6.

Number of True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) images calculated for algal leaf spot disease from the data obtained from the confusion matrix of most ideal model, SqueezeNet.

However, some researchers have suggested that Inception V3 is an effective model for image classification. For example, Chugh et al. (2020)’s potato plant diseases categorization and, Xia et al. (2017)’s flower classification results have shown the highest accuracy of Inception V3 in image classification. However, in this study, Inception V3 has shown considerable precision compared to other models but lower precision reported than the SqueezeNet (local) model. Khan et al. (2021) suggested using SqueezeNet, a deep learning-based method, to classify six distinct diseases that significantly affect the quality and yield of citrus fruits. Moreover, among the two different deep learning models they have used, SqueezeNet and MobileNetV2, the SqueezeNet model performed more satisfactorily than MobileNetV2, with an accuracy of 96% which is consistent with our results. Similarly, Tiwari et al. (2022) have performed image embedding and classification using pre-trained deep learning architectures, namely, Inception V3, VGG16, VGG19, Painter, SqueezeNet, and DeepLoc. Results have shown that the SqueeseNet model performed well resulting in the highest accuracy of 95.14%. This fact was also proven by high precision, recall, and F1 score values of the SqueeseNet model which is mostly similar to our study results. Moreover, from the results of our study, SqueezeNet, Inception V3, Painters, VGG-16, and DeepLoc are the ascending order of AI models aligned according to the precision value obtained from each model, which the mostly similar pattern was obtained from the study conducted by Tiwari et al. (2022).

Conclusion

In conclusion, the precisions of the neural networks of the five AI models, namely, Inception v3, SqueezeNet [local], VGG-16, Painters, and DeepLoc were 0.807, 0.901, 0.780, 0.800, and 0.771, respectively. Performance metrics of each model were taken. From the results, the SqueezeNet (local) model was selected as the optimal AI model for tea disease detection using tea leaf images owing to its high level of precision, F1, recall values, and good performance throughout the confusion matrix. Our results suggest that the Orange 3 visual programming software is a feasible and convenient software package for deep learning-based disease detection and is beneficial for agricultural development. Further, the SqueezeNet model is suggested as the most ideal AI model for tea leaf disease detection through images. In future studies, it is more appropriate to use the same AI models to examine the accuracy of the models for various diseases of different plants, and it is more crucial to investigate how different image types affect the functioning of AI models.

Acknowledgements

This work was supported by Korea Institute of Planning and Evaluation for Technology in Food, Agriculture and Forestry (IPET) and Korea Smart Fam R&D Foundation (KosFarm) through Smart Farm Innovation Technology Development Progarm, funded by Ministry of Agriculture, Food and Rural Affairs (MAFRA) and Ministry of Science and ICT (MSIT), Rural Development Administration (RDA) (421009043HD020).

References

1
Barbedo J.G.A. 2016, A review on the main challenges in automatic plant disease identification based on visible range images. Biosyst Eng 144:52-60. doi:10.1016/j.biosystemseng.2016.01.017 10.1016/j.biosystemseng.2016.01.017
2
Chakruno P., S. Banik, and K. Sumi 2022, Important diseases of tea (Camellia sinensis L.) and their integrated management. In Diseases of Horticultural Crops: Diagnosis and Management, vol 4. Apple Academic Press, USA, pp 119-138. doi:10.1201/9781003160472-7 10.1201/9781003160472-7
3
Chugh G., A. Sharma, P. Choudhary, and R. Khanna 2020, Potato leaf disease detection using InceptionV3. Int Res J Eng Technol 7:1363-1366.
4
Datta S., and N. Gupta 2023, A novel approach for the detection of tea leaf disease using deep neural network. Procedia Comput Sci 218:2273-2286. doi:10.1016/j.procs.2023.01.203 10.1016/j.procs.2023.01.203
5
Demšar J., and B. Zupan 2012, Orange data mining fruitful and fun. Inf Družba IS 6:1-486.
6
Demšar J., T. Curk, A. Erjavec, C. Gorup, T. Hocevar, M. Milutinovic, M. Mozina, M. Polajnar, M. Toplak, and A. Staric, et al. 2013, Orange data mining toolbox in Python. J Mach Learn Res 14:2349-2353.
7
Guo T., J. Dong, H. Li, and Y. Gao 2017, Simple convolutional neural network on image classification. In IEEE 2017 2nd International Conference on Big Data Analysis (ICBDA), pp 721-724. doi:10.1109/ICBDA.2017.8078730 10.1109/ICBDA.2017.8078730
8
Hidayatuloh A., M. Nursalman, and E. Nugraha 2018, Identification of tomato plant diseases by leaf image using Squeezenet model. In 2018 International Conference on Information Technology Systems and Innovation (ICITSI), pp 199-204. doi:10.1109/ICITSI.2018.8696087 10.1109/ICITSI.2018.8696087
9
Hu G., X. Yang, Y. Zhang, and M. Wan 2019, Identification of tea leaf diseases by using an improved deep convolutional neural network. Sustain Comput Inform Syst 24:100353. doi:10.1016/j.suscom.2019.100353 10.1016/j.suscom.2019.100353
10
Ishak A., K. Siregar, R. Ginting, and M. Afif 2020, Orange software usage in data mining classification method on the dataset lenses. In IOP Conference Series: Materials Science and Engineering (IOP Publishing) 1003(1):012113. doi:10.1088/1757-899X/1003/1/012113 10.1088/1757-899X/1003/1/012113
11
Jiang X., Y. Pang, X. Li, J. Pan, and Y. Xie 2018, Deep neural networks with elastic rectified linear units for object recognition. Neurocomputing 275:1132-1139. doi:10.1016/j.neucom.2017.09.056 10.1016/j.neucom.2017.09.056
12
Kaggle Data Science Company 2017, https://www.kaggle.com/ Accessed 03 May 2023.
13
Kansara D., and V. Sawant 2020, Comparison of traditional machine learning and deep learning approaches for sentiment analysis. In Advanced Computing Technologies and Applications: Proceedings of 2nd International Conference on Advanced Computing Technologies and Applications-ICACTA Springer, Singapore, pp 365-377. 10.1007/978-981-15-3242-9_35
14
Keith L., W.H. Ko, and D.M. Sato 2006, Identification guide for diseases of tea (Camellia sinensis): Plant Disease PD-33. University of Hawaii, Honolulu, HI, USA.
15
Khan E., M.Z.U. Rehman, F. Ahmed, and M.A. Khan 2021, Classification of diseases in citrus fruits using SqueezeNet. In IEEE 2021 International Conference on Applied and Engineering Mathematics (ICAEM), pp 67-72. doi:10.1109/ICAEM53552.2021.9547133 10.1109/ICAEM53552.2021.9547133
16
Kimutai G., and A. Förster 2022, Tea sickness dataset. Mendeley Data V2. doi:10.17632/j32xdt2ff5.2 10.17632/j32xdt2ff5.2
17
Latha R.S., G.R. Sreekanth, R.C. Suganthe, R. Rajadevi, S. Karthikeyan, S. Kanivel, and B. Inbaraj 2021, Automatic detection of tea leaf diseases using deep convolution neural network. In 2021 International Conference on Computer Communication and Informatics (ICCCI), pp 1-6. doi:10.1109/ICCCI50826.2021.9402225 10.1109/ICCCI50826.2021.9402225
18
Mahesh B. 2020, Machine learning algorithms-a review. Int J Sci Res 9:381-386. doi:10.21275/ART20203995 10.21275/ART20203995
19
Mikołajczyk A., and M. Grochowski 2018, Data augmentation for improving deep learning in image classification problem. In 2018 International Interdisciplinary PhD Workshop (IIPhDW), pp 117-122. 10.1109/IIPHDW.2018.8388338
20
Mirza A.H. 2018, Computer network intrusion detection using various classifiers and ensemble learning. In 2018 26th Signal Processing and Communications Applications Conference (SIU), pp 1-4. doi:10.1109/SIU.2018.8404704 10.1109/SIU.2018.8404704
21
Mohapatra S., and T. Swarnkar 2021, Comparative study of different orange data mining tool-based AI techniques in image classification. In S Das, MN Mohanty, eds, Advances in Intelligent Computing and Communication: Lecture Notes in Networks and Systems, vol 202. Springer, Singapore, pp 611-620. doi:10.1007/978-981-16-0695-3_57 10.1007/978-981-16-0695-3_57
22
Nanehkaran Y.A., D. Zhang, J. Chen, Y. Tian, and N. Al-Nabhan 2020, Recognition of plant leaf diseases based on computer vision. J Ambient Intell Human Comput pp 1-18. doi:10.1007/s12652-020-02505-x 10.1007/s12652-020-02505-x
23
Neyshabur B., S. Bhojanapalli, D. McAllester, and N. Srebro 2017, Exploring generalization in deep learning. Adv Neural Inf Process Syst 30.
24
Nusrat I., and S.B. Jang 2018, A comparison of regularization techniques in deep neural networks. Symmetry 10(11):648. doi:10.3390/sym10110648 10.3390/sym10110648
25
Patro V.M., and M.R. Patra 2014, Augmenting weighted average with confusion matrix to enhance classification accuracy. Trans Mach Learn Artif Intell 2(4):77-91. doi:10.14738/tmlai.24.328 10.14738/tmlai.24.328
26
Ratra R., and P. Gulia 2020, Experimental evaluation of open source data mining tools (WEKA and Orange). Int J Eng Trends Technol 68(8):30-35. doi:10.14445/22315381/IJETT-V68I8P206S 10.14445/22315381/IJETT-V68I8P206S
27
Raut P., and A. Dani 2020, Correlation between number of hidden layers and accuracy of artificial neural network. In Advanced Computing Technologies and Applications: Proceedings of 2nd International Conference on Advanced Computing Technologies and Applications-ICACTA. Springer, Singapore, pp 513-521. doi:10.1007/978-981-15-3242-9_49 10.1007/978-981-15-3242-9_49
28
Shafi I., J. Ahmad, S.I. Shah, and F.M. Kashif 2006, Impact of varying neurons and hidden layers in neural network architecture for a time frequency application. In 2006 International Multitopic Conference, pp 188-193. doi:10.1109/INMIC.2006.358160 10.1109/INMIC.2006.358160
29
Sharma S, S. Sharma, and A. Athaiya 2020, Activation functions in neural networks. Int J Eng Appl Sci 4(12):310-316. doi:10.33564/IJEAST.2020.v04i12.054 10.33564/IJEAST.2020.v04i12.054
30
Shi Y., T. ValizadehAslani, J. Wang, P. Ren, Y. Zhang, M. Hu, and H. Liang 2022, Improving imbalanced learning by pre-finetuning with data augmentation. In Fourth International Workshop on Learning with Imbalanced Domains: Theory and Applications, pp 68-82.
31
Shrestha A., and A. Mahmood 2019, Review of deep learning algorithms and architectures. IEEE Access 7:53040-53065. doi:10.1109/ACCESS.2019.2912200 10.1109/ACCESS.2019.2912200
32
Shruthi U., V. Nagaveni, and B.K. Raghavendra 2019, A review on machine learning classification techniques for plant disease detection. In 5th International Conference on Advanced Computing and Communication Systems (ICACCS), pp 281-284. doi:10.1109/ICACCS.2019.8728415 10.1109/ICACCS.2019.8728415
33
Sibi P., S.A. Jones, and P. Siddarth 2013, Analysis of different activation functions using back propagation neural networks. J Theor Appl Inf Technol 47:1264-1268
34
Singh R., N. Sharma, and R. Gupta 2023, Proposed CNN model for tea leaf disease classification. In 2023 2nd International Conference on Applied Artificial Intelligence and Computing (ICAAIC), pp 53-60. doi:10.1109/ICAAIC56838.2023.10140680 10.1109/ICAAIC56838.2023.10140680
35
Singh V., N. Sharma, and S. Singh 2020, A review of imaging techniques for plant disease detection. Artif Intell 4:229-242. doi:10.1016/j.aiia.2020.10.002 10.1016/j.aiia.2020.10.002
36
Sladojevic S., M. Arsenovic, A. Anderla, D. Culibrk, and D. Stefanovic 2016, Deep neural networks-based recognition of plant diseases by leaf image classification. Comput Intell Neurosci 2016:1-11. doi:10.1155/2016/3289801 10.1155/2016/3289801
37
Szandała T. 2021, Review and comparison of commonly used activation functions for deep neural networks. In: A Bhoi, P Mallick, CM Liu, V Balas, eds, Bio-inspired Neurocomputing. Studies in Computational Intelligence, vol 903. Springer, Singapore, pp 203-224. doi:10.1007/978-981-15-5495-7_11 10.1007/978-981-15-5495-7_11
38
Tian Y., and Y. Zhang 2022, A comprehensive survey on regularization strategies in machine learning. Inf Fusion 80:146-166. doi:10.1016/j.inffus.2021.11.005 10.1016/j.inffus.2021.11.005
39
Tiwari R.G., A. Misra, and N. Ujjwal 2022, Image Embedding and Classification using Pre-Trained Deep Learning Architectures. In 2022 8th International Conference on Signal Processing and Communication (ICSC), pp 125-130. doi:10.1109/ICSC56524.2022.10009560 10.1109/ICSC56524.2022.10009560
40
Tripathi M. 2021, Analysis of convolutional neural network-based image classification techniques. J Innov Image Proc 3(2):100-117. doi:10.36548/jiip.2021.2.003 10.36548/jiip.2021.2.003
41
Uzair M., and N. Jamil 2020, Effects of hidden layers on the efficiency of neural networks. In 2020 23rd International Multitopic Conference (INMIC), pp 1-6. doi:10.1109/INMIC50486.2020.9318195 10.1109/INMIC50486.2020.9318195
42
Vaishnav D., and B.R. Rao 2018, Comparison of machine learning algorithms and fruit classification using orange data mining tool. In 2018 3rd International Conference on Inventive Computation Technologies (ICICT), pp 603-607. doi:10.1109/ICICT43934.2018.9034442 10.1109/ICICT43934.2018.9034442
43
Xia X., C. Xu, and B. Nan 2017, Inception-v3 for flower classification. In 2017 2nd International Conference on Image, Vision, and Computing (ICIVC), pp 783-787. doi:10.1109/ICIVC.2017.7984661 10.1109/ICIVC.2017.7984661
페이지 상단으로 이동하기