Legume Research

  • Chief EditorJ. S. Sandhu

  • Print ISSN 0250-5371

  • Online ISSN 0976-0571

  • NAAS Rating 6.80

  • SJR 0.391

  • Impact Factor 0.8 (2024)

Frequency :
Monthly (January, February, March, April, May, June, July, August, September, October, November and December)
Indexing Services :
BIOSIS Preview, ISI Citation Index, Biological Abstracts, Elsevier (Scopus and Embase), AGRICOLA, Google Scholar, CrossRef, CAB Abstracting Journals, Chemical Abstracts, Indian Science Abstracts, EBSCO Indexing Services, Index Copernicus
Legume Research, volume 47 issue 8 (august 2024) : 1404-1411

Efficient Faba Bean Leaf Disease Identification through Smart Detection using Deep Convolutional Neural Networks

Hie Yong Jeong1, In Seop Na2,*
1Department of Artificial Intelligence Convergence, Chonnam National University, Republic of Korea.
2Division of Culture Contents, Chonnam National University, Republic of Korea.
  • Submitted25-01-2024|

  • Accepted15-04-2024|

  • First Online 16-05-2024|

  • doi 10.18805/LRF-798

Cite article:- Jeong Yong Hie, Na Seop In (2024). Efficient Faba Bean Leaf Disease Identification through Smart Detection using Deep Convolutional Neural Networks . Legume Research. 47(8): 1404-1411. doi: 10.18805/LRF-798.

Background: Legumes, such as lentils, field peas, Faba beans and chickpeas, are high in vitamins, fiber, important minerals and protein and can help avoid obesity and cardiovascular illnesses. They also contribute to ecosystem services, such as nitrogen fixation and resilience to environmental stresses. Despite a 60% increase in global pulse production from 2000 to 2021, a demand-supply gap, especially in South Asia, raises concerns about nutritional access. Since illnesses are currently an issue to the food security of faba beans, machine learning is required for efficient disease identification.

Methods: This research employs Convolutional Neural Networks (CNNs) for robust Faba bean leaf disease identification. The CNN model is trained with diverse images representing specific diseases. The study focuses on diseases like Chocolate Spot, Faba Bean Gall, Rust and Healthy leaves. Image processing involves resizing, grayscale conversion and labeling. The CNN architecture includes eight convolutional layers, four max-pooling layers and three dropout layers. The model is trained using 80% of the dataset, validated using 20% and tested for accuracy.

Result: The CNN model achieves an accuracy of 99.37% during training and 89.69% during validation after 75 epochs. Confusion matrix and classification report illustrate the model’s performance. It shows high precision, recall and F1 scores for each class, indicating balanced performance. Chocolate Spot and Rust exhibit the highest precision and F1 scores. The overall accuracy is 91%, comparable to other studies on Faba bean disease detection. The study presents a CNN-based disease identification system for Faba beans, demonstrating high accuracy and balanced performance across different diseases. The model’s effectiveness is comparable to other advanced techniques. The research highlights the potential of machine learning in optimizing disease management for Faba beans. Future work could explore a broader range of diseases and incorporate hybrid machine learning algorithms for further improvement.

Legumes such as chickpeas, Feba beans, lentils and field peas (often referred to as “poor man’s meat” due to their richness in protein, essential elements, dietary fiber and vitamins) play a crucial role in preventing health issues like cardiovascular diseases and obesity when included in daily diets (Crujeiras et al., 2010; Bouchenak and Lamri-Senhadji, 2013). Also, they have significant contributions in providing ecosystem services, such as nitrogen fixation, deep root systems and resilience to environmental stresses involving carbon sequestration and mitigation of greenhouse gas emissions, reducing groundwater pollution and enhancing crop resilience in the face of climate change (Hazra et al., 2020; Jena et al., 2022). Faba beans are vital contributors to global food security and provide essential nutrients to human and livestock diets (Kwon et al., 2018).

Though the global pulse production saw a surge between 2000 and 2021 with a 60% increase, surpassing the world population growth (30%), the persistent demand-supply gap, especially in South Asia, is a concerning issue (CGAIR, 2023). This scarcity has led to rising local prices, indicating potential challenges in meeting the nutritional demands of the growing population. Additionally, the increased reliance on pulse imports, coupled with the susceptibility of legumes to diseases in rainfed conditions, raises concerns about the stability of the global supply chain. Balancing the surge in production with effective disease management strategies becomes crucial for ensuring sustained access to this vital source of nutrition.

This research aims to develop a robust and reliable Feba bean leaf disease identification system using a Convolutional Neural Network. Training the model with diverse images will enable it to detect patterns indicative of specific diseases. This tool will assist farmers in making accurate and data-driven decisions to mitigate the impact of diseases on Faba bean crops. This study anticipates enhancing agricultural productivity, optimizing resource utilization and securing global food supplies. 
 
Related work
 
Legume plants are frequently affected by diseases caused by different microorganisms like fungi, bacteria, viruses and parasites. These diseases limit the full potential of these plants, leading to significant losses in yield. Sometimes, the losses can be as high as 100% in the absence of effective crop management practices (Darai et al., 2017; Jones, 2020; Urva, 2021).

Subsequently, researchers are turning towards technological solutions for the early and accurate identification of crop diseases. Various automated methods, categorized as direct and indirect, have been proposed for disease detection. Among the indirect methods, are optical imaging techniques, such as fluorescence and hyperspectral imaging which provide valuable insights into the physiological and morphological changes in legume plants caused by diseases. For example, parameters like morphological changes and transpiration rates are measured using these techniques to identify diseases and predict crop health (Suárez et al.,  2022). Over time, the analysis of the large datasets generated by these imaging technologies has seen an increase in the use of machine learning algorithms, particularly from the year 2010 (Chen et al., 2020). Deep learning techniques such as Convolutional Neural Networks (CNNs) have emerged as powerful tools in the field of image-based disease detection in plants (Min et al., 2024). CNNs are a class of deep learning models designed for pattern recognition in visual data, making them well-suited for tasks like plant disease identification. CNNs employ convolutional layers to automatically extract relevant features from images. This feature extraction capability is crucial for discerning subtle visual cues indicative of diseases, such as discoloration, lesions, or abnormal growth patterns. A model based on deep learning is suggested for identifying plant species and classifying plant leaf diseases. Convolutional Neural Networks (CNNs) are neural networks that are highly layered and usually contain basic function layers such as convolution layers, pooling layers and a classification layer. The first practical use of a modern Convolutional Neural Network was proposed as LeNet-5. CNNs vary in terms of how these basic layers are structured and packaged, as well as the approach used to train the network. Aside from the legume crop, artificial intelligence has been used in many other fields, such as big data analysis and animal research (Na et al., 2024; Kim and Kim, 2023; Porwal et al., 2024; Wasik and Pattinson, 2024).

In addition to CNN, various models often integrate multiple approaches where the strengths of different methods are combined to achieve improved accuracy and reliability (Cho, 2024). These systems are known as hybrid systems. Hybrid systems may incorporate traditional image processing techniques, such as segmentation and morphological operations, alongside deep learning methods like CNNs and can fuse information from various sources, including spectral, spatial and temporal data, to enhance the overall disease detection process. Bedi and Gole (2021) came up with a way to automatically detect Bacterial Spot disease in peach plants using their leaf images. Their method involved a combination of a convolutional autoencoder (CAE) and CNN, resulting in a hybrid system. With just 9,914 training parameters, their model achieved an impressive 99.35% training accuracy and 98.38% testing accuracy. The main advantage of this approach was that it required fewer training parameters, which ultimately reduced the analysis time. Abed et al., (2021) proposed a framework that detects bean leaves and diagnoses diseases within the detected leaves using a pre-trained ResNet34 encoder. Five different deep-learning models were evaluated to identify the healthiness of bean leaves. The effectiveness of the suggested framework was assessed utilizing a dataset comprising 1295 images categorized into three classes. The Densenet121 model demonstrated superior performance, achieving a CAR of 98.31%, Sensitivity of 99.03%, Specificity of 96.82%, Precision of 98.45%, F1-Score of 98.74% and an AUC of 100%. These findings collectively emphasize the significant impact of cutting-edge technologies and machine learning in enhancing disease management processes. For instance, Manschadi et al., (1998) utilized a model that was created for beans cultivated in cooler climates with ample water and nutrients. Through subsequent updates incorporating region-specific features and validation using field data from northern Syria, where beans were subjected to varying water levels, the refined model demonstrated accurate predictions of bean growth and yield in drought-prone areas. This highlighted the importance of accounting for regional nuances in optimizing predictive models for plant diseases. The present study extends the application of deep CNNs to address disease detection challenges, emphasizing a tailored approach for Faba beans.   
The method used for building a deep convolutional neural network (CNN) model for Faba bean leaf disease identification is described in the following sections. It is divided into several significant segments, the first of which is the gathering of images for deep neural network classification.
 
Dataset
 
The CNN model requires a sizable dataset to be processed. In this work, the data is captured with the help of agricultural experts. The Dataset contains 8021 images of faba bean leaf images captured in the fields, the detail is summarized in Table 1. It is separated into four categories: three disease categories and one healthy category. The disease groups include Healthy, Rust, Faba Bean Gall and Chocolate Spot leaf images (Fig 1 and Table 1). In this study, the entire dataset was split into training, testing and validation data in an 80:20 ratio.

Fig 1: Images from faba beans leaf dataset (a) Healthy (b) Rust (c) Gall and (d) Chocolate spot images.



Table 1: Classification of diseases in Faba beans.


 
Image processing and labelling
 
Image pre-processing was used to enhance or modify the raw images that the CNN classifier needed to process before training the model. Images acquired from various sources might have various dimensions, so it would be necessary to resize and rescale the pictures to make sure that the dimensions of the images are the same. Considering the computational cost of handling larger-sized images, this procedure is essential for consistency as well as for speeding up the training process. Scaling the data to a standardized size and format is essential before putting it into the network. It is standard procedure to use 224 × 224 input images in reliable models. A meaningful comparison with state-of-the-art models is facilitated by aligning the image size of our network with these standard dimensions.

So, the images were first resized to 224 × 224 pixels to normalize its size. The pictures were then converted to grayscale. For the explicit learning of the training data features, a significant amount of training data is needed at this pre-processing stage. The next stage involved sorting the photos of Faba bean leaves according to type and labeling each photo with the appropriate disease acronym. In this instance, the test collection and training dataset displayed four classes (Table 1).
 
Training dataset
 
This step involved applying the Convolutional Neural Network (CNN) process to generate a model for performance evaluation using image data as input. The steps involved in normalizing images of Faba bean leaves are depicted in Fig 2.

Fig 2: Classification of CNN model.


 
Convolutional neural network (CNN) model
 
In the CNN model, images can be efficiently filtered by the convolution operation because of its matrix structure. A convolutional layer, input layer, fully connected layer, pooling layer, drop-out layer and a final linked dataset classification layer are among the layers used in the Convolutional Neural Network for data training. An order of operations is mapped onto the input test set by each layer. Table 2 provides the details of the critical architectural and training parameters of a Convolutional Neural Network (CNN) employed in this study.

Table 2: Hyper-parameter of CNN model.



The CNN incorporated eight convolutional layers, each employing 32 and 64 filters to extract intricate features from input data. The architecture is further enriched with four max-pooling layers to down-sample spatial dimensions effectively. Dropout, a regularization technique, is strategically applied with rates of 0.25 and 0.4 in specific layers, enhancing the model’s generalization ability. Uniform weight assignment, Rectified Linear Unit (ReLU) activation function and a low learning rate of 0.00001 contribute to the network’s stability. The training regimen spans 75 epochs, each involving a batch size of 32 instances. Fig 3 visually presents the comprehensive architecture.

Fig 3: CNN model Architecture with convolution and pooling layer specifications.


 
Feature extraction process
 
The present work employs a Convolutional Neural Network (CNN) to classify Faba bean leaf diseases. The study is conducted within the Jupyter Notebook environment of the Anaconda platform. There are four different classes in the categorization test. The Sequential API from the Keras package is utilized in the construction of the neural architecture. The input shape parameter is set to (224, 224, 3), reflecting the dimensions of the RGB input images, after a Conv2D layer is initialized with 32 filters of 3×3 dimensions. A crucial component of feature extraction is the introduction of non-linearity to the model through the subsequent use of the Rectified Linear Unit (ReLU) activation function. After the convolutional layer, a MaxPooling2D layer with a 2´2 pool size is deliberately included to improve feature abstraction and enable spatial dimension downsampling. This pattern of architecture is repeated twice, this time with extra Conv2D layers that have 64 and 128 filters, respectively, along with max-pooling and ReLU activation. The ability to extract spatial information hidden in the input images and capture complex structure are made possible by these layers. The addition of a flattening layer, which converts the three-dimensional feature maps into a single one-dimensional vector, marks an important change. Then, a dense layer with 128 neurons that uses ReLU activation acts as a feature aggregator, adding a substantial amount of complexity to the model. A Dropout layer with a dropout rate of 0.5 is purposefully inserted into the architecture to fix the overfitting problem. This layer effectively improves the model’s capacity for generalization by methodically deactivating 50% of neurons during the training phase.
 
Convolution layer
 
Convolution is a type of specialized linear operation that involves multiplying a filter and an input matrix element by element, followed by a summation at each location in a feature map. A 3´3 kernel is used for the entire input matrix. The output value in the corresponding position of the output tensor, also known as a feature map, is obtained by summing the element-wise products of each kernel element and the corresponding input tensor element at each location. Over a two-dimensional input image (I) and two-dimensional kernel (K), the convolution operation is defined as follows:
 
 
Where
m and n = Kernel (K) coordinates.
i and j = Image (I) coordinates.

After every convolution operation with Woutput × Houtput × Doutput,  the output size of the convolution layer is calculated as follows:
 
 
 Where
K = Number of filters applied to the image matrix.
F = Filter size.
P = Number of zero paddings.
S = Number of stride sizes.
 
Pooling layer
 
Max-pooling is the most popular type of pooling technique. It reduces the number of parameters in the network and eases computational demands by surveying larger image areas, which lowers the resolution of an output from a convolutional layer. The idea behind max-pooling is that the network recognizes unique elements like edges, curves and circles for a given image. The hypothesis states that more activated pixels have higher values. As a result, max-pooling chooses the most active pixels, forward-propagating these high values while eliminating less active pixels with lower values.
 
Performance evaluation parameters
 
a) Accuracy
 
The ratio of the total number of accurate predictions including both true positives and true negatives to the total number of predictions is one of the most widely used performance evaluation metrics. It usually indicates whether a model is being trained correctly and provides an estimate of its general performance. The following is the accuracy calculation formula:
 
 
TP = True positive.
TN = True negative.
FP = False positive.
FN = False negative.
 
b) Precision
 
The frequency of accurate model predictions is indicated by this parameter. The calculation involves dividing the total number of positive labels by the number of accurate positive predictions.
 
 
c) Recall or sensitivity
 
It calculates the classifier’s completeness. It is the proportion between the total number of positive reviews in the dataset and the number of correctly predicted positive observations. Discovering the most positive labels is the aim of computing these metrics. The following is the recall calculation formula:
 
 
 
d)    F1 Score

An F1-score of 1 indicates perfect performance, while a score of 0 indicates complete failure. The following formula is used to get the F-measure:
 
 
 
Steps used in algorithm
 
• Collect the color images of the leaf of faba beans.
• Utilize a convolutional neural network (CNN)-based segmentation to generate a mask from the given color image.
• Overlay the original color image with the generated mask to produce a new masked image.
• Divide the masked image into smaller regions known as tiles (Ktiles).
• Classify each tile (Ktiles) from the masked image into the category of “Faba bean.”
• Identify and analyze the classified tiles (Ktiles) to pinpoint regions indicative of a diseased part of the leaf.
• Conclude the process, having completed the steps for disease detection.
The Convolutional Neural Network (CNN) framework used in this study consisted of eight convolutional layers with 32 and 64 filters, four Max-pooling layers, three dropout layers with 0.25 and 0.4 settings and two fully-connected layers with 1024 hidden nodes and four output nodes, as shown graphically in Fig 4. A total of 1,11,69,476 training parameters are included in the overall architecture.
 
The training and validation loss comparison is shown in Fig 4 (a) and the training accuracy and validation accuracy comparison is shown in Fig 4 (b). It is evident that, after 75 epochs, an accuracy rate of 99.37% has been achieved during the training phase. While the accuracy rate for validation phase is 89.69%. Thus, it is sensible to conclude that, under the research methodology, more iterations would result in higher data accuracy.  It is worth noting that the number of epochs rises proportionally with the lengthening of the training phase.

Fig 4: (a) Training loss vs validation loss (b) Training accuracy vs validation accuracy.


 
Confusion matrix performance
 
For a particular set of test data, the confusion matrix is a matrix that is used to assess the extent to which the classification models perform. Fig 5 displays the confusion matrix with faba bean test data for the following categories: Chocolate spot, Faba bean Gall, Healthy and Rust. It shows the performance of the presented algorithm on the testing dataset as a visual representation. The confusion matrix shows the values of true positives, true negatives, false positives and false negatives for each class.
 

Fig 5: Confusion matrix with faba bean test data for categories: Chocolate spot, Faba bean Gall, Healthy and Rust.



Classification report
 
Table 3 shows the classification performance the present model with metrics such as precision, recall and F1-score. It was observed that the model performed well on the classification task.

Table 3: Classification accuracies, precision, F1 score and support.



The model is well-balanced in terms of precision and recall with few false positives or false negatives. Hence, the model is able to accurately identify the correct class for most of the images. The accuracies in present study were comparable with other studies on disease detection in Faba beans. Slimani et al., (2023) obtained precision of 95.1% by using YOLOv8 for classification of rust disease in fava bean fields of eastern Morocco. Other performance indicators were also high and almost equivalent to the present study with mean Average Precision (mAP), recall and F1 score of 93.7%, 90.3% and 92%, respectively. In another study, Salau et al., (2023) achieved accuracies of 92.1% and 98.14% with raw and preprocessed data, respectively.

In the present study, the average precision, recall and F1-score were all high (around 0.91), indicating that the model could accurately identify the correct class for most of the images. However, the model performed slightly better in some classes than others. Chocolate Spot and Rust had the highest precision and F1 scores, while Faba Bean Gall had a slightly lower recall. This could be due to several factors, such as the number of images in each class or the inherent difficulty of distinguishing between certain classes.
This study addresses a recent area of research in digital image processing by using a convolutional neural network (CNN) technique to identify diseases of the Faba bean leaf. The basic processes of image acquisition, preprocessing, representation, interpretation and recognition are followed by all digital image processing methods, including CNN. This entails channel ordering, pixel value normalization, resizing the dimensions and image data augmentation. Following preprocessing, a Canny edge detector was used to segment those images and determine the region of interest. Using the segmented images, the model is trained to produce representative features. the developed classifier’s robustness using F1-score, accuracy, precision and recall is assessed. For this study, four disease classes-Chocolate Spot, Faba Bean Gall, Rust and Healthy are taken into account.

High recall, F1-scores and precision for each category-Chocolate Spot, Faba Bean Gall, Healthy and Rust-indicate that the classification model performs well across a number of classes. The precision and recall metrics show how well the model predicts each class, while the former emphasizes how well it captures instances of each category. The model’s ability to produce accurate predictions across the whole dataset is indicated by its overall accuracy of  91%. With unweighted and weighted evaluations, respectively, the macro and weighted averages provide additional evidence for the model’s resilience. All of these metrics confirm that the model is dependable in correctly categorizing examples within the specified classes, offering insightful information about its overall effectiveness and emphasizing its potential  for real-world use. 
This study was limited to four distinct diseases that affect Faba bean leaves. But it’s important to recognize that faba beans are vulnerable to several diseases that affect not just the leaves but also the stems and roots. It is therefore recommended that future research examine a wider range of illnesses that could impact various faba bean parts. Furthermore, the concatenated model’s fully connected layer classifier was employed in this study. To compare and improve the overall performance of the model, hybrid machine learning algorithms, such as support vector machines, random forests and decision trees can be used in the future.
This research was funded by the Korea Institute of Marine Science and Technology Promotion (KIMST) funded by the Ministry of Oceans and Fisheries, Korea (RS-2022-KS221676, Development of Digital Flow-through Aquaculture System).
All authors contributed toward data analysis, drafting and revising the paper and agreed to be responsible for all the aspects of this work.
The database generated and /or analysed during the current study are not publicly available due to privacy, but are available from the corresponding author on reasonable request.
Author(s) declare that all works are original and this manuscript has not been published in any other journal.
The authors declare that they have no conflicts of interest to report regarding the present study.

  1. Abed, S.H., Al-Waisy, A.S., Mohammed, H.J. and Al-Fahdawi, S. (2021). A modern deep learning framework in robot vision for automated bean leaves diseases detection. International Journal of Intelligent Robotics and Applications. 5(2): 235- 251.

  2. Bedi, P. and Gole, P. (2021). Plant disease detection using hybrid model based on convolutional autoencoder and convolutional neural network. Artificial Intelligence in Agriculture. 5: 90- 101.

  3. Bouchenak, M. and Lamri-Senhadji, M. (2013). Nutritional quality of legumes and their role in cardiometabolic risk prevention: A review. Journal of Medicinal Food. 16(3): 185-198.

  4. CGIAR Initiative on Foresight. (2023). What do we know about the future of pulses in global and regional agri-food systems? https://www.cgiar.org/news-events/news/what-do-we-know- about-the-future-of-pulses-in-global-and-regional-agri- food-systems/ 

  5. Chen, X., Xie, H., Zou, D. and Hwang, G.J. (2020). Application and theory gaps during the rise of artificial intelligence in education. Computers and Education: Artificial Intelligence. 1: 100002.

  6. Cho, O.H. (2024). An evaluation of various machine learning approaches for detecting leaf diseases in agriculture. Legume Research. 47(4): 619-627. https://doi.org/10.18805/LRF-787.

  7. Crujeiras, A.B., Goyenechea, E. and Martínez, J.A. (2010). Fruit, Vegetables and Legumes Consumption: Role in Preventing and Treating Obesity. In: Bioactive Foods in Promoting Health. Academic Press. (pp. 359-380).

  8. Darai, R., Ojha, B.R. and Dhakal, K.H. (2017). Disease management of major grain legumes and breeding strategies in Nepal. Adv. Plants Agric. Res. 6(1): 1-7. 

  9. Hazra, K.K., Nath, C.P., Ghosh, P.K. and Swain, D.K. (2020). Inclusion of legumes in rice-wheat cropping system for enhancing carbon sequestration. Carbon Management in Tropical and Sub-Tropical Terrestrial Systems. pp 23-36.

  10. Jena, J., Maitra, S., Hossain, A., Pramanick, B., Gitari, H.I., Praharaj, S., et al. (2022). Role of legumes in cropping system for soil ecosystem improvement. Ecosystem Services: Types, Management and Benefits. Nova Science Publishers, Inc, 415.

  11. Jones, R.A. (2020). Disease pandemics and major epidemics arising from new encounters between indigenous viruses and introduced crops. Viruses. 12(12): 1388.

  12. Kim, E.J., Kim, J.Y. (2023). Exploring the online news trends of the metaverse in South Korea: A data-mining-driven semantic network analysis. Sustainability. 15: 16279. https:// doi.org/10.3390/su152316279. 

  13. Kwon, S.J., Kim, D.G., Kim, J.M., Kang, K.Y., Lee, M.K., Hong, M.J., Ryu, J. (2018). Phytochemical compounds and antioxidant activity in the grain of selected faba bean (Vicia faba) genotypes. Plant Breeding and Biotechnology. 6(1): 65-73.

  14. Manschadi, A.M., Sauerborn, J., Stützel, H., Göbel, W. and Saxena, M.C. (1998). Simulation of faba bean (Vicia faba L.) growth and development under Mediterranean conditions: Model adaptation and evaluation. European Journal of Agronomy. 9(4): 273-293.

  15. Min, P.K., Mito, K. and Kim, T.H. (2024). The evolving landscape of artificial intelligence applications in animal health. Indian Journal of Animal Research. https://doi.org/10.18805/ IJAR.BF-1742. 

  16. Na, J.C., Kim, E.J., Kim, J.Y. (2024). Unveiling metaverse social trends: Analysing big data regarding online sports news with LDA-based topic modelling. Revista De Psicología Del Deporte (Journal of Sport Psychology). 33(1): 115- 125. https://rpdonline.com/index.php/rpd/article/view/1533.

  17. Porwal, S., Majid, M., Desai, S. C., Vaishnav, J. and Alam, S. (2024). Recent advances, challenges in applying artificial intelligence and deep learning in the manufacturing industry. Pacific Business Review (International). 16(7): 143-152.

  18. Salau, A.O., Abeje, B.T., Faisal, A.N. and Asfaw, T.T. (2023). Faba Bean Disease Detection Using Deep Learning Techniques. In 2023 International Conference on Cyber Management and Engineering (CyMaEn), IEEE. (pp. 344-349).

  19. Slimani, H., El Mhamdi, J. and Jilbab, A. (2023). Artificial intelligence- based detection of fava bean rust disease in agricultural settings: An innovative approach. International Journal of Advanced Computer Science and Applications. 14(6). DOI: 10.14569/IJACSA.2023.0140614.

  20. Suárez, J.C., Vanegas, J.I., Contreras, A.T., Anzola, J.A., Urban, M.O., Beebe, S.E. and Rao, I.M. (2022). Chlorophyll fluorescence imaging as a tool for evaluating disease resistance of common bean lines in the Western Amazon Region of Colombia. Plants. 11(10): 1371.

  21. Urva, P. (2021). Detection and classification of grain crops and legumes disease: A survey. Sparklinglight Transactions on Artificial Intelligence and Quantum Computing (STAIQC). 1(1): 41-55.

  22. Wasik, S. and Pattinson, R.  (2024). Artificial intelligence applications in fish classification and taxonomy: Advancing our understanding of aquatic biodiversity. FishTaxa. 31: 11-21. 

Editorial Board

View all (0)