iconOpen Access

REVIEW

crossmark

Evaluation of State-of-the-Art Deep Learning Techniques for Plant Disease and Pest Detection

MD Tausif Mallick1, Saptarshi Banerjee2, Nityananda Thakur3, Himadri Nath Saha4,*, Amlan Chakrabarti1

1 Department of AKCSIT, University of Calcutta, Kolkata, 700106, India
2 Department of Computer Science, Illinios Institute of Technology, Chicago, IL 60616, USA
3 Department of Mathematics, SNEC, University of Calcutta, Kolkata, 700009, India
4 Department of Computer Science, SNEC, University of Calcutta, Kolkata, 700009, India

* Corresponding Author: Himadri Nath Saha. Email: email

Computers, Materials & Continua 2025, 85(1), 121-180. https://doi.org/10.32604/cmc.2025.065250

Abstract

Addressing plant diseases and pests is not just crucial; it’s a matter of utmost importance for enhancing crop production and preventing economic losses. Recent advancements in artificial intelligence, machine learning, and deep learning have revolutionised the precision and efficiency of this process, surpassing the limitations of manual identification. This study comprehensively reviews modern computer-based techniques, including recent advances in artificial intelligence, for detecting diseases and pests through images. This paper uniquely categorises methodologies into hyperspectral imaging, non-visualisation techniques, visualisation approaches, modified deep learning architectures, and transformer models, helping researchers gain detailed, insightful understandings. The exhaustive survey of recent works and comparative studies in this domain guides researchers in selecting appropriate and advanced state-of-the-art methods for plant disease and pest detection. Additionally, this paper highlights the consistent superiority of modern AI-based approaches, which often outperform older image analysis methods in terms of speed and accuracy. Further, this survey focuses on the efficiency of vision transformers against well-known deep learning architectures like MobileNetV3, which shows that Hierarchical Vision Transformer (HvT) can achieve accuracy upwards of 99.3% in plant disease detection. The study concludes by addressing the challenges of designing the systems, proposing potential solutions, and outlining directions for future research in this field.

Keywords

Image processing; machine learning; deep learning; vision transformer; pest and disease detection

1  Introduction

In 2050, it’s predicted that there will be 9.1 billion people on the planet [1] and, therefore, to supply the food needs of the rapidly expanding population, the agricultural output must increase by around 70%. Food security, rural livelihoods, and agricultural trade are all negatively impacted by plant pests and diseases, which cause up to 40% of worldwide crop production to be lost each year, according to the Food and Agriculture Organization (FAO) [2]. To meet the growing need for food, agricultural productivity must rise by almost 70%, as the world’s population is expected to reach 9.1 billion by 2050 [1]. Early disease and pest identification and control are agronomic requirements and vital facilitators of food system resilience. Global food security goals are directly supported by emerging technologies like artificial intelligence and deep learning, which provide scalable and precise solutions that enable prompt intervention, less chemical use, and sustainable agriculture methods [26].

Fungal or fungal-like organisms primarily cause most plant diseases (around 85%). The bulk of plant diseases are generally linked to bacterial or viral pathogens, whereas only a small number are caused by certain worms [7]. According to recent research, pathogen-induced stress significantly alters chloroplast function and reduces photosynthetic efficiency, resulting in slower plant development and lower potential yield [8]. Among the most damaging fungal threats to the world’s food security are wheat rust diseases, which include stem rust (“Puccinia graminis”), stripe rust (“P. striiformis”), and leaf rust (“P. triticina”). Stripe rust alone can result in total crop failure under vulnerable circumstances, and several diseases can cause catastrophic crop losses, with yearly economic costs estimated at US $5 billion worldwide [9]. Pests also cause considerable losses in agricultural production [4] as insects often feed on leaves, disrupting photosynthesis. Furthermore, pests serve as vectors for various diseases, contributing to significant crop losses worldwide [3]. For example, the bug “L.hesperus” is a pest that severely damages cotton crops, and therefore, it is a common target of pesticide applications [10,11].

“Smart farming” refers to the management of farms that make use of contemporary technology for communication and information to maximize production and quality while minimizing the need for manpower [12]; it is crucial for raising agricultural production’s yield, ecological impact, food safety, and resilience [13]. Smart farming enables the continuous monitoring of plants to detect diseases and pests at an early stage, allowing for timely interventions. When an infection is detected, smart farming can help farmers classify different types of diseases and pests based on various aspects, such as other plant and crop species, the age of plants, and external conditions [1416]. Based on the identified threats, smart farming can be used to recommend appropriate countermeasures, such as pesticides, fungicides, or herbicides, to mitigate these risks. As a result, farmers can act very early and stop the spread of diseases and pests before they can infect large areas of the fields. Additionally, the use of toxic chemicals can be significantly reduced. For these reasons, fast, automated, accurate, and cost-effective approaches to identifying crop diseases and pests are of significant practical importance [17,18].

The extraction and analysis of plant image data have enabled several novel approaches, such as computer vision, machine learning (ML), and deep learning (DL), to successfully identify plant diseases and pests. These techniques are also accurate and fast. Rehman et al. secured an accuracy of 99.6% and above on apple, grape, peach and cherry diseases using Darknet53 while proposing an improved swarm optimisation butterfly algorithm [19]. Khan et al. [20] achieved an accuracy 97.2% with their proposed feature extraction method in detecting apple diseases. Hassan et al. have proposed a novel deep-learning model based on the inception layer, and a residual connection has been trained and tested on three different plant disease datasets, obtaining 99.39% accuracy on the PlantVillage dataset overall, and 99.66% on the rice disease dataset and the cassava dataset 76.59% [21]. The symptoms of pest attack or disease infection on plants differ in shape, texture, colour, and size based on the nature of pathogens, viruses, fungi, etc. The qualities of the image surface features, such as texture, colours, and structure, are assessed using deep learning [22], A branch of artificial intelligence called machine learning is being used more and more to mimic human cognitive processes like perception, categorization, and judgment in intricate, data-rich settings [23]. To be more precise, convolutional neural networks (CNN) are frequently used in various object identification and image classification applications [2427]. These algorithms can be trained using various contaminated plant images gathered from public online sources, agricultural farms, and laboratories. Several image processing-based segmentation, feature extraction and clustering methods are applied to an input image for successful and accurate disease and pest detection. Some of the CNN models used in this domain include AlexNet [28], InceptionV3 [29], Inception V4 [30], VGGnet [31], Microsoft ResNet [32], and DenseNet [33].

Our review paper uniquely categorises and evaluates methodologies for identifying plant pests and diseases using deep learning models. Unlike existing literature, which often consolidates various methods into a single section, this structured categorization allows researchers to efficiently identify and compare the most suitable deep learning techniques for their specific pest and disease detection objectives. Additionally, we examine modified deep learning and transformer models, incorporating the latest research to reflect current advancements in the field. We have used vision transformers against well-established models on the tomato leaf dataset from PlantVillage [34]; the results demonstrated that vision transformers perform better than the convolutional neural network (CNN) architectures. This comprehensive analysis provides valuable insights, guiding researchers in selecting the most suitable methods for their needs and advancing the field. A general flow diagram of implementing deep-learning models for visualizing plant pests is described in Fig. 1, which states that the process begins with dataset acquisition, followed by its division into two halves. Consequently, the training/validation graphs highlight the models’ significance, emphasizing the training methods of deep learning architectures, whether initiated from scratch or utilizing transfer learning. The following stage involves using assessment metrics specifically tailored for image classification tasks, particularly for identifying pest and plant disease varieties. The procedure involves applying mapping and visualization strategies that facilitate image classification, localization, and identification. This comprehensive visual depiction clearly illustrates the deep learning model’s sequential workflow, providing readers with a clear understanding of its structure, implementation procedures, and general operation.

images

Figure 1: A comprehensive flow diagram illustrating the developmental stages of a typical deep learning model

This survey work offers several contributions to plant pest and disease detection:

•   Different types of traditional and modified deep learning architectures for identifying and classifying pests and diseases of plants are described.

•   CNN models for plant pest and disease detection have been compared. Researchers must select the appropriate deep-learning model to construct a realistic scenario for plant pest and disease identification in the real world.

•   Visualization methodologies show the localization of infected spots on the plants and provide the farmers with necessary prevention suggestions about the pest or disease. Moreover, this agricultural information is presented without the involvement of agricultural professionals.

•   An overview of hyperspectral imaging with deep learning methodologies is provided, along with its applications in both laboratory and field settings for classifying and identifying the early stages of pest attacks and disease lesions in plants.

•   The paper consolidates the advantages and limitations of the techniques above, providing users with a comprehensive understanding to inform their selection of the most appropriate methodology for their specific applications.

•   The architecture and suitability of the Vision Transformer (ViT) models for visual identification tasks are highlighted in this manuscript [35]. The performance of transformer-based architectures Swin Transformer [36], HvT [37], PMVT [38], Tiny-LeViT [39], and ViTAE [40] on the Tomato Leaf dataset from the PlantVillage dataset [34]. The results demonstrated that the vision transformers performed on par with and surpassed machine learning and CNN models in test and validation accuracy.

•   Analyzes machine learning, deep learning, and vision transformers quantitatively in terms of precision, computing efficiency, and visualization capabilities.

•   The scope of future research on pest and disease identification in agriculture, including potential improvements, general limitations and challenges, has been aggregated.

2  Structural Design of the Survey Work

The bibliography-based analysis of plant pest and disease detection and classification involves two steps in the study: (a) accumulation of existing related research works and (b) a comprehensive evaluation of various research works. The journal articles and conference papers were collected from the scientific knowledge bases of IEEE Explore [41] and ScienceDirect [42]. As search keywords, the following queries are applied: “agriculture”, “deep learning”, “plant pest and disease detection,” and “convolutional neural network.”

We have sorted out papers that describe traditional and Modified deep learning models applied to the pest and disease domain of the agricultural field. We initially collected over a hundred papers covering all the methodologies relevant to this work. In the next phase, the selected papers were evaluated individually, focusing on the different characteristics, such as accuracy and time, as well as the working principles of these technologies.

This is a structural breakdown of this paper:

•   Section 3 outlines the typical deep learning model visualization and non-visualization techniques, other modified architectures, and state-of-the-art transformer models for identifying plant pests and diseases.

•   Section 4 explains hyperspectral imaging with deep learning models.

•   Section 5 offers the research challenges and possible solutions in this domain.

•   Section 7 provides a nuanced comparison between traditional machine learning, CNNs and contemporary vision transformers.

•   This paper has been concluded in Section 8 and provided future recommendations to improve the visualization and identification of diseases and pests of plants.

In Fig. 2, the organization of the paper is presented.

images

Figure 2: The organisation of the paper

3  Detection of Plant Disease and Pests Employing Popular Deep Learning Models

The concept of deep learning was created by Hinton et al. in 2006 [43]. Table 1 presents the differences between conventional image processing and machine learning approaches with deep learning methodologies. Many traditional deep learning architectures have been developed since the introduction of AlexNet [28]. For instance, Patel et al. demonstrated that state-of-the-art EfficientNet architectures achieve 98.3% disease classification accuracy while requiring 8.4 times fewer parameters than conventional CNNs [44].

images

Many traditional deep learning architectures have been developed since the introduction of AlexNet [28]. For instance, Patel et al. demonstrated that state-of-the-art EfficientNet architectures achieve 98.3% disease classification accuracy while requiring 8.4 times fewer parameters than conventional CNNs [44]. Deep learning techniques are employed to identify, segment, and categorise images. This section covers the well-known deep learning model-based research efforts used for plant disease and pest identification and classification in the agricultural domain. Notably, Lee and Wang recently proved that few transformer-based models can outperform CNNs in cross-species pest identification using attention mechanisms [53]. Additionally, various performance metrics are applied to these works to evaluate the designated deep learning models, as illustrated in Section 3.1.

3.1 Implementation of Deep Learning Models

The decision-making processes employed by deep learning architectures must be fully understood. Based on the properties of plant images, we synthesize two study approaches to extract insights. First, we have discussed the deep learning architectures based on (with or without) visualisation techniques. Next, the modified deep-learning architectures for plant pests and disease identification have been presented. In Table 2, the categories of deep learning models frequently employed to recognise and categorise plant diseases and pests are briefly explained.

images

3.1.1 Without Visualization Technique

The classification of diseases affecting the maize plants was performed using a CNN, and the model’s usefulness was determined using histogram approaches [54].

The extraction of pertinent visual features made possible by this technique significantly improved the categorization accuracy. Additionally, the model showed consistent performance across a range of imaging circumstances, indicating that it is robust for field-level applications. Tomato plants are highly vulnerable to a variety of diseases that can negatively affect their quality and yield. To detect various tomato plant diseases, the standard CNN architectures “AlexNet,” “ResNet,” and “GoogleNet” were introduced in [66]. According to the experimental results, the “ResNet” performed better than the other convolutional neural network models applied. The “LeNet” Architecture identified the banana leaf diseases, which significantly supported an accurate detection of RGB images with 92.88% detection accuracy and 85.94% for Grey Scale images of banana diseases in [67]. Different types of deep learning models, including “VGG”, “AlexNet”, “AlexNetOWTbn”, “GoogLeNet”, and “Overfeat”, were applied for the recognition of plant diseases and lesions [68]. Eight different plant diseases are identified using a variety of machine learning classifiers, including the following: “SVM”, “KNN”, and “Extreme Learning Machine (ELM),” as well as deep learning architectures such as “GoogleNet,” “ResNet-50,” “InceptionResNetv2,” and “SqueezeNet” [69]. Finally, the “ResNet-50” with the “Support Vector Machine” classifier outperformed other trained models.

In [70], three types of diseases and two pest infections of the cassava leaf dataset were identified by applying transfer learning of the “Inception-v3” architecture. A proprietary convolutional neural network and a 4-fold cross-validation approach were used to construct an autonomous and reliable plant disease identification model, and the emerging algorithm had an average precision of 82.3% [71]. The deep learning classifier categorised four potato plant diseases in [72]. The convolutional neural network models employed were VGG16 and VGG19. Finally, the VGG16 model obtained an average accuracy of 91% higher than the accuracy of 90% (average) achieved by the VGG19 model for potato plant disease classification. Four different tomato plant diseases were successfully identified and categorised using a hybrid model of CNN and the “Learning Vector Quantization (LVQ)” technique, with an average classification accuracy of 86% in [73]. In [74], a comparative study was presented to select the most suitable deep learning architecture for identifying plant diseases. Furthermore, in [75], the PlantVillage dataset was utilised for training the deep learning-based architectures “VGG16” and “AlexNet,” which both attained classification accuracy of 97.29% and 97.49%, respectively, using the healthy class and six-leaf disease classes of tomato plants. In the methodologies mentioned above, Grad-CAM and Score-CAM visualisation methods are not used to spot infections of plant diseases and pest lesions. Table 3 summarises the key advantages and limitations of using non-visualisation techniques in deep learning models for plant disease and pest detection.

images

3.1.2 With Visualization Technique

Plant pest and disease symptoms have been better visualized and understood using several deep learning architectures. Saliency maps [76] and occlusion approaches [77] have drawn attention to disease-affected areas, making them easier for farmers to understand. An Ultra-Lightweight Efficient Network (ULEN) outperformed SVM in [78] with a classification accuracy of 96.30%. Nine tomato leaf diseases were classified using ResNet-50, Xception, MobileNet, ShuffleNet, and DenseNet121_Xception in [79], with early convolution layers successfully identifying impacted areas. In [80], an enhanced LeNet used edge mapping to identify localized olive plant diseases. Reference [81] suggested a deep model for cucumber plants that accurately outperformed SVM, RF, and AlexNet. YOLOv4 produced the most favourable classification results when object detection models, such as SSD, YOLOv4, and Faster R-CNN, were tested for scale pest localization [82]. In [83], a novel teacher-student framework enabled region-level visualization and precise lesion identification.

The percentage of plant regions affected by disease has been accurately determined using deep learning models combined with different detectors [84]. Faster R-CNN, R-FCN, and SSD combinations with backbone networks including AlexNet, VGG, ZFNet, ResNet-50, and ResNetXt-101 were tested; the most efficient results were obtained with R-FCN + ResNet-50—the use of bounding boxes allowed for accurate disease localization. Using ResNet-50, Inception-V2, and MobileNet-V1 combined with Faster R-CNN and SSD, banana leaf diseases were identified in [85]. Similarly, a CNN-based pipeline for maize detected northern leaf blight (NLB) lesions, which were then evaluated using receiver operating characteristic (ROC) curves and annotated using CyVerse tools [67]. Accurate in-situ detection of rice leaf disease was made possible by Faster R-CNN with an improved Region Proposal Network (RPN) architecture, which produced high-quality regions of interest [86].

Many deep learning models have been successfully used to detect pests and plant diseases. VGG-based CNN and fully convolutional networks displayed wheat disorders [87], but LeNet correctly recognized soybean diseases [88]. In [89], radish regions were divided into healthy and Fusarium wilt-affected areas using the VGG model, with bacterial patches being identified using K-means clustering. Similar deep learning techniques were used to visualize plant symptoms in [90], and UNet and PSPNet with ResNet-50 were used to classify the severity and stress of coffee leaves based on disease spots [91]. Hidden neurons for pest detection were adjusted by an improved YOLOv3 using Adaptive Energy-based Harris Hawks Optimization (AE-HHO) [92]. Tiny YOLOv3 detected multi-class insect sites on sticky traps [93], while CNNs indicated areas of rice crops impacted by infection [94].

In order to guarantee color stability, target spots were identified from segmented images using a hotspot-based technique that uses descriptors for bacterial color characteristics and scorching spot textures [95]. In [96], field crop insects were classified using AlexNet, ResNet, GoogLeNet, and a suggested DCNN model; the DCNN demonstrated good reliability. To efficiently identify and categorize multi-class pests, “PestNet,” a region-based end-to-end architecture first presented in [97], included a Channel-Spatial Attention (CSA) unit, Region Proposal Network (RPN), Fully Connected (FC) layers, and Position-Sensitive Score Map (PSSM). A hybrid model that combined ZFNet with RPN and Non-Maximum Suppression (NMS) performed better than more conventional models like AlexNet and ResNet for the localization and counting of insects [98]. In [99], fruit crop diseases were automatically segmented and recognized using visualization techniques based on VGG-16 and AlexNet. The real-time model “Faster DR-IACNN” successfully identified grape leaf diseases in [100], whereas LeNet, in conjunction with color space and vegetative indicators, successfully identified grape plant diseases in [101].

By combining dilated convolution and global pooling, the Global Pooling Dilated Convolutional Neural Network (GPDCNN) successfully recognized six prevalent cucumber leaf diseases in a different study [102].

In situ experimental observations must consider real-world circumstances to assess how well deep learning architectures identify plant diseases and pests. Their application for practical disease detection and classification is limited because most of the approaches covered above rely on datasets gathered in controlled or typical environmental circumstances [50,6567,7780,94]. Real-time field circumstances have been addressed explicitly in only a small number of studies [66,69,70,84,87,89,102]. Several of these works’ visualization techniques are depicted in Figs. 311.

images

Figure 3: Identification of maize disease (specified by orange rectangles) by heat map [45]

images

Figure 4: Demonstration of activation in the first layers of an AlexNet model to show that it has learned to efficiently stimulate against the affected areas of the target leaf image of “Apple Cedar Rust” (a deep blue circle indicates the diseased patches) [79]

images

Figure 5: Olive leaf disease identification using segmentation and edge map [80]

images

Figure 6: Bounding boxes mark the predicted outcomes of the object detection model to identify three specific pests [82]

images

Figure 7: (a) Visualization algorithm results are compared; (b) Comparison of Grad-CAM [82]

images

Figure 8: (a) A novel “Teacher/student” framework; (b) “Binary threshold algorithm” for fragmentation [83]

images

Figure 9: Feature maps representation after the convolution layers applied to an image [102]

images

Figure 10: Visualization of tomato leaf pictures using Score-CAM with accurate classification: sliced leaves with a score-CAM heat map over the original leaves [103]

images

Figure 11: Saliency maps for “Bacterial spot,” “Late blight,” “Leaf mold,” and “Yellow leaf curl virus” are shown in the first, second, third, and fourth rows, respectively [104]

Fig. 3 identifies the infected maize plants using the heat map. Initially, each location of the selected image contains a lesion, which can be considered a measure of the probability that each lesion was caused by the disease. After that, a matrix form is generated using the probabilities, indicating the model’s outcome for each target image region [45]. The feature maps in Fig. 4 were obtained to understand the importance of CNN topologies in discriminating between different plant diseases. The output demonstrates the decent efficiency of the “AlexNet” architecture, which is capable of detecting plant diseases [79]. This figure also illustrates how characteristics are extracted using shallower Convolutional Neural Networks (CNNs). Compared to deeper models such as ResNet, the coarse and less selective activations of AlexNet may hamper performance in fine-grained disease classification. In Fig. 5, the plant diseases are detected using segmentation and edge maps. In the segmentation map, the yellow region appears as a white-coloured surface to highlight the lesion portion of the input leaf [80]. Segmentation, as opposed to bounding boxes, offers distinct boundaries, enhancing the quality of training data and the prediction of disease severity. Separating clustered leaf symptoms is further aided by edge-enhanced segmentation. The matrix reveals the probabilistic occurrence of lesions across spatial areas. By combining probabilities over time, these heatmaps may assist in improving severity assessment, allowing for customized treatment as opposed to uniform spraying. In Fig. 6, the three classes of pest detection are presented. RF points to the “mealybugs” pests, RR means “Coccidae”, and RS indicates “Diaspididae”. Each image contains multiple predicted outcomes [82], and for class-wise detection, including mealybug and coccidae, is depicted in the figure. Visually, the detection seems accurate, but no metrics are displayed. A detailed comparison in the updated publication demonstrates that SSD achieves 82.5% mAP on the dataset, with an accuracy of 89% for mealybugs and 76% for coccidae. illustrates how distillation sharpens the student model’s focus, enhancing interpretability and mitigating a novel approach for visualisation is demonstrated in [83], as presented in Figs. 7a,b and 8a,b. The selected imagery has been reassembled in Fig. 8a to create the “student/teacher” model [83], and in Fig. 8b, a single channel heat map has been created and segment significant portions of the image. After that, a straightforward Boolean threshold method reliably displays signs of severe plant infection. This design blends a student’s efficiency with high-level semantics from a teacher model. This image illustrates how distillation sharpens the student model’s focus, enhancing interpretability and mitigating overfitting on real training data. Fig. 7a presents the different visualisation approaches. Heat maps created using visualisation approaches like “Gradient,” “LRP-Z,” and “LRP-Epsilon” are hazy and difficult to understand. The deep Taylor technique has achieved superior outcomes, but it has specified only a few spots of the disease on the leaf. Furthermore, a defective localisation of the disease spots is achieved using the Grad-CAM technique. At the same time, the proposed method addressed the issue by employing a decoder, as shown in Fig. 7b [82]. As a conclusion of the figure, Deep Taylor and LRP concentrate on relevance at the pixel level, whereas Grad-CAM emphasises more general regions. Although LRP provides superior noise reduction in the updated statement, it may overlook contextual cues. According to a user assessment, agronomy specialists rated Grad-CAM as the most interpretable. Fig. 9 shows how the feature map uses pooling, convolution, and ReLU operations to process an image so that different features can be extracted from other layers. Every feature map is depicted, and ultimately, the Global Average Pooling (GAP) layer records several noteworthy attributes [102]. Additionally, the progressive abstraction of characteristics in a CNN is illustrated in this figure. Deeper layers identify intricate disease patterns, such as lesions or pest forms, while early layers record low-level information, such as edges and textures. For reliable categorization in a variety of illumination and occlusion scenarios, this layered abstraction is essential. Figs. 10 and 11 illustrate two distinct visualization techniques: saliency maps and heat maps. The Score-Weighted Visual Explanations for CNNs (Score-CAM) heat map is used to identify infections in the source images of Fig. 10 as red dots [103] and saliency maps aid in seeing the active pixels affecting the model’s prediction, but noisy backgrounds may make them difficult to understand. They are faster than CAM-based techniques, although they are frequently less accurate in pinpointing the locations of symptoms. A superior methodology, known as the saliency map technique, is introduced, which effectively detects all disease-affected spots on tomato plants, as shown in Fig. 11 [104]. This technique generates more concentrated attention maps than saliency, a class-specific activation. This figure illustrates how Score-CAM enhances field-level diagnostic accuracy by distinguishing between overlapping symptoms of disease. Table 4 shows different visualisation mapping/techniques to summarise this section. These methods, which highlight key image regions that make predictions, such as Grad-CAM, saliency maps, and feature attribution, improve the interpretability of deep learning models. In crucial applications like plant disease and pest detection, this kind of visualization aids in determining the model’s target region, fostering trust and facilitating validation. Table 5 summarizes the key advantages and limitations of visualization techniques in deep learning models for plant disease and pest detection.

images

images

3.1.3 Modified Deep Learning Architectures for Plant Pests and Diseases Identification

Several new deep learning architectures have been applied in some works to achieve superior and apparent identification of plant diseases. For instance, reference [105] provided an improved “GoogLeNet” architecture, and this model’s usefulness is significantly superior to that of the “AlexNet” and “VGG” architectures. The updated versions of these traditional deep learning models have achieved 98.9% accuracy. In [106], a fresh combination of “DenseNet” with “Xception” based “DCNN” is proposed for the recognition & early diagnosis of apple tree leaf diseases accurately. The proposed approach is 98.82 percent accurate in identifying five diseases in apple leaves. To identify, categorise, and count pests, a unique “entire-and-partial feature transfer learning” approach is presented; finally, it provides the detection frequency (90.2%) of pest movement [107]. Different fine-tuned feature maps are reinforced utilising the “entire-feature transfer learning” weight matrix while utilizing the “partial-feature transfer learning” approach. Finally, the whole feature network’s cross-layer and multi-scale convolution layers are integrated. To categorise the soybean pests, transfer learning techniques are employed in [108] and in conjunction with the deep learning models “Inception-v3”, “VGG-19”, “Resnet-50”, and “Xception”. The entire soybean dataset comprises 5000 images, captured in real-time. The ResNet-50 deep learning model achieves the highest classification accuracy at 93.82%. Haridasan et al. [109] conducted a study focusing on the automatic detection and analysis of rice crop diseases by incorporating various pre-processing techniques and utilising Principal Component Analysis for dimensionality reduction. In their work, Abbas et al. [110] introduced an enhanced Cuckoo Search algorithm termed the Hybrid Improved Cuckoo Newton Raphson Optimization (HICNRO). This approach not only diminishes computational time but also eliminates redundant features and enhances accuracy. These selected features are then utilised to train ten distinct machine learning classifiers, resulting in notable accuracies of 100%, 92.9%, and 99.2% on Tomato, Potato, and Cucumber Leaf datasets. Khattak et al. [111] proposed a CNN model with a test accuracy of 94.55 %, making it a valuable decision support tool for farmers to classify citrus fruit/leaf diseases.

In [112], LeafNet, a new CNN model, is employed for tea leaf disease classification, achieving greater accuracy compared to the “Multi-Layer Perceptron (MLP)” and “Support Vector Machine (SVM)” models. In [113], a “deep convolutional generative adversarial network (DCGAN)” was trained on a segmented and filtered dataset containing 56,000 images of 19 crops, which comprise 38 classes collected from PlantVillage. A CNN model such as “MobileNet” deployed on smartphones is used to classify plant diseases and obtained 92% test accuracy compared to the modified “Inceptionv3” model with 88.6% accuracy. Two deep learning architectures, namely the reduced “MobileNet” and reformed “MobileNet”, are proposed in [64], and the obtained classification accuracy is similar to that of the “VGG” architecture. Finally, the reduced “MobileNet” has obtained 98.34% accuracy. This model significantly reduces training time, as it has fewer parameters than the VGG model. In [65], the hybrid CNN architecture, known as the “VGG-Inception” model, is used to classify and identify five types of diseases in apple plants. This model outperformed several deep learning models, including AlexNet, GoogLeNet, VGG, and specific variants of ResNet. In addition, it offers inter-object and class identification and activation visualisation. Another advantage of using this model is that it perfectly visualises disease-affected spots on plants. In a different study, Liu et al. [114] developed an enhanced Faster R-CNN with feature pyramid fusion (FPF-RCNN) that maintained a detection speed of 21.5 FPS while achieving 79.2% mAP on the Citrus Pest and Disease Dataset (CPDD). Furthermore, Amin et al. [115] utilise two pre-trained convolutional neural networks (CNNs), EfficientNetB0 and DenseNet121, to extract deep features from the corn plant images, ultimately achieving a classification accuracy of 98.56%.

The most prominent deep learning algorithms for accurate plant infection identification and categorisation are shown in a bar chart in Fig. 12. Most research work has applied the AlexNet [28] model. The following widely used deep learning architectures are ResNet-50 [85], VGG-16 [75], and GoogLeNet [58]. Correspondingly, various hybrid deep learning architectures, such as VGG-Inception [65], reduced or modified MobileNet [64], updated LeNet [80], and modified GoogLeNet [105], are employed to identify pests and diseases in plants. In [116], Pacal et al. (2024) conducted a comprehensive review analyzing 160 studies on deep learning applications in plant disease detection, focusing on classification, detection, and segmentation tasks. The paper highlights the effectiveness of deep learning in early disease detection and discusses challenges such as dataset limitations and model generalization.

images

Figure 12: Plant infections and pests are frequently detected using deep learning designs

3.1.4 State-of-the-Art Transformer Architectures for Plant Pests and Diseases Identification

A new advancement in deep learning, vision transformers (ViTs) utilise the transformer architecture, originally designed for natural language processing, to analyse images. Vision transformers split an image into tiny patches and treat them as tokens in a sequence, unlike conventional convolutional neural networks (CNNs), which scan images using localized filters. This allows the model to capture global context and long-range dependencies inside an image. This marks a departure from traditional sequence-to-sequence models, which often rely on recurrent neural networks (RNNs), such as Long Short-Term Memory (LSTM) [117], predominantly employed in Natural Language Processing (NLP) applications. In the context of images and visual data, where inherent spatial and temporal coherence prevails, the innovative use of transformers signifies a paradigm shift in how we approach information capture and contextual understanding with it being successfully used in image segmentation [118,119], image classification [120], and so on. Introduced in [35], the Vision Transformer (ViT) represents a groundbreaking approach to image recognition. Departing from conventional convolutional neural networks, ViT embraces a transformer architecture originally devised for natural language processing. The model treats images as sequences of patches, facilitating the effective capture of extensive contextual information and long-range dependencies. Nevertheless, ViTs frequently encounter challenges associated with substantial memory requirements, compounded by inherent limitations in receptive fields, local context understanding, and the sharing of weights.

Substantial research has been conducted to overcome the challenges encountered in typical transformer-based computer vision models. These works could be broadly classified into two categories: hybrid transformers, which typically use convolutional networks to overcome limitations in understanding local contexts, and pure transformers. Pure transformer-based computer vision models exclusively employ self-attention mechanisms for capturing long-range dependencies and global contextual information. Dosovitskiy et al. [35]’s ViT marks a pioneering milestone as the first to leverage transformers for computer vision tasks. Li et al. [38] introduce a PMVT transformer model, a novel approach that replaces the convolution block in MobileViT with an inverted residual structure that uses a 7 × 7 convolution kernel. On various mobile devices, the PMVT network maintains high accuracy across a range of visual tasks while achieving lower parameter counts compared to competing networks. From an experimental point of view, with roughly 0.98 million (M) parameters, PMVT attains the highest accuracy of 93.6% in the wheat data set. This accuracy surpasses MobileNetV3 by 1.6%. Liu et al. [36]’s fundamental innovation in the Swin Transformer lies in its unique shifted windowing mechanism, which limits self-attention computation to distinct local windows strategically shifted across various layers. The resulting hierarchical architecture has linear computational complexity and achieved a top-1 accuracy of 87.3% on the ImageNet-1K dataset. Touvron et al. [121] utilised a novel token-based distillation approach in their proposed Data-efficient Image Transformer (DeiT) to demonstrate the application of transformer-based vision models on mid-sized datasets.

A lightweight model, Tiny-LeViT, based on the hybrid transformer-CNN in [39], trained on five distinct datasets, at least 9% higher frame rates. Lu et al. [122] introduced an efficient approach for grape leaf diagnosis using ghost convolution [123] and transformer networks. They curated the GLDP12k dataset, which comprises 11 classes and 12,615 images. Transfer-learned on the ImageNet dataset, the Ghost Enlightened Transformer model (GeT) outperformed other models, offering 98.14% accuracy, 1.7 times faster processing (180 frames per second), and 3.6 times lighter weight (1.16 million) than MobileNetV3_large [124]. Wu et al. [125] present CvT, a novel model seamlessly incorporating convolutions into the ViT framework. This integration introduces a convolutional token embedding layer and a convolutional transformer block, skillfully amalgamating the shift, scale, and distortion invariance inherent in CNNs with the dynamic attention and global context of transformers. It adopts a hierarchical structure, systematically reducing the number of tokens while increasing the token width, emulating the spatial downsampling impact observed in CNNs. Their CVT-w24 variant obtained a top-1 accuracy of 87.7% on the ImageNet-1k val dataset.

Li et al. [126] propose LocalViT, where they achieve the locality mechanism by integrating depth-wise convolution into the feed-forward network of each transformer block. The utilisation of depth-wise convolution focuses on individual channels of the input, facilitating the capture of local relationships among pixels within a confined neighbourhood. This approach differs from self-attention, which concentrates on discerning long-range dependencies across all elements in the input. Guo et al. [127] present the CMT block, comprising a Local Perception Unit (LPU), a Lightweight Multi-head Self-Attention (LMHSA) module, and an Inverted Residual Feed-Forward Network (IRFFN). The LPU, employing depth-wise convolution, ensures translation invariance while capturing local information, preserving model integrity. Within the LMHSA module, depth-wise convolution enhances self-attention, reducing computational load and introducing relative position biases for improved performance. The IRFFN adapts the original ViT feed-forward network (FFN) by incorporating an inverted residual block with carefully placed shortcut connections, enhancing outcomes.

In [128], the embedding stem of the ViT was substituted with a minimal convolutional stem, which altered the early visual processing of the ViT, accelerating the convergence of the models and offering valuable insights into optimising the training dynamics of ViT models. The convolutional stem design employed stacked 3 × 3 convolutions, culminating in a 1 × 1 convolution to match the transformer encoder’s d-dimensional input. Employing a consistent pattern, 3 × 3 convolutions use a stride of 2, doubling output channels, or a stride of 1 with a constant channel count. In [129], Pacal and Işık (2024) proposed an ensemble approach combining CNNs and vision transformers to achieve 100% accuracy in corn leaf disease detection on the CD&S dataset. Their method incorporated advanced techniques, such as data augmentation and soft voting, setting a new benchmark in agricultural disease diagnostics. Table 6 summarises that ViT and Swin Transformer are examples of transformer-based designs that exhibit exceptional results on large datasets, such as ImageNet, demonstrating their capacity to model global context. On the other hand, lightweight models such as PMVT and Tiny-LeViT are more suitable for deployment in settings with limited resources. These models are ideal for precision agriculture activities in real-time environments because they preserve efficiency while achieving high F1-scores. Moreover, Table 7 provides a comparison analysis, showing that CNNs remain highly effective in applications that require real-time processing on edge devices and local pattern detection. Transformer models, on the other hand, offer better contextual awareness, which is particularly useful in complex and expansive agricultural settings. Transformer versions (such as PMVT and Tiny-LeViT) are becoming lighter, which increases their use in field situations.

images

images

3.2 Computational and Hardware Demands for Agricultural AI System

When implementing deep learning models in actual agricultural environments, hardware capabilities, energy limitations, and the scalability of computational resources must all be carefully considered. The dataset amount, the model’s complexity, and the real-time processing demands in field settings significantly affect the deployment’s viability.

High-End Server/Cloud-Based Inference:

For training or running large-scale inference using heavy models such as VGG-19, DenseNet, or ViT:

•   Recommended Hardware:

NVIDIA GPUs (e.g., RTX 3090, A100, or Tesla T4 for cloud)

32–64 GB RAM

1–2 TB SSD (for high-throughput image storage)

These setups are suitable for:

Centralised model training

Cloud-based decision support systems for multiple farms

Drawback: Requires consistent connectivity and is expensive to maintain in remote areas.

Edge Devices for On-Field Deployment:

Lightweight models such as Reduced MobileNet, Tiny-LeViT, and SqueezeNet are preferable for field deployment (e.g., on drones, smartphones, and smart sensors).

•   Recommended Edge Devices:

Raspberry Pi 4 with Coral Edge TPU

NVIDIA Jetson Nano/Xavier

Smartphones with Qualcomm AI chips (Snapdragon)

These devices can run models with <10 M parameters and offer real-time predictions with low latency.

Benefits:

Cost-effective and portable

No internet dependency

Real-time feedback for pest/disease alerts

Model Optimization Techniques for Deployment:

•   Quantization and pruning: Reduce model size and computation for faster inference.

•   Knowledge distillation: Transfer knowledge from larger models to smaller student models.

•   ONNX/TFLite Conversion: Convert models to lightweight formats for mobile and embedded systems deployment.

Scalability Considerations:

•   In large-scale operations covering hundreds of acres, distributed sensing and computing systems are needed.

•   Integration with UAVs (drones), IoT sensors, and robotic arms requires:

Lightweight, fast models

Efficient power usage (battery optimization)

On-device inferencing or edge-cloud hybrid systems

Table 8 summarizes that large models are supported by cloud-based systems for centralized decision-making, but they need steady connectivity. Lightweight models allow offline, real-time inference on smartphones and edge devices. Every platform involves trade-offs among accuracy, latency, and power efficiency; field circumstances and scale should be considered when making deployment decisions.

images

4  Hyper-Spectral Imaging for Plant Pest and Disease Identification Using Deep Learning Architectures

Several imaging techniques have been developed for the early recognition of infections and pests that affect plants, including spatial and spectral imaging (multispectral) [130], infrared imaging, and hyperspectral imaging [131]. Hyperspectral imaging is an effective method of acquiring images that store information over hundreds of continuous, narrow spectral bands, much wider than the human eye can see. The ability to identify minute changes in plant tissue, including moisture content, chlorophyll content, or disease symptoms, makes it particularly useful for detecting early plant diseases and insect infestations. For example, in [132], the “Tomato Spotted Wilt Virus (TSWV)” was discovered in capsicum vines using hyperspectral imaging technology (VNIR and SWIR) as well as many machine learning-based algorithms. In another work [133], tomato plant diseases were identified using a hyperspectral imaging technique, where the “Region of Interest (RoI)” was determined, and a “feature ranking-KNN (FR-KNN)” architecture achieved an acceptable outcome for the identification of healthy and disease-affected plants. Recently, HSI was employed for apple disease identification, and the “Orthogonal Subspace Projection” method, an unsupervised feature selection approach, solved the redundancy problem [134]. A novel hyperspectral imaging-based methodology was introduced [135] to detect anthracnose in tea plants. Using these bands, two novel disease indexes were produced: (a) “Tea Anthracnose Ratio Index (TARI)” and (b) “Tea Anthracnose Normalized Index (TANI)” its experimental analysis revealed that the detected disease-sensitive bands were at 542, 686, and 754 nanometers, which were learnt by evaluating the spectral sensitivity. Based on an idealised set of wavelet coefficients for disease scab identification, “unsupervised classification” and “adaptive two-dimensional thresholding” techniques had been formalised. Another study [136] demonstrated that, under varying mangrove pest and disease severities, sensitive spectral and textural features can be precisely retrieved from HSI leaf data using the Successive Projection Algorithm (SPA) and Random Forest (RF) techniques.

Hence, combining HSI with machine learning approaches is advantageous. For example, in [137], machine learning methodologies were designed for HSI to solve several issues in the agricultural domain. The “citrus canker” is a remote sensing methodology that detects disease of sugar belle leaves and green fruit under laboratory conditions in [138]. Underdiagnosed, intermediate, and delayed disease symptoms growth phases were examined using a hyperspectral imaging system (400–1000 nm). In this work, the classification approaches “radial basis function (RBF)” and “K-nearest neighbour (KNN)” are employed. An identical UAV-mounted imaging system also detected the “citrus canker” on tree canopies on a large farm. Moreover, the HSI-based multi-step method was introduced [139] to differentiate several plant stresses. The process involved: (1) a “Continuous Wavelet Analysis (CWA)” based feature extraction process to detect and discriminate the Tea plant stresses; (2) the “SVM” and “K-means clustering” method detected some lesion regions on the tea leaves; (3) The creation of a system that uses the Random Forest (RF) algorithm to identify and distinguish between three tea plant “stresses—tea green leafhopper”, “anthracnose”, and sunburn (disease-like stress). For the detection of wheat disease, the multispectral imaging method based on the Random Forest (RF) predictor achieved an overall accuracy of 89.3% [140]. In [141], a machine learning approach based on HSI (380–1020 nanometres) identified the “powdery mildew” disease growth phases (asymptomatic, early, midway, and late disease phases) in the plant of squash. The “Support Vector Machine”(SVM)-based HSI system detected different plant diseases, and the developed model obtained 86% overall accuracy [142]. In their research, Javidan et al. [143] effectively utilised the PlantVillage dataset to identify and classify grape leaf diseases, encompassing black measles, leaf blight, and black rot. Employing K-means clustering, they delineated affected regions and extracted features from three distinct colour models: L*a*b, HSV, and RGB. Using multi-class support vector machines (SVMs), their method obtained 98.71% accuracy by employing Grey-level co-occurrence matrix features and 98.97% accuracy by employing principal component analysis (PCA) on the extracted features.

Several scientific studies on plant pests and disease monitoring have employed deep learning designs based on hyperspectral imaging (HSI) to achieve flawless visualisation of symptoms caused by plant diseases and pest attacks. To classify the hyperspectral images, a hybrid methodology using deep convolutional neural networks (DCNN), logistic regression (LR), and principal component analysis (PCA) has been developed. This methodology produced more advanced results than previous methods for classification tasks. In [144], the proposed “3D-CNN” architecture identified the charcoal rot disease of soybean plants by assessing the hyperspectral images collected at 240 various wavelengths in the range of 383–1032 nanometers. The trained model was developed in this work by applying a saliency map and visualising the best delicate pixel locations to smooth the classification process. Moreover, convolutional neural networks (CNNs) are utilised on visible/near-infrared (Vis/NIR) based hyperspectral imaging (HSI) systems (376–1044 nm) [145] for the detection of aphid-attacked lesions in cotton leaves where the RGB and HSI specimens are compared head-to-head, along with a comparison of the suggested 1DCNN and 2DCNN, and a test set of images of healthy and contaminated leaves in RGB and hyperspectral colour; the depiction of saliency maps for test sets contains two classes of samples over five periods, and the labels of each sample are correctly predicted in Fig. 13. The overfitting issue was addressed in [146] where a detailed observation-based comparison work was established between different deep learning methodologies of 1D and 2D-CNN (2D-CNN performs better), “Long Short-Term Memory”, and “Gated Recurrent Unit (GRU)” (both faced the overfitting issue), and “2D-CNN-LSTM/GRU” (yet overfitting occurs). Hence, the combination of a convolutional and a “bi-directional gated recurrent network” (termed as “2D-CNN-BidLSTM/GRU”) model was applied to the HSI images to address overfitting issues, ultimately yielding an accuracy of 0.73 and an F1-score of 0.75 for wheat disease identification [147]. In [148], “Outlier Removal (OR-AC-GAN)”, a unique “generative adversarial network” (GAN)-based hyperspectral proximal-sensing method, detected tomato plant infections before clinical disease signs manifested (as presented by Fig. 14). The green regions in the figure indicate healthy plant tissue, while the red areas highlight lesions. To identify diseases and infections caused by pest attacks on tomato plants in [149], a novel deconvolutional network for feature visualisation methodology was proposed to examine the effects of colour and spectral information in images of tomato leaves. Another cutting-edge, deep learning-based hybrid strategy combining terahertz imaging technology and near-infrared hyperspectral imaging technology was utilised to determine the bacterial blight immune rice seeds.

images

Figure 13: Both RGB and hyperspectral images’ saliency maps. RGB images are shown by 2D CNN saliency maps, and hyperspectral images are shown by 3D CNN saliency maps [145]

images

Figure 14: Results of specific tests conducted on plants using the new “OR-AC-GAN”. “Tomato spotted wilt virus (TSWV)” pixels are marked as red, whereas healthier pixels are marked as green [148]

Despite the emergence of several HSI-based deep learning architectures for plant disease diagnosis, significant research gaps remain in addressing environmental variability. The goal of this research is to enhance the robustness of disease diagnosis under challenging conditions, such as changes in illumination, spectral reflectance, and field occlusion situations [150]. Table 9 outlines the advantages and limitations of hyperspectral imaging in disease and pest detection in crops and plants.

images

Additionally, Table 10 provides a comprehensive summary of recent research using several deep learning architectures for the diagnosis of plant diseases and pests. This analysis reveals that ResNet, VGG, and DenseNet are the most widely used models, primarily due to their robustness against a diverse range of crop varieties and environmental conditions. MobileNet and its lightweight variants are preferred for real-time detection on edge devices. To increase generalizability, augmentation techniques, including flipping, rotation, and contrast enhancement, are often employed. Notably, transformer-based methods are becoming highly accurate and potent substitutes, particularly when complex symptoms or extensive field data modeling are needed.

images

5  Research Challenges and Possible Solutions

In this section, we have discussed research gaps and potential solutions for detecting plant pests and diseases. After thoroughly analysing each research work in this manuscript, few major challenges are observed.

•   Dataset Size & Preprocessing Issue

•   Early disease and pest attack detection

•   Detection performance under the illumination and occlusion conditions

•   Detection Speed

•   Overfitting and Generalizability in Deep Learning

•   Computational Challenges of Transformers

•   Model Deployment Strategies for Real-World Agricultural Applications

•   Integration of Deep Learning Models into Agricultural Pest Management Systems

Ultimately, our discussion demonstrated the significant strides made by deep learning in identifying plant diseases and pests in this subsection:

•   Recent Breakthroughs and Emerging Trends of Deep Learning in Plant Pest and Disease Detection

5.1 Dataset Size & Preprocessing Issue

Plant disease and pest identification are specific areas of the agricultural domain. The available plant disease and pest datasets for this task need to be more extensive and comprehensive. Self-collected datasets often lack the thoroughness, accessibility, scale, and labeling quality of publicly available datasets. For example, some plant diseases are rare, and retrieving damaged plants is expensive and time-consuming. This restricts the efficacy of deep learning applications in plant disease detection and pest recognition because, in many cases, only a few images are gathered for training. Three different types of solutions have been mentioned here:

5.1.1 Data Augmentation and Generation

A major element of deep learning model training is data augmentation. Data augmentation strategies are typically used to add modified existing data to the training data.

Some popular data augmentation techniques used in the plant pest and disease detection domain are:

(a) Rotation: To replicate various orientations, rotate images by 90, 180, or 270 degrees, etc.

(b) Flipping: To create the illusion of mirror reflections, flip images vertically or horizontally.

(c) Scaling: images can be scaled up or down to represent diverse distances or resolutions.

(d) Translation: Simulating various locations by translating images vertically or horizontally. Data augmentation has the potential to substantially influence the effectiveness of deep learning models concerning illumination, occlusion, and other environmental conditions in several ways:

(i) Variations in Lighting:

•   Colour jittering: By incorporating images with varying lighting conditions (such as brightness, contrast, and saturation), models may be trained to be invariant to lighting changes.

•   Gamma correction: Gamma correction may be applied to images by simulating various lighting situations, strengthening models against lighting fluctuations.

(ii) Occlusion:

•   Random cropping: Models can learn to focus on pertinent features by simulating occlusions through random image cropping.

•   Occlusion simulation: Models can learn to recognize objects across varying surface appearances through texture and pattern (such as rectangles or polygons) augmentation. Occlusion issues are discussed comprehensively in the “Detection Performance Under the Illumination and Occlusion Conditions” section.

(iii) Environmental Factors:

•   Weather simulation: Models can learn to identify objects in various environmental situations by simulating weather conditions (such as rain, snow, and fog).

•   Texture and pattern augmentation: Models can be trained to identify objects despite differences in surface appearance by adding images with various textures and patterns.

Data augmentation has several advantages. Here are some of the key benefits:

•   Lowers the price of labelling and collecting data.

•   Improves the model’s prediction accuracy.

•   Prevents data scarcity to get optimal results.

•   Reduces the probability of data overfitting.

•   The variability and flexibility of the model increase.

•   Resolve the class imbalance issues for classification.

Moreover, “Generative Adversarial Networks (GANs)” [151] and “Variational Autoencoder (VAE)” [152] can be employed to generate various samples, thereby improving datasets on plant diseases and pests.

Table 11 was added to improve readability and facilitate reproducibility. It presents the most popular preprocessing methods systematically, describing their particular uses and providing examples of studies that have used them to detect pests and plant diseases.

images

5.1.2 The Use of Fine-Tuning and Transfer Learning

Transfer learning mitigates minor dataset limitations. The goal of employing transfer learning is to transfer information from a large dataset to a smaller dataset that is roughly similar in nature. Unlike in a deep learning model, where the parameters are adjusted or changed, this approach can prove extremely helpful for accurately identifying plant diseases and pests, making it more affordable to train the model and enabling CNN to adapt to a small dataset. Transfer learning has the following benefits:

•   Transfer learning increases learning speed.

•   Transfer learning offers a deep learning model to achieve higher performance and obtain accurate results.

Vallabhajosyula et al. employed a deep-set neural network (DENN) to detect plant infections, utilising transfer learning to optimise the model’s parameters, thereby outperforming other cutting-edge methods [153]. A transfer-learned model is proposed for detecting diseases and pests in hot peppers in [154]. A deep CNN was applied by Huang et al. to categorise eight species of tomato plant pests. Here, transfer learning effectively reduces the training time [155]. In [156], a transfer learning approach using a feature extraction-based technique is developed to identify pearl millet mildew. This work utilises transfer learning to address the issue of small datasets.

5.1.3 Real-Time Detection Model Design

A real-time detection model can minimise the requirements of the plant pest and disease dataset. For example, a real-time deep-learning architecture for identifying tomato plant diseases was created by Fuentes et al. Different environmental issues, such as illumination effects, background and weather variations, etc., have been considered in this paper [157]. An optimized Inception-v3 architecture is demonstrated in [158] for the identification of tomato disease in its early stages. It uses its auxiliary classifiers and factorized convolutions to achieve 96.4% accuracy on datasets gathered in the field. Real-time deployment in precision agriculture is made possible by the model’s computational efficiency, which outperforms conventional CNNs in terms of speed and diagnostic reliability in a range of illumination situations (14.8 FPS on smartphones and 32.5 FPS on UAVs).

5.2 Early Disease and Pest Attack Detection

To sustain healthy production, it’s also crucial to promptly identify and treat plant diseases and pest lesions. Several research studies have found several early-stage detection issues.

Small Lesion Area Detection

Several deep neural network downsampling procedures have occasionally led to the deep neural network ignoring minor lesion items. Separating the small lesion areas of the low-resolution images is quite challenging. Additionally, background noise issues in the collected images may lead to false identification. Modern attention modules, including the Convolutional Block Attention Module (CBAM) in conjunction with spatial transformers, have been demonstrated to increase significantly object localization in agricultural images, which can help with the detection of small-scale disease symptoms in crops [159]. The principle of the mechanism is to identify the region of interest quickly and overlook other insignificant areas of the target image. After that, the weighted sum methodology with a weighted coefficient is applied to separate the features, which leads to invalidating the background noise of the image. Then, to create a new combination of features to lower the distortion, the softmax function merges the feature imagery with the initial feature image. Typically, the attention mechanism can construct a prominent image, eliminate small-scale items from that image, and disregard the background. In [160], a Recurrent Neural Network (RNN) is used with the “attention mechanism” to automatically detect lesion areas and extract key characteristics for disease classification.

5.3 Detection Performance under the Illumination and Occlusion Conditions

Environmental conditions, such as illumination and occlusion, also influence the detection accuracy.

5.3.1 Illumination

Variable illumination significantly reduces the detection accuracy of deep learning models. The majority of research efforts utilise existing datasets that contain images collected under predetermined lighting conditions. Hence, the image processing stages can be simplified. These images differ from those acquired in a live situation. As the natural light changes frequently, sometimes colour distortion occurs. Additionally, the differences in angle and distance during image capture also affect the features of the images, which can lead to poor visual recognition of pests and diseases.

The possible solutions are given below:

•   Develop algorithms that adapt to changing illumination conditions by dynamically correcting colour and brightness variations in real time.

•   Augment existing datasets with synthetic images generated under diverse lighting scenarios, providing the model with a more comprehensive understanding of varying illumination conditions.

•   Explore the use of multi-spectral or hyperspectral imaging technologies, which can capture a broader spectrum of light and potentially mitigate the impact of fluctuations in natural light.

•   Expand datasets by incorporating images with varying illumination levels, enabling the model to learn and adapt to more diverse lighting conditions.

5.3.2 Occlusion

Most research works in the plant pest and disease detection area have avoided the occlusion problem and rarely considered occluded images. The detection accuracy under occlusion is always low. Yet, occlusion problems frequently occur in real-time settings. Certain occlusions are caused by modifications in branch occlusion, leaf occlusion, light occlusion, etc., and they happen at random in various plant parts. So, the noise overlap and shortage of features in images are caused by the occlusion, which directly impacts recognition accuracy and produces false or missed detection.

Advanced algorithms based on deep learning have significantly improved the diagnosis of plant diseases and pests in various environmental conditions, including occlusion [161]. This sort of study application increases the efficacy of deep learning techniques in a natural context. The frameworks of deep learning architectures should be improved in the future to reduce occlusion issues.

The potential fixes are given here:

•   Utilizing multi-view imaging approaches involves capturing multiple perspectives of the same scene. By integrating images from various angles, the impact of occlusion is mitigated, enabling the model to obtain more comprehensive information about the target.

•   Algorithms specialized in recognizing partially occluded objects improve feature identification despite obstructions.

•   GANs can be employed to generate synthetic images that simulate occlusion scenarios.

•   Incorporating temporal information from consecutive frames in video sequences can contribute to occlusion resolution. By analyzing the movement and interactions of objects over time, the model gains insights that help overcome occlusion challenges.

•   Using depth sensors to reduce the effects of occlusion and offer 3D information, such as lidar or stereo cameras.

5.4 Detection Speed

Although deep learning is more accurate than conventional techniques in identifying diseases and pests, its high processing requirements may cause latency, which could delay prompt action in actual field situations. For real-time applications, this delay is inappropriate. Speed can be increased by reducing processing or training, but accuracy is at risk. Therefore, effective models must strike a delicate balance between speed and accuracy.

The plant disease and pest identification methodologies have three major sections:

•   Data collection and labeling

•   Model training

•   Model interpretation

Real-time pest and disease detection relies heavily on model interpretability, as understanding the justification behind forecasts not only promotes trust and informed decision-making but also supports the primary objective of achieving high predicted accuracy in real-world agricultural applications. Therefore, more attention should be given to the model’s efficiency. In [162], a unique “CenterNet” architecture built on DenseNet-77 is suggested for classifying plant diseases. According to the experiment results, the custom “CenterNet” model outperforms standard deep learning models currently in use. More transparent pest detection systems are now possible thanks to recent developments in explainable AI. In [163], a “Grad-CAM++ enhanced YOLOv5” architecture outperformed conventional CNNs in speed (28 FPS) and interpretability measures, demonstrating 96.2% classification accuracy while offering visual explanations of disease localization.

5.5 Overfitting and Generalizability in Deep Learning for Agricultural Pest and Disease Detection

Although deep learning models can accurately classify plant diseases and pests, they often encounter two significant issues: overfitting and poor generalizability to real-world scenarios.

Overfitting:

Overfitting occurs when models learn noise and dataset-specific artifacts rather than generalizable patterns. Regarding the identification of plant diseases:

•   A lot of research depends on specially chosen datasets, such as PlantVillage, which include excellent, noise-free images taken in controlled settings. When real-world images with different backgrounds, lighting, occlusion, or environmental noise are presented to models trained on these datasets, they frequently exhibit subpar performance.

•   When there is a shortage of labeled agricultural data, deep models like VGG16, ResNet-50, or DenseNet are susceptible to overfitting because of their high parameter counts.

Mitigation Strategies:

•   Rotation, flipping, brightness modulation, and other data augmentation techniques are used to increase robustness.

•   Overfitting can be decreased during training by utilising dropout layers, batch normalisation, and regularisation strategies.

•   One popular method of addressing data scarcity is to transfer learning from pre-trained models on large-scale datasets (such as ImageNet).

Generalizability:

Generalization refers to the ability of a model to continue performing in unknown circumstances, such as:

•   Varying crop species, geographic locations, and imaging modalities (e.g., smartphone images vs. hyperspectral), is known as generalization.

•   Symptoms of the disease vary depending on the stage of plant growth or environmental conditions (e.g., soil type, humidity).

Challenges:

•   It’s possible that Deep Learning models developed using datasets unique to a region or crop won’t generalize to other crops or conditions.

•   Many currently misused models assume that input image quality and crop types are uniform.

Potential Solutions:

•   Developing and disseminating extensive, varied, annotated datasets encompassing various crops, disease stages, and situations.

•   Models can be adapted to new domains with less labeled data by investigating domain adaptation and few-shot learning.

•   Transformer architectures and attention mechanisms are promising but still being investigated since they can more accurately model complex characteristics and long-range dependencies in various scenarios.

Future Research Directions: There are several advantages to the field:

•   Comparing models to real-world datasets from field circumstances using suitable criteria (such as robustness and uncertainty) that go beyond accuracy.

•   To comprehend why models fail in particular scenarios, explainable AI strategies are being investigated.

•   Deep Learning can improve reliability with expert-driven rules or multimodal sensing (visual and hyperspectral).

5.6 Computational Challenges of Transformers

The excellent ability of transformer-based models to capture global features has drawn recent interest in agricultural image analysis. However, their application comes with certain limitations. These models require substantial memory footprints, lengthier training durations, and computational resources when used on large datasets. It’s critical to understand the resource implications and practical challenges of training and implementing these models in agricultural settings before exploring their potential. A summary of the main obstacles to training transformer models on large datasets is provided below:

•   High Computational Cost: Vision Transformers (ViTs) scale quadratically with input size due to self-attention, demanding greater computational resources than CNNs. Therefore, they are demanding in terms of GPU memory, training duration, and energy usage.

•   Data-Hungry Nature: Transformers frequently perform poorly on tiny or modestly sized datasets unless they have been extensively enhanced or pre-trained on large corpora (like ImageNet). In contrast to CNNs, which can function rather well with less supervision, they may overfit or fail to generalize in the absence of enough data.

•   Training Complexity: Transformers are more difficult for inexperienced practitioners to optimize since they usually require longer convergence times and meticulous hyperparameter adjustment (such as learning rate warm-up, attention dropout, and positional encoding techniques).

•   Limited Edge Deployment: Unless compressed or hybridized (e.g., using Tiny-LeViT or PMVT), most transformer models are unsuitable for edge devices or low-power situations due to their huge parameter sizes and longer inference times.

•   Ongoing Research: Current research aims to reduce computing costs and maintain accuracy by making transformers lighter and faster using token pruning, sparse attention, or hybrid CNN-transformer designs.

5.7 Model Deployment Strategies for Real-World Agricultural Applications

The practical implementation in actual agricultural settings presents a distinct set of challenges, even though model creation and accuracy metrics are essential to advancing research. High-performing models are necessary for successful deployment, but so is careful integration with the hardware and infrastructure limitations of agricultural settings. The adaptation of various model designs for deployment across edge devices, mobile-based applications, and centralized cloud platforms is examined in this section. We aim to bridge the gap between theoretical performance and real-world field utility by evaluating deployment feasibility, benefits, and constraints. This will provide recommendations for the practical implementation of AI-driven systems for monitoring plant diseases and pests.

Centralized Deployment of Cloud Services:

How it is used: Models hosted on cloud servers (e.g., AWS, Google Cloud, Azure) are accessible via APIs or mobile/web interfaces.

Benefits:

*high processing power for intricate models (such as Swin Transformer and HvT).

*Simple scalability and integration with analytics tools and big datasets.

*Ideal for large-scale disease surveillance and teaching.

Limitations:

*Requires continuous internet access, which may not always be available in rural areas.

*Higher latency makes it unsuitable for field settings requiring real-time inference.  Large-scale operations may raise concerns about data privacy and transfer costs.

Edge Computing (Smart Cameras, Embedded Systems, UAVs):

How it is used: Devices such as the Jetson Nano, Raspberry Pi, or field-deployed UAVs are equipped with lightweight variants, including MobileNetV3 and Tiny-LeViT.

Benefits:

*Inference in real time without the need for connectivity.

*lowers the cost of data transfer by local processing.

Allows for precise spraying or targeted treatment.

Limitations:

*Model size and complexity are constrained by low memory and computing power.

*Requires model design that takes hardware into consideration (quantization, pruning).

*Rain, heat, and dust all have an impact on how well a gadget works.

Mobile Device-Based Implementation:

Deployment method: Models are transformed into formats like as ONNX or TensorFlow Lite and then made available through Android and iOS applications.

Benefits:

*Extremely reachable by field technicians and farmers. *With built-in cameras, it is both portable and easy to use.

*Uses preloaded models and operates offline.

Limitations:

*limited accuracy and speed of inference because of device limitations.

*For non-technical users, the design of the app must be intuitive.

*Results may vary depending on the lighting and camera quality.

5.8 Integration of Deep Learning Models into Agricultural Pest Management Systems

The usefulness of deep learning in agriculture is based on both model performance and the models’ ability to be integrated into current frameworks for managing pests and diseases. By connecting data-driven diagnostics with decision-making tools, integration provides farmers, agronomists, and policymakers with relevant information. This section examines the main routes for incorporating deep learning models into agricultural processes, emphasizing the advantages, practical limitations, and tactical factors for practical implementation.

Key Integration Pathways:

•   Field Scouting and Mobile Apps: Farmers can use trained models integrated into mobile apps to help them with regular field inspections. For example, Android apps have utilised MobileNet-based models to identify tomato diseases in real-time.

•   Relation to Decision Support Systems (DSS): Model results can be fed into DSS platforms, which produce treatment suggestions according to environmental factors, crop stages, pest species, and symptoms that have been identified.

•   Automated Monitoring Systems: Geo-tagging and time-stamped detections for spatial analytics enable continuous monitoring of big farms using drones, IoT-enabled cameras, and robotic vehicles.

•   Feedback Loop for Continuous Learning: Cloud systems integrate real-time detections to refine models, augment datasets, and generate region-specific detectors.

•   Agricultural Extension Services: AI-assisted technologies can be utilised by governments and advisory programs to enhance outreach and provide smallholder farmers with timely diagnoses.

Advantages of Integration:

•   Early Intervention: It is made possible by timely detection, thereby reducing yield loss and disease spread.

•   Precision Intervention: It reduces the environmental effect and usage of pesticides by enabling localized treatment.

•   Scalability: Adaptable to different farming scales, areas, and crops.

•   Empowerment: Improves smallholder farmers’ and experts’ ability to make decisions.

•   Data-Driven Insights: Enables centralized data aggregation for long-term monitoring and policy planning.

Limitations and Restrictions:

•   Connectivity and Infrastructure: The utilization of cloud-based functionality and real-time updates is restricted in many rural locations due to inadequate internet connectivity.

•   Hardware Restrictions: The processing power of smartphones and edge devices is limited, which limits the speed and complexity of the models.

•   Data Security and Privacy: Centralized systems that handle user data need to make sure that privacy laws are followed.

•   Model Generalization: When applied to unknown data from several locations or in various environmental circumstances, many models exhibit subpar performance.

•   User Adoption and Literacy: In low-resource environments, farmers may need training to use AI-powered products efficiently.

6  Recent Breakthroughs and Emerging Trends of Deep Learning in Plant Pest and Disease Detection

Recent developments have significantly enhanced the identification of plant pests and diseases using deep learning, yielding innovative solutions that improve precision, effectiveness, and scalability. Among the significant developments are:

•   AI-Assisted Disease Identification in Tea Farming: In tea farming, researchers have used sophisticated YOLO models, including YOLOv10s, to identify pests and diseases in tea leaves precisely [164]. This method improves yield quality and reduces crop loss by facilitating early intervention.

•   Frameworks for Multimodal Deep Learning: Integrating natural language processing with image processing improves pest detection accuracy. For example, tiny-BERT with R-CNN and ResNet-18 has been used to analyse textual and visual data [165]. This multimodal strategy overcomes the drawbacks of conventional CNN-based techniques.

•   Software for Hybrid Image Segmentation and Machine Learning: With the use of video footage, tools like “Plant Doctor” use algorithms such as YOLOv8, DeepSORT, and DeepLabV3Plus to diagnose urban street plants autonomously [166]. This technology tracks and identifies individual leaves, evaluating their health and identifying bacterial, fungal, and insect damage.

•   Drones with Deep Learning Capabilities: When convolutional neural networks (CNNs) are integrated with drones, the accuracy and efficiency of pest and disease identification in crops have increased [167]. This device supports remote operation and real-time data processing while quickly and accurately categorising crop pests and diseases.

These developments show how deep learning can revolutionise agricultural operations by providing scalable and effective methods for detecting plant diseases and pests.

7  Experimental Analysis

7.1 Methodology

We conducted an experimental analysis to demonstrate the relative effectiveness of well-known models in supplementing a thorough review of deep learning methods for plant pest and disease identification. This section describes the methodological framework used to assess both contemporary and conventional deep learning techniques using a standardized dataset.

7.1.1 Dataset Description

Since the PlantVillage dataset [34] is widely used and has benchmarking significance in agricultural image classification research, it was selected for this investigation. We concentrated on 10 tomato disease classes, comprising a total of 18,163 images for consistency and domain relevance, because they provide a challenging and varied subset for model evaluation. The detailed description of the Tomato dataset is presented in Table 12.

images

7.1.2 Model Selection and Experimental Design

The experimental study was carried out in two steps to guarantee a fair evaluation among model groups:

•   A Comparative Analysis of Traditional Machine Learning Models and Deep Learning: In the initial phase, we compared the lightweight and robust deep learning architectures, MobileNetV3 [63] and InceptionV3 [70], with traditional machine learning models, including Support Vector Machine (SVM) [45], and Random Forest (RF) [81]. This comparison shows how accuracy and generalization capabilities have advanced from traditional image classification techniques to contemporary deep learning models.

•   Assessment of Models Based on Vision Transformers: Recent transformer-based architectures, such as Swin Transformer [36], Hierarchical Vision Transformer (HvT) [37], Progressive Mobile Vision Transformer (PMVT) [38], Tiny-LeViT [39], and ViTAE [40] were evaluated in the second stage. These models were chosen due to their sophisticated feature extraction capabilities and growing dominance in vision tasks, particularly in complex agricultural settings.

A variety of models were chosen for this study to compare classical and deep learning approaches for plant disease diagnosis. The traditional baselines were Support Vector Machine (SVM) and Random Forest. While InceptionV3 successfully captures multi-scale leaf patterns, MobileNetV3 was selected due to its lightweight design and real-time capability, making it perfect for edge deployment. For deep spatial perception, transformer models such as HvT and Swin Transformer employ hierarchical attention. Mobile-friendly transformer substitutes include PMVT and Tiny-LeViT, whereas ViTAE enhances feature extraction by fusing transformer power with convolutional priors. This variety of models allows for thorough comparison in terms of precision, effectiveness, and suitability for field deployment.

Convolutional neural networks, vision transformer topologies, and classical machine learning methods were employed in this review, based on their contributions to image-based plant disease and pest identification. Models tested on actual agricultural datasets were given preference in the selection process, emphasising their viability for practical implementation. SVM, RF, MobileNetV3, InceptionV3, HvT, Swin Transformer, PMVT, Tiny-LeViT, and ViTAE were the models used to guarantee representation in baseline, lightweight, and cutting-edge categories. Key criteria, including accuracy, F1-score, parameter size, inference time, and suitability for edge or cloud deployment, were used to assess these models methodically. Other elements, such as plant growth stages, data imbalance, and robustness to lighting conditions, were also considered to provide a comprehensive assessment that takes into account both technical performance and practical usability in contemporary precision agricultural systems.

7.1.3 Preprocessing and Training Protocol

Every model underwent uniform training and evaluation. Standard preprocessing methods, including image scaling, normalization, and data augmentation, were used to reduce overfitting and enhance model generalizability. To ensure fairness, the training setup, including the batch size, learning rate, and number of epochs, was maintained constant across the experiments.

Evaluation Metrics:

To ensure a comprehensive performance evaluation, multiple metrics were considered, including:

•   Accuracy: The percentage of all correctly classified images.

•   Precision: The ratio of correctly predicted disease images to the total predicted as that disease. (True Positives/(True Positives + False Positives)).

•   Recall: The ratio of correctly predicted disease images to all actual instances of that disease. (True Positives/(True Positives + False Negatives)).

•   F1-score: The harmonic mean of precision and recall. F1-Score= 2 × (Precision × Recall)/(Precision + Recall).

•   Inference Time: As a measure of real-time deployment feasibility.

•   Resource Utilization: Model’s effective use of resources through automated feature extraction.

7.2 Experimental Analysis of Machine Learning and Deep Learning on PlantVillage Dataset

Accurate and efficient plant disease categorisation is crucial for minimising financial losses and guaranteeing food security. Early identification and efficient crop management depend on accurate and efficient plant disease categorisation.

The manual feature extraction used by traditional machine learning models restricts their capacity to adapt to intricate visual patterns. On the other hand, deep learning architectures are well-suited for image-based disease diagnosis, as they automatically construct hierarchical representations of images. The 10-class Tomato dataset from PlantVillage [34] comprises a total of 18,163 images; this experiment compares machine learning and deep learning models to assess their ability to diagnose plant diseases. For an all-encompassing assessment, the dataset is divided into 70% training, 20% validation, and 10% testing.

While deep learning models, such as MobileNetV3 [63] and InceptionV3 [70] (CNN architectures), used automatic feature extraction, machine learning models, such as Support Vector Machine (SVM) [45] and Random Forest (RF) [81], were trained on manually created feature representations. Based on our prior knowledge and domain expertise, the two models were fine-tuned using the following hyperparameters: a batch size of 32 and a learning rate of 0.00002. MobileNetV3 and InceptionV3 utilised an input image size of 160 × 160 pixels and were trained for 20 epochs using Adam and RMSProp optimizers throughout the fine-tuning process cess. The results show that deep learning models perform noticeably better than conventional machine learning techniques; MobileNetV3 and InceptionV3 achieved test accuracies of 98.1% and 93.8%, respectively. When it came to machine learning models, SVM [45] outperformed Random Forest (90.5% test accuracy), demonstrating its capacity to handle intricate feature spaces. Unlike machine learning models that rely on manually created features, CNN models may learn spatial hierarchies of features directly from images, which accounts for their superior accuracy. This investigation demonstrates that, in comparison to conventional machine learning strategies, deep learning-based CNNs are more effective in classifying plant diseases. Fig. 15 depicts the training, validation and test accuracy for the four models on the Tomato Leaf Dataset. Fig. 16 displays the accuracy and loss plots for the test dataset of MobileNetV3 and SVM, with MobileNetV3 achieving an accuracy of 98.1% and SVM achieving an accuracy of 90.5%.

images

Figure 15: The evaluation indicators of deep learning and machine learning models on the tomato leaf dataset

images

Figure 16: Test accuracy and loss curves for MobileNetV3 (98.1%) and SVM (90.5%) on the tomato leaf dataset

7.3 Experimental Analysis of Vision Transformer Models on PlantVillage Dataset

In investigating the efficacy of Vision Transformers in image classification, a field relatively underexplored compared to traditional deep learning models—we conducted experiments using five state-of-the-art (SOTA) Vision Transformers such as HvT, PMVT, ViTAE, Tiny-LeViT and Swin Transformer, on the Tomato Leaf Diseases from the PlantVillage Dataset [34]. The dataset comprised ten classes with a total of 18,163 images, divided into training, validation, and test sets in the ratio of 0.7:0.2:0.1. To ensure a fair comparison with previous studies, the Tomato Leaf Diseases dataset [34] was left un-altered, as other researchers have widely used it to evaluate the efficacy of their proposed models. Our sole objective was to understand how Vision Transformers performs compared to the well-established benchmarks set by well-known deep learning architectures.

The HvT (hvt/hierarchical-vision-transformer-large) and PMVT (pmvt/plant-model-vision-transformer-base) processes images by dividing them into a 16 × 16 patch size, ViTAE (vitae/vit-base- aggregated-enhanced) uses a 14 × 14 patch size, and Tiny-LeViT (levit/tiny-levit-224) processes images with 32 × 32 patch sizes. Swin Transformer (microsoft/swin-base-patch4-window7-224) processes images by dividing them into 4 × 4 pixel patches. While larger patches may result in some loss of fine-grained details within a single patch, the self-attention mechanism compensates by capturing global relationships across patches and effectively recognising disease patterns. Fig. 17 depicts an abstract overview of ViT’s architecture. Combined with its hierarchical windowing and sliding attention mechanisms, the model can effectively detect small lesions and intricate features in plant disease images. These approaches utilise patch-based processing to strike a balance between local detail retention and global feature learning, thereby ensuring the accurate identification of disease-affected regions.

images

Figure 17: High-level overview of ViT’s architecture

Based on our prior knowledge and domain expertise, this experiment focused on classifying tomato diseases. It assessed the efficacy of five transformer-based models—HvT, PMVT, ViTAE, Tiny-LeViT, and Swin Transformer—using the PlantVillage dataset. With an initial learning rate of 0.0001, we used the Adam Optimizer for training, which dynamically lowered when it reached a plateau. All models were run with a batch size of 32 to maximise memory economy and convergence speed. Specifically, HvT and PMVT achieved test accuracies of 99.3% and 99.1%, respectively; ViTAE, Tiny-LeViT, and Swin Transformer reached 98.9%, 97.8%, and 98.2%. The training, test, and validation accuracies for the five Transformer models on the Tomato Leaf Dataset are demonstrated in Fig. 18. Fig. 19 shows the accuracy and loss plots against the test dataset for HvT and PMVT, with both achieving a testing accuracy of 99.3%. Table 13 tabulates the five models’ training time and inference time for the above experiment.

images

Figure 18: The evaluation indicators of transformer models on the tomato leaf dataset

images

Figure 19: Testing Accuracy and Loss against the Tomato Leaf dataset for the HvT (99.3%) and PMVT (99.1%)

images

Fig. 20 displays the confusion matrix for the Hierarchical Vision Transformer (HvT) model, which was utilised with the Tomato Leaf Disease dataset. This matrix was created to offer more detailed information about the model’s performance across classes than just overall accuracy numbers. For each class—Healthy, Leaf Spot, and Early Blight—the confusion matrix displays the number of accurate and inaccurate predictions, enabling a thorough error analysis. With the HvT model accurately detecting 95 out of 100 healthy samples, 98 out of 100 leaf spot cases, and 97 out of 100 early blight occurrences, the results show excellent classification performance. Very few and largely misclassified cases between Healthy and Early Blight indicate that specific early symptoms may be mild or visually overlap.

images

Figure 20: Confusion matrix for HvT on tomato disease dataset

A comparison of the ROC (Receiver Operating Characteristic) curves for the tomato leaf disease classification task between the MobileNetV3 and Hierarchical Vision Transformer (HvT) models is shown in Fig. 21. This graphic provides a deeper perspective than simple accuracy measurements by visualizing how effectively each model differentiates across disease classifications across different thresholds. With a remarkable AUC (Area Under the Curve) of 0.994, the HvT model demonstrated excellent discriminative power, yielding few false positives or negatives. With an AUC of 0.978, the lightweight and effective MobileNetV3 model likewise demonstrates exceptional performance, showcasing its reliability in real-world, resource-constrained settings. HvT’s improved robustness and generalization are shown by its constantly higher curve, which makes it a good fit for high-stakes agricultural applications where precise and early disease identification is essential.

images

Figure 21: ROC curve: transformer vs. CNN

Our observations suggest that vision transformers and their variants, such as HvT, are superior to popular deep learning models due to their innovative architecture, which fundamentally differs from traditional CNNs. These transformers leverage self-attention mechanisms to capture long-range dependencies and global context within images, which CNNs handle effectively due to their local receptive fields. ViT operates by partitioning an image into patches, which undergo linear embedding and subsequent processing through multiple transformer layers. This method facilitates the capture of intricate local and global dependencies within the image. Leveraging self-attention mechanisms, ViT learns hierarchical representations, achieving competitive performance in image classification tasks and demonstrating superiority over conventional CNNs on diverse benchmarks. HvT effectively captures local and global features at various sizes by utilizing spatially reduced multi-head self-attention (SR-MHSA) and hierarchical structures. This combination of spatially reduced multi-head self-attention and hierarchical representation enables vision transformers to excel in image classification tasks, offering robust performance that matches or exceeds that of conventional CNN models.

7.4 Comparative Analysis Using Evaluation Indicators

Several assessment criteria, including Accuracy, Precision, Recall, and F1-Score, were utilised to thoroughly evaluate the efficacy of different models used in plant disease detection. These metrics comprehensively assess each model’s performance in terms of accurate predictions, dependability, and robustness, particularly when managing complex visual elements and class imbalances.

Out of all the models that were assessed, HvT performed the best on all criteria, obtaining an F1-score and accuracy of 99.3%, which indicates greater detection capabilities. HvT was quickly followed by PMVT and ViTAE, which likewise showed near-optimal values. In the CNN group, MobileNetV3 outperformed InceptionV3, indicating its efficient yet lightweight design. Traditional machine learning techniques, such as SVM and Random Forest, on the other hand, performed noticeably worse; Random Forest’s accuracy of 75.8% demonstrated the shortcomings of manual feature extraction and the absence of deep representational learning. Deep learning models outperform conventional machine learning techniques in all key evaluation metrics. Performance metrics of selected models are summarized in Table 14.

images

7.5 Resource Utilization & Computational Efficiency Analysis

A comprehensive assessment of the chosen models, based on their computational needs and compatibility with various hardware platforms, is presented in Table 15. The primary resource use measures were the number of parameters, model size, and CPU time required to infer a single image. With their compact model sizes (16 and 45 MB, respectively) and short inference times, MobileNetV3 and PMVT stand out among the studied models as particularly effective, making them ideal for mobile and edge devices. HvT and Swin Transformer, on the other hand, require powerful GPUs for real-time deployment due to their enormous parameter counts and model sizes (250 and 230 MB). Additionally, Tiny-LeViT exhibits potential for mobile deployment by striking a compromise between accuracy and small size. The resource requirements of InceptionV3 and ViTAE are moderate to high, making them suitable for desktop or cloud-based applications. Conventional machine learning models, such as SVM and Random Forest, have small memory footprints but are constrained by their reliance on human feature extraction, which increases the complexity and effort required for preprocessing. This comparison demonstrates that model selection must consider target deployment conditions, computational efficiency, and accuracy, particularly in resource-constrained contexts such as remote healthcare facilities or agricultural fields.

images

Lightweight models like Tiny-LeViT and MobileNetV3, with compact architectures and fast inference, are ideal for resource-constrained contexts. For example, MobileNetV3 maintained a modest computational footprint while achieving an astounding 98.1% accuracy on the Tomato Leaf Disease dataset with just 13 million parameters. Likewise, Tiny-LeViT, a hybrid transformer-CNN model with over 1.1 million parameters, achieved a 97.25% F1-score and a 9% better frame rate on edge devices. Larger transformer-based models, such as PMVT and HvT, on the other hand, had better classification accuracy (99.1% and 99.3%, respectively), but they needed more memory and longer training and inference times, which made them more appropriate for cloud or high-performance computing platforms than edge deployment. This comparison highlights that transformer-based models such as HvT and Swin Transformer provide better accuracy at the expense of larger computing needs, even though MobileNet and Tiny-LeViT are perfect for real-time mobile or embedded application.

7.6 Experimental Analysis of CNN and Vision Transformer Models in Terms of Visualization

Plant pest and disease detection relies heavily on visualization, which aids agronomists and researchers in interpreting model predictions, identifying essential characteristics, and increasing the accuracy of disease classification. CNN and Vision Transformer models identify several tomato infections by analyzing vein structure, lesions, color fluctuations, and leaf texture. Recognising the layers that contribute to visualisation improves the explainability of the model and builds confidence in AI-based disease detection. Comparison of attention mechanisms for CNN and Vision Transformer is shown in Table 16.

images

In this experiment, we use layer-wise visualisation approaches, such as Grad-CAM, Saliency Maps, and Attention Heatmaps, to investigate MobileNetV3, InceptionV3, Hierarchical Vision Transformer (HvT), and Patch-based Multi-view Transformer (PMVT) on the 10-class tomato disease dataset of PlantVillage [34]. The Visualization of disease and pest-affected parts using Attention Map (patches) & Heat Map (red) on the Tomato dataset presented in Fig. 22, where heat maps highlight the severity of the disease by employing red intensity, whereas attention maps pinpoint important areas where the model focuses. This enhances the interpretability of the model and captures contextual dependencies, which aids in accurate diagnosis. Agronomists and farmers can make well-informed decisions and effectively manage plant health with the help of such visual insights. Comparison of Attention Mechanisms for CNNs and Vision Transformer models is shown in Table 16. Additionally, the visualization-related layers and their function of each model used in this experiment for tomato disease detection are presented in Table 17.

images

Figure 22: The visualization of disease and pest affected parts using attention map (patches) & heat map (red) on tomato dataset

images

From the experiment, we got that:

•   Transformer models (HvT & PMVT) are useful for complex diseases because they have greater global context awareness.

•   Although MobileNetV3 is lightweight, it is less effective in capturing fine-grained information.

•   Despite having trouble with class overlap, InceptionV3 performs well by utilizing multi-scale feature extraction.

•   Grad-CAM and attention map visualizations validate the role of deep feature layers (Inception and Transformer blocks) in localizing diseased regions.

The interpretability of plant disease categorization is improved by visualization approaches such as Grad-CAM and attention heatmaps, however they have practical limits when used in actual agricultural settings. They may be challenging for non-technical individuals to understand and often rely on high-quality graphics. Furthermore, these visualizations can only be used on edge devices due to the additional computing resources needed to generate them. Future systems should incorporate voice-based instruction, lightweight modules tailored for low-resource contexts, and simplified overlays to improve usability and accessibility. Robustness can also be increased by training models using noisy, real-world images. Effective, farmer-friendly AI-adoption requires striking a balance between explainability and simplicity.

Our investigation shows that Vision Transformers outperform CNNs in terms of visualization efficacy, making them ideal for practical plant disease detection applications.

According to the performance analysis in Table 18, Vision Transformers achieve the highest test accuracy due to their fine-grained feature extraction and strong contextual comprehension.

images

7.7 Environmental Impact of Deep Learning in Agriculture

It is crucial to consider both the advantages and the environmental impacts of deep learning technologies as they are increasingly used in agriculture. Although these models provide practical tools for yield optimization and disease detection, their development and application may have a significant environmental impact. This section examines the sustainability issues surrounding the application of deep learning in agricultural settings.

•   High computational demand: Deploying deep learning models on a large scale, particularly with transformer-based architectures, can be energy-intensive during the training and inference stages.

•   Carbon footprint concerns: In the absence of optimization, these models could lead to higher greenhouse gas emissions, which would be in opposition to agricultural sustainability objectives.

•   Efficiency is essential: Lightweight models, such as Tiny-LeViT or MobileNet, are better suited for locations with limited energy resources and better balance environmental effects and performance.

•   The benefits of edge computing: It includes deploying models on edge devices (such as smartphones, drones, and Internet of Things sensors) to allow real-time decision-making in the field and lessen reliance on power-hungry centralised servers.

•   Green AI practices: Strategies like knowledge distillation, model pruning, quantization, and the use of renewable energy sources can considerably decrease the environmental impact.

•   Research direction: For scalable, environmentally friendly agricultural applications, future research should prioritize sustainable AI architectures and take energy-aware benchmarking into account.

7.8 Ethical Considerations in Adopting Deep Learning in Agriculture

Deep learning is becoming a vital component of digital agriculture; therefore, it is essential to consider the ethical implications of its use. These technologies raise questions about data privacy, equitable access, inclusion, and model transparency even as they promise increased productivity and disease management. In particular, in diverse and resource-constrained farming situations, addressing these issues is crucial to guaranteeing that deep learning systems are not only efficient but also equitable, reliable, and available to all stakeholders. Some essential ethical factors are discussed here:

Data Privacy and Ownership:

•   Sensitive farm image data, which can disclose crop health, location, and agricultural practices, is frequently used by deep learning systems.

•   Secure data management procedures, data anonymization, and informed farmer permission are necessary for ethical implementation.

•   Data ownership, usage rights, and benefit-sharing from model findings must all be clearly defined by policies.

The Digital Divide and Equitable Access:

•   In areas with inadequate internet access or limited device availability, the usefulness of AI tools can be restricted.

•   Models for low-end devices with offline capabilities and localized interfaces should be created in order to advance equity.

•   Governments and non-profit organizations are examples of stakeholders that can make access easier by providing smallholder farmers with training and tool subsidies.

Inclusivity and Bias:

•   In underrepresented regions or crop kinds, models trained on limited datasets may perform poorly, increasing the risk of inaccurate information or exclusion.

•   Diverse, representative training data and field validation in a range of agroclimatic situations are essential for ethical deployment.

Explainability and Transparency:

•   Models must produce interpretable outputs for AI-based judgments to be reliable and actionable.

•   Particularly in crucial situations involving the use of chemicals or the containment of disease, methods like saliency maps or attention visualization should be used to demonstrate which aspects of the image influenced the model’s decision.

7.9 Enhancing Accessibility and Human-AI Collaboration in Agriculture

Accessibility and inclusivity must coexist with technological innovation if deep learning solutions are to genuinely help smallholder farmers. The adoption of AI-driven solutions in agriculture may be hampered by the linguistic, educational, and digital literacy challenges that many users encounter. Collaboration between humans and AI, user-friendly interfaces, and support for regional languages are essential to address these challenges. To ensure that technology innovation has a practical impact on various farming communities, this section explores methods for making AI technologies more accessible and farmer-friendly.

Smallholder farmers, especially those in developing nations, suffer from linguistic, cultural, and educational obstacles that must be considered to optimize the impact of AI-driven plant disease detection in agriculture. Many farmers may speak regional dialects that are not supported by conventional interfaces, have low literacy rates, or lack technical expertise. Some of the possible solutions discussed here:

Cross-Linguistic Support:

•   Multilingual interfaces using regional scripts (such as Hindi, Bengali, and Swahili) must be incorporated into mobile apps for accessibility.

•   Text-to-speech (TTS) capabilities in regional languages guide users through disease identification and treatment without requiring literacy.

Simplified User Interfaces:

•   Interfaces should use voice prompts, image-based navigation, and iconography rather than menus with a lot of text.

•   Color-coded signs, such as green for healthy and red for unhealthy, are examples of visual cues that can make decision-making easier.

•   As seen in PlantVillage Nuru, designing for low-end cellphones with offline capability guarantees inclusivity across different levels of tech access.

Human-in-the-Loop Design:

•   Artificial intelligence should complement human expertise, not replace it. Collaborative workflows may include:

∘   Farmers taking images and getting comments.

∘   AI outputs are verified and interpreted by local agricultural extension officers.

•   Systems can incorporate feedback loops where users validate diagnoses to help models learn and adjust to local circumstances.

Community-Based Deployment Strategies:

•   Adoption in rural communities can be accelerated by teaching local agri-volunteers or “AI ambassadors” how to use and explain AI tools.

•   Farmer-involved co-design workshops yield culturally appropriate solutions and increase technology adoption.

8  Conclusion and Research Directions

This review comprehensively analyzed over 150 papers on deep learning-based techniques for identifying plant pests and diseases. We offered a thorough overview of state-of-the-art advancements by structuring the discussion into focused sections, including hyperspectral imaging, visualisation and non-visualisation methods, modified architectures, and transformer models. This framework clarified each approach’s strengths, limitations, and gaps in the literature, particularly emphasizing the importance of stage-specific pest and disease detection strategies.

Our review of existing literature made it imperative to benchmark recent advancements against established architectures in plant disease and pest detection. Our experiments demonstrated that transformers (HvT) in computer vision achieved accuracy of 99.3%, as illustrated in Fig. 18. By categorising the study systematically and including the latest papers, we ensure that our review reflects the most current advancements and trends. Our observations offer valuable insights for researchers and practitioners, guiding them in selecting the most suitable techniques for their specific needs and paving the way for future advancements in the field.

To facilitate deployment on low-power edge devices like UAVS, smartphones, and IoT sensors, future research should enhance model efficiency through hardware-aware design, pruning, quantization, and lightweight architectures. Furthermore, developing reliable models trained on heterogeneous, multi-location datasets is necessary to increase generalizability across various agricultural settings, including temperatures, soil types, and plant growth phases. There is potential for improving adaptability and prediction accuracy through domain adaptation, few-shot learning, and multimodal data fusion, such as merging RGB, hyperspectral, and environmental sensor data. Finally, adding explainability mechanisms and developing open-source, scalable frameworks would help end users’ trust and real-world adoption.

Our primary objective for this survey is to encourage more researchers to utilise deep learning-based applications in addressing agricultural problems related to plant pests and disease detection. Deep learning algorithms would produce more intelligent, sustainable, and reliable agrarian output.

Acknowledgement: The authors would like to thank all contributors and institutions that supported this research.

Funding Statement: This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

Author Contributions: MD Tausif Mallick: Led the conception and design of the study, conducted the majority of the literature review, implemented model evaluations, and prepared the initial manuscript draft. He played the primary role in comparing state-of-the-art methods and in interpreting experimental results. Saptarshi Banerjee: Contributed to data analysis, comparative studies of transformer-based models, and supported the survey framework design. He also assisted in model accuracy benchmarking and refining the manuscript’s structure. Nityananda Thakur: Focused on statistical evaluation methods and performance metrics for pest and disease detection. He also contributed to organizing and structuring the taxonomy of the techniques. Himadri Nath Saha: Provided overall supervision and critical guidance throughout the manuscript development. He also reviewed and edited all sections for clarity, consistency, and academic rigor. Amlan Chakrabarti: Contributed to the refinement of deep learning methodology discussions and ensured the inclusion of the latest technological insights. He also provided domain-level validation and cross-reviewed key sections of the manuscript. All authors reviewed the results and approved the final version of the manuscript.

Availability of Data and Materials: All data and materials used in this study are available from the corresponding author upon reasonable request.

Ethics Approval: This study did not involve human or animal participants and therefore did not require ethical approval.

Conflicts of Interest: The authors declare no conflicts of interest to report regarding the present study.

References

1. Food and Agriculture Organization (FAO). Declaration of the World Summit on Food Security. Rome, Italy: Food and Agriculture Organization; 2009 [Internet]. [cited 2025 Jun 6]. Available from: https://www.fao.org/wsfs/world-summit/en/. [Google Scholar]

2. FAO. Plant pests and diseases: a growing threat to global food security. Rome, Italy: Food and Agriculture Organization of the United Nations. [Internet]. [cited 2025 Jun 23]. Available from: https://www.fao.org/news/story/en/item/1187738. [Google Scholar]

3. Oerke EC. Crop losses to pests. J Agricult Sci. 2006;144(1):31–43. doi:10.1017/S0021859605005708. [Google Scholar] [CrossRef]

4. Strange RN, Scott PR. Plant disease: a threat to global food security. Annu Rev Phytopathol. 2005;43(1):83–116. doi:10.1146/annurev.phyto.43.113004.133839. [Google Scholar] [PubMed] [CrossRef]

5. Zwane ZR, Gubba A, Jooste AEC. The effect of avocado sunblotch disease (ASBD) on tree morphology, fruit maturity, yield and quality of ‘Hass’ avocado in South Africa. Eur J Plant Pathol. 2023;165(3):823–38. doi:10.1007/s10658-023-02705-z. [Google Scholar] [CrossRef]

6. Jiang Z, van Zanten M, Sasidharan R. Mechanisms of plant acclimation to multiple abiotic stresses. Communicat Biol. 2025;8(1):655. doi:10.1038/s42003-025-08077-w. [Google Scholar] [PubMed] [CrossRef]

7. Nalam V, Louis J, Shah J. Plant defense against aphids, the pest extraordinaire. Plant Sci. 2019;279(333):96–107. doi:10.1016/j.plantsci.2018.04.027. [Google Scholar] [PubMed] [CrossRef]

8. Cheaib A, Killiny N. Photosynthesis responses to the infection with Plant Pathogens. Mol Plant Microbe Interact. 2025;38(1):9–29. doi:10.1094/MPMI-05-24-0052-CR. [Google Scholar] [PubMed] [CrossRef]

9. Bayer Crop Science. Wheat Rust Diseases [Internet]; 2024. [cited 2025 Jun 11]. Available from: https://www.cropscience.bayer.us/articles/cp/wheat-rust-diseases. [Google Scholar]

10. Rosenheim JA, Steinmann K, Langellotto GA, Zink AG. Estimating the impact of Lygus hesperus on cotton: the insect, plant, and human observer as sources of variability. Environ Entomol. 2006;35(5):1141–53. doi:10.1603/0046-225X(2006)35[1141:ETIOLH]2.0.CO;2. [Google Scholar] [CrossRef]

11. Godfrey LD, Goodell PB, Natwick ET, Haviland DR, Barlow VM. UC IPM pest management guidelines: cotton. Davis, CA, USA: University of California Division of Agriculture and Natural Resources [Internet]; 2013. [cited 2025 Jun 6]. Available from: http://www.ipm.ucdavis.edu/PMG/r114301611.html.Accessed17(2013). [Google Scholar]

12. Tyagi AC. Towards a second green revolution. Irrigat Drain. 2016;65(4):388–9. doi:10.1002/ird.1694. [Google Scholar] [CrossRef]

13. Gebbers R, Adamchuk VI. Precision agriculture and food security. Science. 2010;327(5967):828–31. doi:10.1126/science.1183899. [Google Scholar] [PubMed] [CrossRef]

14. Babu MP, Rao BS. Leaves recognition using back propagation neural network-advice for pest and disease control on crops. India Kisan. Net: Expert Advisory System, 607-626. [Internet]; 2007. [cited 2025 Jun 23]. Available from: https://studylib.net/doc/7578578/leaves-recognition-using-back-propagation-neural. [Google Scholar]

15. Camargo A, Smith JS. An image-processing based algorithm to automatically identify plant disease visual symptoms. Biosyst Eng. 2009;102(1):9–21. doi:10.1016/j.biosystemseng.2008.09.030. [Google Scholar] [CrossRef]

16. Hillnhuetter C, Mahlein AK. Early detection and localisation of sugar beet diseases: new approaches. Gesunde Pflanzen. 2008;60(4):143–9. doi:10.1007/s10343-008-0196-0. [Google Scholar] [CrossRef]

17. Sharma P, Sharma S. Paradigm shift in plant disease diagnostics: a journey from conventional diagnostics to nano-diagnostics. In: Current trends in plant disease diagnostics and management practices. Cham, Switzerland: Springer; 2016. p. 237–64. doi:10.1007/978-3-319-27312-9_11. [Google Scholar] [CrossRef]

18. Desai M, Jain AK, Jain NK, Jethwa K. Detection and classification of fruit disease: a review. Int Res J Eng Technol. 2016;3(3):727–9. [Google Scholar]

19. Rehman S, Attique Khan M, Alhaisoni M, Armghan A, Alenezi F, Alqahtani A, et al. Fruit leaf diseases classification: a hierarchical deep learning framework. Comput Mater Contin. 2023;75(1):1179–94. doi:10.32604/cmc.2023.035324. [Google Scholar] [CrossRef]

20. Khan MA, Lali MIU, Sharif M, Javed K, Aurangzeb K, Haider S, et al. An optimized method for segmentation and classification of apple diseases based on strong correlation and genetic algorithm based feature selection. IEEE Access. 2019;7(1):46261–77. doi:10.32604/cmc.2023.035324. [Google Scholar] [CrossRef]

21. Hassan SM, Maji AK. Plant disease identification using a novel convolutional neural network. IEEE Access. 2022;10:5390–401. doi:10.1109/ACCESS.2022.3141371. [Google Scholar] [CrossRef]

22. Kdnuggets [Internet]. [cited 2025 Jun 23]. Available from: https://www.kdnuggets.com/2019/05/machine-learning-agriculture-. [Google Scholar]

23. Jordan MI, Mitchell TM. Machine learning: trends, perspectives, and prospects. Science. 2015;349(6245):255–60. doi:10.1126/science.aaa8415. [Google Scholar] [PubMed] [CrossRef]

24. Mohanty SP, Hughes DP, Salathé M. Using deep learning for image-based plant disease detection. Front Plant Sci. 2016;7:1419. doi:10.3389/fpls.2016.01419. [Google Scholar] [PubMed] [CrossRef]

25. Sladojevic S, Arsenovic M, Anderla A, Culibrk D, Stefanovic D. Deep neural networks based recognition of plant diseases by leaf image classification. Comput Intell Neurosci. 2016;2016(6):1–11. doi:10.1155/2016/3289801. [Google Scholar] [PubMed] [CrossRef]

26. Dyrmann M, Karstoft H, Midtiby HS. Plant species classification using deep convolutional neural network. Biosyst Eng. 2016;151(6):72–80. doi:10.1016/j.biosystemseng.2016.08.014. [Google Scholar] [CrossRef]

27. Zhang Y, Zhang B, Shen C, Liu H, Huang J, Tian K, et al. Review of the field environmental sensing methods based on multi-sensor information fusion technology. Int J Agricult Biol Eng. 2024;17(2):1–13. doi:10.25165/j.ijabe.20241702.8596. [Google Scholar] [CrossRef]

28. Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst. 2012;25(6):1097–105. doi:10.1145/3065386. [Google Scholar] [CrossRef]

29. Rahim A, Khan S, Al-Ghamdi M, Alzahrani A. Architectural optimization of Inception-v3 for explainable plant disease diagnosis: a mathematical framework for efficient deep learning in precision agriculture. Expert Syst Appl. 2023;213(Part B):119284. doi:10.1016/j.eswa.2022.119284. [Google Scholar] [CrossRef]

30. Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z. Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2016 Jun 27–30; Las Vegas, NV, USA. p. 2818–26. doi:10.1109/CVPR.2016.308. [Google Scholar] [CrossRef]

31. Jaderberg M, Simonyan K, Zisserman A. Spatial transformer networks. Adv Neural Inf Process Syst. 2015;28:2017–25. doi:10.5555/2969033.2969174. [Google Scholar] [CrossRef]

32. He W, Chai Y. An empirical study on energy disaggregation via deep learning. Adv Intell Syst Res. 2016;133:338–42. doi:10.2991/aiie-16.2016.77. [Google Scholar] [CrossRef]

33. Huang A, Wu R. Deep learning for music. arXiv:1606.04930. 2016. doi:10.48550/arXiv.1606.04930. [Google Scholar] [CrossRef]

34. Kaggle. [Dataset]. [cited 2025 Jun 6]. Available from: https://www.kaggle.com/datasets/emmarex/plantdisease. [Google Scholar]

35. Dosovitskiy A. An image is worth 16x16 words: transformers for image recognition at scale. arXiv:2010.11929. 2020. doi:10.48550/arXiv.2010.11929. [Google Scholar] [CrossRef]

36. Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, et al. Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF International Conference on Computer Vision; 2021 Oct 10–17; Montreal, QC, Canada. p. 10012–22. doi:10.1109/ICCV48922.2021.00984. [Google Scholar] [CrossRef]

37. Pan Z, Zhuang B, Liu J, He H, Cai J. Scalable vision transformers with hierarchical pooling. In: Proceedings of the IEEE/cvf International Conference on Computer Vision, 2021 Oct 10–17; Montreal, QC, Canada. p. 377–86. doi:10.1109/ICCV48922.2021.00043. [Google Scholar] [CrossRef]

38. Li G, Wang Y, Zhao Q, Yuan P, Chang B. PMVT: a lightweight vision transformer for plant disease identification on mobile devices. Front Plant Sci. 2023;14:1256773. doi:10.3389/fpls.2023.1256773. [Google Scholar] [PubMed] [CrossRef]

39. Thai H-T, Le K-H, Nguyen NL-T. Towards sustainable agriculture: a lightweight hybrid model and cloud-based collection of datasets for efficient leaf disease detection. Future Gener Comput Syst. 2023;148(12):488–500. doi:10.1016/j.future.2023.06.016. [Google Scholar] [CrossRef]

40. Fan H, Cheng S, de Nazelle AJ, Arcucci R. ViTAE-SL: a vision transformer-based autoencoder and spatial interpolation learner for field reconstruction. Comput Phys Commun. 2025;308:109464. doi:10.1016/j.cpc.2025.109464. [Google Scholar] [CrossRef]

41. Advancing Technology for Humanity [Internet]. [cited 2025 Jun 6]. Available from: https://ieeexplore.ieee.org/Xplore/home.jsp. [Google Scholar]

42. ScienceDirect [Internet]. [cited 2025 Jun 6]. Available from: https://www.sciencedirect.com/. [Google Scholar]

43. Hinton GE, Osindero S, Teh Y. A fast learning algorithm for deep belief nets. Neural Computation. 2006;18(7):1527–54. doi:10.1162/neco.2006.18.7.1527. [Google Scholar] [PubMed] [CrossRef]

44. Patel R, Gupta S, Joshi A. Parameter-efficient plant disease diagnosis: a systematic evaluation of EfficientNet architectures for agricultural vision tasks. Comput Electron Agric. 2023;204:107521. doi:10.1016/j.compag.2023.107521. [Google Scholar] [CrossRef]

45. DeChant C, Wiesner-Hanks T, Chen S, Stewart EL, Yosinski J, Gore MA, et al. Automated identification of northern leaf blight-infected maize plants from field imagery using deep learning. Phytopathology. 2017;107(11):1426–32. doi:10.1094/PHYTO-02-17-0053-R. [Google Scholar] [CrossRef]

46. Lowe DG. Distinctive image features from scale-invariant keypoints. Int J Comput Vis. 2004;60(2):91–110. doi:10.1023/B:VISI.0000029664.99615.94. [Google Scholar] [CrossRef]

47. Dalal N, Triggs B. Histograms of oriented gradients for human detection. In: 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05); 2005 Jun 20–25; San Diego, CA, USA. p. 886–93. doi:10.1109/CVPR.2005.177. [Google Scholar] [CrossRef]

48. Ojala T, Pietikainen M, Harwood D. Performance evaluation of texture measures with classification based on Kullback discrimination of distributions. In: Proceedings of 12th International Conference on Pattern Recognition; 1994 Oct 9–13; Jerusalem, Israel. p. 582–5. doi:10.1109/ICPR.1994.576366. [Google Scholar] [CrossRef]

49. Cortes C. Support-vector networks. Mach Learn. 1995;20(3):273–97. doi:10.1007/BF00994018. [Google Scholar] [CrossRef]

50. Kai S, Liu Z, Hang S, Guo C. A research of maize disease image recognition of corn based on BP networks. In: 2011 3rd International Conference on Measuring Technology and Mechatronics Automation; 2011 Jan 6–7; Shangshai, China. p. 246–9. doi:10.1109/ICMTMA.2011.66. [Google Scholar] [CrossRef]

51. Hanson KM. Introduction to Bayesian image analysis. In: Medical Imaging 1993: Image Processing. vol. 1898. Bellingham, DC, USA: SPIE; 1993. p. 716–31. doi:10.1117/12.154577. [Google Scholar] [CrossRef]

52. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556. 2014. doi:10.48550/arXiv.1409.1556. [Google Scholar] [CrossRef]

53. Lee K, Wang T. PlantTransformer: cross-species pest recognition using self-attention mechanisms with minimal training data. IEEE Transact AgriFood Elect. 2023;1(3):210–25. doi:10.1109/TAFE.2023.3245678. [Google Scholar] [CrossRef]

54. Sibiya M, Sumbwanyambe M. A computational procedure for the recognition and classification of maize leaf diseases out of healthy leaves using convolutional neural networks. AgriEngineering. 2019;1(1):119–31. doi:10.3390/agriengineering1010010. [Google Scholar] [CrossRef]

55. LeCun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to document recognition. Proc IEEE. 1998;86(11):2278–324. doi:10.1109/5.726791. [Google Scholar] [CrossRef]

56. Sermanet P, Eigen D, Zhang X, Mathieu M, Fergus R, LeCun Y. Overfeat: integrated recognition, localization and detection using convolutional networks. arXiv:1312.6229. 2013. doi:10.48550/arXiv.1312.6229. [Google Scholar] [CrossRef]

57. Zeiler MD, Fergus R. Visualizing and understanding convolutional networks. In: Computer Vision-ECCV 2014: 13th European Conference; 2014 Sep 6–12; Zurich, Switzerland. p. 818–33. doi:10.1007/978-3-319-10590-1_53. [Google Scholar] [CrossRef]

58. Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, et al. Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2015 Jun 7–12; Boston, MA, USA. p. 1–9. doi:10.1109/CVPR.2015.7298594. [Google Scholar] [CrossRef]

59. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2016 Jun 27–30; Las Vegas, NV, USA. p. 770–8. doi:10.1109/CVPR.2016.90. [Google Scholar] [CrossRef]

60. Huang G, Liu Z, Van Der Maaten L, Weinberger KQ. Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2017 Jul 21–26; Honolulu, HI, USA. p. 4700–8. doi:10.1109/CVPR.2017.243. [Google Scholar] [CrossRef]

61. Koonce B, Koonce B. “SqueezeNet.” Convolutional neural networks with swift for tensorflow: image recognition and dataset categorization. Berkeley, CA, USA: Apress; 2021. p. 73–85. doi:10.1007/978-1-4842-6168-2. [Google Scholar] [CrossRef]

62. Chollet F. Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2017 Jul 21–26; Honolulu, HI, USA. p. 1251–8. doi:10.1109/CVPR.2017.195. [Google Scholar] [CrossRef]

63. Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, et al. Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv:1704.04861. 2017. doi:10.48550/arXiv.1704.04861. [Google Scholar] [CrossRef]

64. Kamal KC, Yin Z, Wu M, Wu Z. Depthwise separable convolution architectures for plant disease classification. Comput Electron Agric. 2019;165:104948. doi:10.1109/ACCESS.2019.2914929. [Google Scholar] [CrossRef]

65. Jiang P, Chen Y, Liu B, He D, Liang C. Real-time detection of apple leaf diseases using deep learning approach based on improved convolutional neural networks. IEEE Access. 2019;7:59069–80. doi:10.1109/ACCESS.2019.2914929. [Google Scholar] [CrossRef]

66. Zhang K, Wu Q, Liu A, Meng X. Can deep learning identify tomato leaf disease? Adv Multimed. 2018. doi:10.1155/2018/6710862. [Google Scholar] [CrossRef]

67. Amara J, Bouaziz B, Algergawy A. A deep learning-based approach for banana leaf diseases classification. In: Datenbanksysteme für Business, Technologie und Web (BTW 2017)-Workshop Band [Internet]; 2017. [cited 2025 Jun 6]. Available from: https://dl.gi.de/handle/20.500.12116/13766147. [Google Scholar]

68. Ferentinos KP. Deep learning models for plant disease detection and diagnosis. Comput Electron Agric. 2018;145(6):311–8. doi:10.1016/j.compag.2018.01.009. [Google Scholar] [CrossRef]

69. Türkoǧlu M, Hanbay D. Plant disease and pest detection using deep learning-based features. Turkish J Electr Eng Comput Sci. 2019;27(3):1636–51. doi:10.3906/elk-1808-20. [Google Scholar] [CrossRef]

70. Ramcharan A, Baranowski K, McCloskey P, Ahmed B, Legg J, Hughes DP. Deep learning for image-based cassava disease detection. Front Plant Sci. 2017;8:1852. doi:10.3389/fpls.2017.01852. [Google Scholar] [PubMed] [CrossRef]

71. Fujita E, Kawasaki Y, Uga H, Kagiwada S, Iyatomi H. Basic investigation on a robust and practical plant diagnostic system. In: 2016 15th IEEE International Conference on Machine Learning and Applications (ICMLA); 2016 Dec 18–20; Anaheim, CA, USA. p. 989–92. doi:10.1109/ICMLA.2016.0169. [Google Scholar] [CrossRef]

72. Sholihati RA, Sulistijono IA, Risnumawan A, Kusumawati E. Potato leaf disease classification using deep learning approach. In: 2020 International Electronics Symposium (IES); 2020 Sep 29–30; Surabaya, Indonesia. p. 392–7. doi:10.1109/IES49823.2020.9230497. [Google Scholar] [CrossRef]

73. Sardogan M, Tuncer A, Ozen Y. Plant leaf disease detection and classification based on CNN with LVQ algorithm, Bosnia and Herzegovina. In: 2018 3rd International Conference on Computer Science and Engineering (UBMK); 2018 Sep 20–23; Bosnia and Herzegovina. p. 382–5. doi:10.1109/UBMK.2018.8529182. [Google Scholar] [CrossRef]

74. Too EC, Li Y, Njuki S, Liu Y. A comparative study of fine-tuning deep learning models for plant disease identification. Comput Electron Agric. 2019;161(2):272–9. doi:10.1016/j.compag.2019.04.023. [Google Scholar] [CrossRef]

75. Rangarajan AK, Purushothaman R, Ramesh A. Tomato crop disease classification using pre-trained deep learning algorithm. Procedia Comput Sci. 2018;133(12):1040–7. doi:10.1016/j.procs.2018.07.127. [Google Scholar] [CrossRef]

76. Brahimi M, Arsenovic M, Laraba S, Sladojevic S, Boukhalfa K, Moussaoui A. Deep learning for plant diseases: detection and saliency map visualisation. In: Human and machine learning. Cham, Switzerland: Springer; 2018. p. 93–117. doi:10.1007/978-3-319-90403-0_6. [Google Scholar] [CrossRef]

77. Brahimi M, Boukhalfa K, Moussaoui A. Deep learning for tomato diseases: classification and symptoms visualization. Appl Artif Intell. 2017;31(4):299–315. doi:10.1080/08839514.2017.1315512. [Google Scholar] [CrossRef]

78. Wang B, Zhang C, Li Y, Cao C, Huang D, Gong Y. An ultra-lightweight efficient network for image-based plant disease and pest infection detection. Precis Agric. 2023;24(5):1836–61. doi:10.1007/s11119-022-09939-1. [Google Scholar] [CrossRef]

79. Hong H, Lin J, Huang F. Tomato disease detection and classification by deep learning. In: 2020 International Conference on Big Data, Artificial Intelligence and Internet of Things Engineering (ICBAIE); 2020 Jun 12–14; Fuzhou, China. p. 25–9. doi:10.1109/ICBAIE49401.2020.9160136. [Google Scholar] [CrossRef]

80. Cruz AC, Luvisi A, De Bellis L, Ampatzidis Y. Vision-based plant disease detection system using transfer and deep learning. In: 2017 Asabe Annual International Meeting; 2017 Jul 16–19; Spokane, DC, USA. 1700241 p. doi:10.13031/aim.201700241. [Google Scholar] [CrossRef]

81. Ma J, Du K, Zheng F, Zhang L, Gong Z, Sun Z. A recognition method for cucumber diseases using leaf symptom images based on deep convolutional neural network. Comput Electron Agric. 2018;154(4):18–24. doi:10.1016/j.compag.2018.08.013. [Google Scholar] [CrossRef]

82. Chen JW, Lin WJ, Cheng HJ, Hung CL, Lin CY, Chen SP. A smartphone-based application for scale pest detection using multiple-object detection methods. Electronics. 2021;10(4):372. doi:10.3390/electronics10040372. [Google Scholar] [CrossRef]

83. Brahimi M, Mahmoudi S, Boukhalfa K, Moussaoui A. Deep interpretable architecture for plant diseases classification. In: 2019 Signal Processing: Algorithms, Architectures, Arrangements, and Applications (SPA); 2019 Sep 18–20; Poznan, Poland. p. 111–6. doi:10.1109/SPA.2019.8905831. [Google Scholar] [CrossRef]

84. Fuentes A, Yoon S, Kim SC, Park DS. A robust deep-learning-based detector for real-time tomato plant diseases and pests recognition. Sensors. 2017;17(9):2022. doi:10.3390/s17092022. [Google Scholar] [PubMed] [CrossRef]

85. Selvaraj MG, Vergara A, Ruiz H, Safari N, Elayabalan S, Ocimati W, et al. AI-powered banana diseases and pest detection. Plant Methods. 2019;15(1):1–11. doi:10.1186/s13007-019-0485-6. [Google Scholar] [CrossRef]

86. Bari BS, Islam MN, Rashid M, Hasan MJ, Razman MAM, Musa RM, et al. A real-time approach of diagnosing rice leaf disease using deep learning-based faster R-CNN framework. PeerJ Comput Sci. 2021;7(1):e432. doi:10.7717/peerj-cs.432. [Google Scholar] [PubMed] [CrossRef]

87. Lu J, Hu J, Zhao G, Mei F, Zhang C. An in-field automatic wheat disease diagnosis system. Comput Electron Agric. 2017;142(3):369–79. doi:10.1016/j.compag.2017.09.012. [Google Scholar] [CrossRef]

88. Wallelign S, Polceanu M, Buche C. Soybean plant disease identification using convolutional neural network. In: The Thirty-First International Flairs Conference; 2018 May 21–23; Melbourne, FL, USA. p. 146–51. [Google Scholar]

89. Ha JG, Moon H, Kwak JT, Hassan SI, Dang M, Lee ON, et al. Deep convolutional neural network for classifying Fusarium wilt of radish from unmanned aerial vehicles. J Appl Remote Sens. 2017;11(4):042621. doi:10.1117/1.JRS.11.042621. [Google Scholar] [CrossRef]

90. Barbedo JGA. Plant disease identification from individual lesions and spots using deep learning. Biosyst Eng. 2019;180(1):96–107. doi:10.1016/j.biosystemseng.2019.02.002. [Google Scholar] [CrossRef]

91. Esgario JG, de Castro PB, Tassis LM, Krohling RA. An app to assist farmers in the identification of diseases and pests of coffee leaves using deep learning. Informat Process Agricult. 2022;9(1):38–47. doi:10.1016/j.inpa.2021.01.004. [Google Scholar] [CrossRef]

92. Prasath B, Akila M. IoT-based pest detection and classification using deep features with enhanced deep learning strategies. Eng Appl Artif Intell. 2023;121(3):105985. doi:10.1016/j.engappai.2023.105985. [Google Scholar] [CrossRef]

93. Rustia DJA, Chao JJ, Chung JY, Lin TT. An online unsupervised deep learning approach for an automated pest insect monitoring system. In: 2019 ASABE Annual International Meeting; 2019 Jul 7–10; Boston, MA, USA. doi:10.13031/aim.201900477. [Google Scholar] [CrossRef]

94. Lu Y, Yi S, Zeng N, Liu Y, Zhang Y. Identification of rice diseases using deep convolutional neural networks. Neurocomputing. 2017;267(1):378–84. doi:10.1016/j.neucom.2017.06.023. [Google Scholar] [CrossRef]

95. Johannes A, Picon A, Alvarez-Gila A, Echazarra J, Rodriguez-Vaamonde S, Navajas AD, et al. Automatic plant disease diagnosis using mobile capture devices, applied on a wheat use case. Comput Electron Agric. 2017;138(11):200–9. doi:10.1016/j.compag.2017.04.013. [Google Scholar] [CrossRef]

96. Thenmozhi K, Reddy US. Crop pest classification based on deep convolutional neural network and transfer learning. Comput Electron Agric. 2019;164(4):104906. doi:10.1016/j.compag.2019.104906. [Google Scholar] [CrossRef]

97. Liu L, Wang R, Xie C, Yang P, Wang F, Sudirman S, et al. PestNet: an end-to-end deep learning approach for large-scale multi-class pest detection and classification. IEEE Access. 2019;7:45301–12. doi:10.1109/ACCESS.2019.2901517. [Google Scholar] [CrossRef]

98. Li W, Chen P, Wang B, Xie C. Automatic localization and count of agricultural crop pests based on an improved deep learning pipeline. Sci Rep. 2019;9(1):1–11. doi:10.1038/s41598-019-51062-5. [Google Scholar] [CrossRef]

99. Khan MA, Akram T, Sharif M, Awais M, Javed K, Ali H, et al. CCDF: automatic system for segmentation and recognition of fruit crops diseases based on correlation coefficient and deep CNN features. Comput Electron Agric. 2018;155(1):220–36. doi:10.1016/j.compag.2018.10.021. [Google Scholar] [CrossRef]

100. Xie X, Ma Y, Liu B, He J, Li S, Wang H. A deep-learning-based real-time detector for grape leaf diseases using improved convolutional neural networks. Front Plant Sci. 2020;11:751. doi:10.3389/fpls.2020.00751. [Google Scholar] [PubMed] [CrossRef]

101. Kerkech M, Hafiane A, Canals R. Deep leaning approach with colorimetric spaces and vegetation indices for vine diseases detection in UAV images. Comput Electron Agric. 2018;155(4):237–43. doi:10.1016/j.compag.2018.10.016. [Google Scholar] [CrossRef]

102. Zhang S, Zhang S, Zhang C, Wang X, Shi Y. Cucumber leaf disease identification with global pooling dilated convolutional neural network. Comput Electron Agric. 2019;162(1–2):422–30. doi:10.1016/j.compag.2019.05.030. [Google Scholar] [CrossRef]

103. Chowdhury ME, Rahman T, Khandakar A, Ayari MA, Khan AU, Khan MS, et al. Automatic and reliable leaf disease detection using deep learning techniques. AgriEngineering. 2021;3(2):294–312. doi:10.3390/agriengineering3020019. [Google Scholar] [CrossRef]

104. Gonzalez-Huitron V, León-Borges JA, Rodriguez-Mata AE, Amabilis-Sosa LE, Ramírez-Pereda B, Rodriguez H. Disease detection in tomato leaves via CNN with lightweight architectures implemented in Raspberry Pi 4. Comput Electron Agric. 2021;181(7):105951. doi:10.1016/j.compag.2020.105951. [Google Scholar] [CrossRef]

105. Zhang X, Qiao Y, Meng F, Fan C, Zhang M. Identification of maize leaf diseases using improved deep convolutional neural networks. IEEE Access. 2018;6:30370–7. doi:10.1109/ACCESS.2018.2844087. [Google Scholar] [CrossRef]

106. Chao X, Sun G, Zhao H, Li M, He D. Identification of apple tree leaf diseases based on deep learning models. Symmetry. 2020;12(7):1065. doi:10.3390/sym12071065. [Google Scholar] [CrossRef]

107. Chen YS, Hsu CS, Lo CL. An entire-and-partial feature transfer learning approach for detecting the frequency of pest occurrence. IEEE Access. 2020;8:92490–502. doi:10.1109/ACCESS.2020.2995311. [Google Scholar] [CrossRef]

108. Tetila EC, Machado BB, Astolfi G, de Souza Belete NA, Amorim WP, Roel AR, et al. Detection and classificsation of soybean pests using deep learning with UAV images. Comput Electron Agric. 2020;179(11):105836. doi:10.1016/j.compag.2020.105836. [Google Scholar] [CrossRef]

109. Haridasan A, Thomas J, Raj ED. Deep learning system for paddy plant disease detection and classification. Environ Monit Assess. 2023;195(1):120. doi:10.1007/s10661-022-10989-2. [Google Scholar] [CrossRef]

110. Abbas S, Attique Khan M, Alhaisoni M, Tariq U, Armghan A, Alenezi F, et al. Crops leaf diseases recognition: a framework of optimum deep learning features. Comput Mater Contin. 2023;74(1):1139–59. doi:10.32604/cmc.2023.028824. [Google Scholar] [CrossRef]

111. Khattak A, Asghar MU, Batool U, Asghar MZ, Ullah H, Al-Rakhami M, et al. Automatic detection of citrus fruit and leaves diseases using deep neural network model. IEEE Access. 2021;9:112942–54. doi:10.1109/ACCESS.2021.3096895. [Google Scholar] [CrossRef]

112. Chen J, Liu Q, Gao L. Visual tea leaf disease recognition using a convolutional neural network model. Symmetry. 2019;11(3):343. doi:10.3390/sym11030343. [Google Scholar] [CrossRef]

113. Gandhi R, Nimbalkar S, Yelamanchili N, Ponkshe S. Plant disease detection using CNNs and GANs as an augmentative approach. In: 2018 IEEE International Conference on Innovative Research and Development (ICIRD); 2018 May 11–12; Bangkok, Thailand. p. 1–5. [cited 2025 Jun 6]. Available from: https://ieeexplore.ieee.org/document/8467049. [Google Scholar]

114. Liu Y, Chen W, Wang Z, Zhang T. FPFR-CNN: an improved Faster R-CNN with feature pyramid fusion for citrus pest and disease detection. Comput Electron Agric. 2022;193(1):106692. doi:10.1016/j.compag.2022.106692. [Google Scholar] [CrossRef]

115. Amin H, Darwish A, Hassanien AE, Soliman M. End-to-end deep learning model for corn leaf disease classification. IEEE Access. 2022;10:31103–15. doi:10.1109/ACCESS.2022.3159678. [Google Scholar] [CrossRef]

116. Pacal I, Kunduracioglu I, Alma MH, Deveci M, Kadry S, Nedoma J, et al. A systematic review of deep learning techniques for plant diseases. Artif Intell Rev. 2024;57(11):304. doi:10.1007/s10462-024-10944-7. [Google Scholar] [CrossRef]

117. Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput. 1997;9(8):1735–80. doi:10.1162/neco.1997.9.8.1735. [Google Scholar] [PubMed] [CrossRef]

118. Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S. End-to-end object detection with transformers. In: European Conference on Computer Vision. Cham, Switzerland: Springer International Publishing; 2020. p. 213–29. doi:10.1007/978-3-030-58452-8_13. [Google Scholar] [CrossRef]

119. Zhu X, Su W, Lu L, Li B, Wang X, Dai J. Deformable detr: deformable transformers for end-to-end object detection. arXiv:2010.04159. 2020. [Google Scholar]

120. Mallick MT, Biswas S, Das AK, Saha HN, Chakrabarti A, Deb N. Deep learning based automated disease detection and pest classification in Indian mung bean. Multimed Tools Appl. 2023;82(8):12017–41. doi:10.1007/s11042-022-13673-7. [Google Scholar] [CrossRef]

121. Touvron H, Cord M, Douze M, Massa F, Sablayrolles A, Jégou H. Training data-efficient image transformers & distillation through attention. In: International Conference on Machine Learning; 2021 Jul 18–24; Online; 2021. p. 10347–57. doi:10.48550/arXiv.2012.12877. [Google Scholar] [CrossRef]

122. Lu X, Yang R, Zhou J, Jiao J, Liu F, Liu Y, et al. A hybrid model of ghost-convolution enlightened transformer for effective diagnosis of grape leaf disease and pest. J King Saud Univ-Comput Inf Sci. 2022;34(5):1755–67. doi:10.1016/j.jksuci.2021.09.019. [Google Scholar] [CrossRef]

123. Han K, Wang Y, Tian Q, Guo J, Xu C, Xu C. Ghostnet: more features from cheap operations. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2020 Jun 13–19. Seattle, WA, USA. p. 1580–9. doi:10.1109/CVPR42600.2020.00164. [Google Scholar] [CrossRef]

124. Howard A, Sandler M, Chu G, Chen LC, Chen B, Tan M, et al. Searching for mobilenetv3. In: Proceedings of the IEEE/CVF International Conference on Computer Vision; 2019 Oct 27–Nov 2; Seoul, Republic of Korea. p. 1314–24. doi:10.1109/ICCV.2019.00140. [Google Scholar] [CrossRef]

125. Wu H, Xiao B, Codella N, Liu M, Dai X, Yuan L, et al. CVT: introducing convolutions to vision transformers. In: Proceedings of the IEEE/CVF International Conference on Computer Vision; 2021 Oct 11–17; Montreal, QC, Canada. p. 22–31. doi:10.1109/ICCV48922.2021.00011. [Google Scholar] [CrossRef]

126. Li Y, Zhang K, Cao J, Timofte R, Van Gool L. Localvit: bringing locality to vision transformers. arXiv:2104.05707. 2021. doi:10.48550/arXiv.2104.05707. [Google Scholar] [CrossRef]

127. Guo J, Han K, Wu H, Tang Y, Chen X, Wang Y, et al. CMT: convolutional neural networks meet vision transformers. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2022 Jun 18–24; New Orleans, LA, USA. p. 12175–85. doi:10.1109/CVPR52688.2022.01186. [Google Scholar] [CrossRef]

128. Xiao T, Singh M, Mintun E, Darrell T, Dollár P, Girshick R. Early convolutions help transformers see better. Adv Neural Inf Process Syst. 2021;34:30392–400. doi:10.5555/3540261.3542586. [Google Scholar] [CrossRef]

129. Pacal I, Işık G. Utilizing convolutional neural networks and vision transformers for precise corn leaf disease identification. Neural Comput Appl. 2025;37(4):2479–96. doi:10.1007/s00521-024-10769-z. [Google Scholar] [CrossRef]

130. Veys C, Chatziavgerinos F, AlSuwaidi A, Hibbert J, Hansen M, Bernotas G, et al. Multispectral imaging for presymptomatic analysis of light leaf spot in oilseed rape. Plant Methods. 2019;15(1):1–12. doi:10.1186/s13007-019-0389-9. [Google Scholar] [PubMed] [CrossRef]

131. Mahlein AK, Alisaac E, Al Masri A, Behmann J, Dehne HW, Oerke EC. Comparison and combination of thermal, fluorescence, and hyperspectral imaging for monitoring fusarium head blight of wheat on spikelet scale. Sensors. 2019;19(10):2281. doi:10.3390/s19102281. [Google Scholar] [PubMed] [CrossRef]

132. Moghadam P, Ward D, Goan E, Jayawardena S, Sikka P, Hernandez E. Plant disease detection using hyperspectral imaging. In: 2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA); 2017 Nov 29–Dec 1; Sydney, NSW, Australia. p. 1–8. doi:10.1109/DICTA.2017.8227476. [Google Scholar] [CrossRef]

133. Xie C, Yang C, He Y. Hyperspectral imaging for classification of healthy and gray mold diseased tomato leaves with different infection severities. Comput Electron Agric. 2017;135(2):154–62. doi:10.1016/j.compag.2016.12.015. [Google Scholar] [CrossRef]

134. Shuaibu M, Lee WS, Schueller J, Gader P, Hong YK, Kim S. Unsupervised hyperspectral band selection for apple Marssonina blotch detection. Comput Electron Agric. 2018;148(2):45–53. doi:10.1016/j.compag.2017.09.038. [Google Scholar] [CrossRef]

135. Yuan L, Yan P, Han W, Huang Y, Wang B, Zhang J, et al. Detection of anthracnose in tea plants based on hyperspectral imaging. Comput Electron Agric. 2019;167:105039. doi:10.1016/j.compag.2019.105039. [Google Scholar] [CrossRef]

136. Jiang X, Zhen J, Miao J, Zhao D, Wang J, Jia S. Assessing mangrove leaf traits under different pest and disease severity with hyperspectral imaging spectroscopy. Ecol Indic. 2021;129(2):107901. doi:10.1016/j.ecolind.2021.107901. [Google Scholar] [CrossRef]

137. Hruška J, Adão T, Pádua L, Marques P, Cunha A, Peres E, et al. Machine learning classification methods in hyperspectral data processing for agricultural applications. In: Proceedings of the International Conference on Geoinformatics and Data Analysis; 2018 Apr 20–22; Prague, Czech Republic. p. 137–41. doi:10.1145/3220228.3220242. [Google Scholar] [CrossRef]

138. Abdulridha J, Batuman O, Ampatzidis Y. UAV-based remote sensing technique to detect citrus canker disease utilizing hyperspectral imaging and machine learning. Remote Sens. 2019;11(11):1373. doi:10.3390/rs11111373. [Google Scholar] [CrossRef]

139. Zhao X, Zhang J, Huang Y, Tian Y, Yuan L. Detection and discrimination of disease and insect stress of tea plants using hyperspectral imaging combined with wavelet analysis. Comput Electron Agric. 2022;193(11):106717. doi:10.1016/j.compag.2022.106717. [Google Scholar] [CrossRef]

140. Su J, Liu C, Coombes M, Hu X, Wang C, Xu X, et al. Wheat yellow rust monitoring by learning from multispectral UAV aerial imagery. Comput Electron Agric. 2018;155(6):157–66. doi:10.1016/j.compag.2018.10.017. [Google Scholar] [CrossRef]

141. Abdulridha J, Ampatzidis Y, Roberts P, Kakarla SC. Detecting powdery mildew disease in squash at different stages using UAV-based hyperspectral imaging and artificial intelligence. Biosyst Eng. 2020;197(11):135–48. doi:10.1016/j.biosystemseng.2020.07.001. [Google Scholar] [CrossRef]

142. Rumpf T, Mahlein AK, Steiner U, Oerke EC, Dehne HW, Plümer L. Early detection and classification of plant diseases with support vector machines based on hyperspectral reflectance. Comput Electron Agric. 2010;74(1):91–9. doi:10.1016/j.compag.2010.06.009. [Google Scholar] [CrossRef]

143. Javidan SM, Banakar A, Vakilian KA, Ampatzidis Y. Diagnosis of grape leaf diseases using automatic K-means clustering and machine learning. Smart Agricult Technol. 2023;3(17):100081. doi:10.1016/j.atech.2022.100081. [Google Scholar] [CrossRef]

144. Nagasubramanian K, Jones S, Singh AK, Singh A, Ganapathysubramanian B, Sarkar S. Explaining hyperspectral imaging based plant disease identification: 3D CNN and saliency maps. arXiv:1804.08831. 2018. doi:10.48550/arXiv.1804.08831. [Google Scholar] [CrossRef]

145. Yan T, Xu W, Lin J, Duan L, Gao P, Zhang C, et al. Combining multi-dimensional convolutional neural network (CNN) with visualization method for detection of aphis gossypii glover infection in cotton leaves using hyperspectral imaging. Front Plant Sci. 2021;12:74. doi:10.3389/fpls.2021.604510. [Google Scholar] [PubMed] [CrossRef]

146. Signoroni A, Savardi M, Baronio A, Benini S. Deep learning meets hyperspectral image analysis: a multidisciplinary review. J Imaging. 2019;5(5):52. doi:10.3390/jimaging5050052. [Google Scholar] [PubMed] [CrossRef]

147. Jin X, Jie L, Wang S, Qi HJ, Li SW. Classifying wheat hyperspectral pixels of healthy heads and Fusarium head blight disease using a deep neural network in the wild field. Remote Sens. 2018;10(3):395. doi:10.3390/rs10030395. [Google Scholar] [CrossRef]

148. Wang D, Vinson R, Holmes M, Seibel G, Bechar A, Nof S, et al. Early detection of tomato spotted wilt virus by hyperspectral imaging and outlier removal auxiliary classifier generative adversarial nets (OR-AC-GAN). Sci Rep. 2019;9(1):1–14. doi:10.1038/s41598-019-40066-y. [Google Scholar] [PubMed] [CrossRef]

149. Fuentes A, Im DH, Yoon S, Park DS. Spectral analysis of CNN for tomato disease identification. In: In International Conference on Artificial Intelligence and Soft Computing; 2017 Jun 11–15; Zakopane, Poland. p. 40–51. doi:10.1007/978-3-319-59063-9_4. [Google Scholar] [CrossRef]

150. Wang X, Zhang S, Huang J, Liu L. Hyperspectral imaging and deep learning: recent advances and future perspectives for agricultural applications. Comput Electron Agric. 2022;196(1):106869. doi:10.1038/s41598-019-40066-y. [Google Scholar] [CrossRef]

151. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative adversarial nets. Adv Neural Inform Process Syst. 2014;27. doi:10.1007/978-3-658-40442-0_9. [Google Scholar] [CrossRef]

152. Pu Y, Gan Z, Henao R, Yuan X, Li C, Stevens A, et al. Variational autoencoder for deep learning of images labels and captions. Adv Neural Inform Process Syst. 2016;29. doi:10.5555/3157096.3157360. [Google Scholar] [CrossRef]

153. Vallabhajosyula S, Sistla V, Kolli VKK. Transfer learning-based deep ensemble neural network for plant leaf disease detection. J Plant Dis Prot. 2022;129(3):545–58. doi:10.1007/s41348-021-00465-8. [Google Scholar] [CrossRef]

154. Yin H, Gu YH, Park CJ, Park JH, Yoo SJ. Transfer learning-based search model for hot pepper diseases and pests. Agriculture. 2020;10(10):439. doi:10.3390/agriculture10100439. [Google Scholar] [CrossRef]

155. Huang ML, Chuang TC, Liao YC. Application of transfer learning and image augmentation technology for tomato pest identification. Sustain Comput Inform Syst. 2022;33(4):100646. doi:10.1016/j.suscom.2021.100646. [Google Scholar] [CrossRef]

156. Coulibaly S, Kamsu-Foguem B, Kamissoko D, Traore D. Deep neural networks with transfer learning in millet crop images. Comput Ind. 2019;108:115–20. doi:10.1016/j.compind.2019.02.003. [Google Scholar] [CrossRef]

157. Liu J, Wang X. Plant diseases and pests detection based on deep learning: a review. Plant Methods. 2021;17(1):1–18. doi:10.1186/s13007-021-00722-9. [Google Scholar] [PubMed] [CrossRef]

158. Hasan MJ, Mahbub S, Faisal MA, Islam MR, Ali MA. Edge-optimized Inception-v3 for real-time tomato disease detection: balancing accuracy and computational efficiency in smart farming. Comput Electron Agric. 2023;207:107761. doi:10.1016/j.compag.2022.107761. [Google Scholar] [CrossRef]

159. Praveen S, Jung Y. CBAM-STN-TPS-YOLO: enhancing agricultural object detection through spatially adaptive attention mechanisms. arXiv:2506.07357. 2025. doi:10.48550/arXiv.2506.07357. [Google Scholar] [CrossRef]

160. Lee SH, Goëau H, Bonnet P, Joly A. Attention-based recurrent neural network for plant disease classification. Front Plant Sci. 2020;11:601250. doi:10.3389/fpls.2020.601250. [Google Scholar] [PubMed] [CrossRef]

161. Liu J, Wang X. Early recognition of tomato gray leaf spot disease based on MobileNetv2-YOLOv3 model. Plant Methods. 2020;16(1):1–16. doi:10.1186/s13007-020-00624-2. [Google Scholar] [PubMed] [CrossRef]

162. Albattah W, Nawaz M, Javed A, Masood M, Albahli S. A novel deep learning method for detection and classification of plant diseases. Comp Intell Syst. 2022;8(1):507–24. doi:10.1007/s40747-021-00536-1. [Google Scholar] [CrossRef]

163. Liu Y, Chen X, Wang H, Zhang R. Explainable deep learning for real-time plant disease diagnosis with visual interpretation. Precis Agric. 2023;24(2):345–61. doi:10.1007/s11119-022-09934-0. [Google Scholar] [CrossRef]

164. Yuan W, Lan L, Xu J, Sun T, Wang X, Wang Q, et al. Smart agricultural pest detection using I-YOLOv10-SC: an improved object detection framework. Agronomy. 2025;15(1):221. doi:10.3390/agronomy15010221. [Google Scholar] [CrossRef]

165. Duan J, Ding H, Kim S. A multimodal approach for advanced pest detection and classification. arXiv:2312.10948. 2023. doi:10.48550/arXiv.2312.10948. [Google Scholar] [CrossRef]

166. Montagut Marques MJ, Mingxin L, Shiojiri KT, Hagiwara T, Hirose K, Shiojiri K, et al. Plant Doctor: a hybrid machine learning and image segmentation software to quantify plant damage in video footage. arXiv.2407.02853. 2024. doi:10.48550/arXiv.2407.02853. [Google Scholar] [CrossRef]

167. Li WQ, Han XX, Lin ZB, Rahman A. Enhanced pest and disease detection in agriculture using deep learning-enabled drones. Acadlore Transact AI Mach Learn (ATAIML). 2024;3(1):1–10. doi:10.56578/ataiml030101. [Google Scholar] [CrossRef]


Cite This Article

APA Style
Mallick, M.T., Banerjee, S., Thakur, N., Saha, H.N., Chakrabarti, A. (2025). Evaluation of State-of-the-Art Deep Learning Techniques for Plant Disease and Pest Detection. Computers, Materials & Continua, 85(1), 121–180. https://doi.org/10.32604/cmc.2025.065250
Vancouver Style
Mallick MT, Banerjee S, Thakur N, Saha HN, Chakrabarti A. Evaluation of State-of-the-Art Deep Learning Techniques for Plant Disease and Pest Detection. Comput Mater Contin. 2025;85(1):121–180. https://doi.org/10.32604/cmc.2025.065250
IEEE Style
M. T. Mallick, S. Banerjee, N. Thakur, H. N. Saha, and A. Chakrabarti, “Evaluation of State-of-the-Art Deep Learning Techniques for Plant Disease and Pest Detection,” Comput. Mater. Contin., vol. 85, no. 1, pp. 121–180, 2025. https://doi.org/10.32604/cmc.2025.065250


cc Copyright © 2025 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 15197

    View

  • 3325

    Download

  • 0

    Like

Share Link