Skip to main content

Advances, Systems and Applications

Detection of cotton leaf curl disease’s susceptibility scale level based on deep learning

Abstract

Cotton, a crucial cash crop in Pakistan, faces persistent threats from diseases, notably the Cotton Leaf Curl Virus (CLCuV). Detecting these diseases accurately and early is vital for effective management. This paper offers a comprehensive account of the process involved in collecting, preprocessing, and analyzing an extensive dataset of cotton leaf images. The primary aim of this dataset is to support automated disease detection systems. We delve into the data collection procedure, distribution of the dataset, preprocessing stages, feature extraction methods, and potential applications. Furthermore, we present the preliminary findings of our analyses and emphasize the significance of such datasets in advancing agricultural technology. The impact of these factors on plant growth is significant, but the intrusion of plant diseases, such as Cotton Leaf Curl Disease (CLCuD) caused by the Cotton Leaf Curl Gemini Virus (CLCuV), poses a substantial threat to cotton yield. Identifying CLCuD promptly, especially in areas lacking critical infrastructure, remains a formidable challenge. Despite the substantial research dedicated to cotton leaf diseases in agriculture, deep learning technology continues to play a vital role across various sectors. In this study, we harness the power of two deep learning models, specifically the Convolutional Neural Network (CNN). We evaluate these models using two distinct datasets: one from the publicly available Kaggle dataset and the other from our proprietary collection, encompassing a total of 1349 images capturing both healthy and disease-affected cotton leaves. Our meticulously curated dataset is categorized into five groups: Healthy, Fully Susceptible, Partially Susceptible, Fully Resistant, and Partially Resistant. Agricultural experts annotated our dataset based on their expertise in identifying abnormal growth patterns and appearances. Data augmentation enhances the precision of model performance, with deep features extracted to support both training and testing efforts. Notably, the CNN model outperforms other models, achieving an impressive accuracy rate of 99% when tested against our proprietary dataset.

Introduction

Plant diseases stand as formidable adversaries, wielding significant influence over food production. They exact a heavy toll on crop yields, inflicting economic losses and at times even obstructing agricultural activities. As emphasized by [1], the effective management and control of diseases are paramount to mitigate output losses and ensure the sustainability of agriculture. This underscores the critical need for ongoing crop monitoring paired with swift and precise disease detection. In tandem with the burgeoning global population, there arises an unprecedented demand for increased food production, as articulated by the Food and Agriculture Organization (FAO) [2]. This imperative must harmonize with the imperative to preserve natural ecosystems, advocating for environmentally friendly farming practices. The challenge is multifaceted: sustaining the nutritional value of food while ensuring its widespread safety [3,4,5,6,7]. Meeting this challenge necessitates the adoption of novel scientific approaches for diagnosing leaf diseases and managing crops. Furthermore, these innovative technologies must extend their reach to encompass large-scale ecosystem surveillance. At the heart of this quest lies the crucial task of accurately identifying crop diseases [8]. Manual techniques, as delineated by Miller et al., fall short when it comes to covering extensive crop areas and providing early insights for decision-making [9]. Consequently, researchers have been relentless in their pursuit of automated and practical solutions for disease detection. Deep Learning (DL)-based models have emerged as formidable allies in this endeavor, transcending the limitations of traditional classification methods [4,5,6,7]. They represent the vanguard of technology in the field of plant disease detection. DL, a sophisticated technique with a track record of success across diverse domains [10], operates by abstracting data at a high level through a series of transformative operations [11]. This transformative approach heralds a new era in the quest to safeguard our crops and nourish our growing global population.

Agriculture constitutes a vital pillar of Pakistan's economy, contributing substantively to its Gross Domestic Product (GDP) by a proportion of around 19% and engaging approximately 38% of its labor force [12]. This economic dependency underscores the pivotal role of crop health in sustaining food security and production. The primacy of cotton cultivation in Pakistan is noteworthy, as it is not only a principal cash crop, adding 0.6% to the GDP, but also contributes 2.4% to the value addition within the agricultural sector [13]. Termed "white gold" in developing economies, the success of cotton growth is conditioned by both environmental factors and prudent human management practices.

However, diseases like the Cotton Leaf Curl Virus (CLCuV) pose serious threats to cotton yields. Timely and accurate disease detection is crucial for implementing appropriate management strategies. A noticeable gap exists in the prevailing methodologies for detecting the Cotton Leaf Curl Disease (CLCuD), which necessitate significant manual intervention. This underscores the exigency for a comprehensive, automated framework capable of discerning the disease across multiple scales, robustly navigating challenging contexts, and furnishing timely guidance for strategic interventions to mitigate economic losses. The transformative potential of deep learning and machine learning paradigms has significantly propelled the detection of plant leaf ailments, notably the Cotton Leaf Curl Disease (CLCuD) [14]. These techniques harness sophisticated algorithms, adept at discerning intricate patterns within extensive datasets, thereby engendering the formulation of accurate and efficacious detection models. The current work aims to implement an optimized deep learning model to describe the state-of-the-art identification and examination of Cotton plant disease detection problems using a particular class of deep learning (DL) called CNN, which extends traditional Artificial Neural Networks (ANN) by adding more "depth" to the network and the various convolutions that enable the data to be successfully applied in various image-related problems [15].

In the domain of CLCuD detection, (CNNs) and analogous deep learning architectures possess the capacity to dissect leaf images, thereby unveiling markers indicative of CLCuD and adeptly discriminating between healthy and afflicted leaves. Additionally, the CLCuD detection sphere embraces machine learning mechanisms such as Random Forest, Support Vector Machines (SVM), and K-Nearest Neighbors (KNN). These algorithms scrupulously scrutinize foliar data to discern patterns characteristic of CLCuD, enabling the classification of leaves into healthy and pathological categories. Noteworthy is the endeavor of [16], who curated a dataset of cotton leaf images, subjected them to noise reduction and contrast enhancement, subsequently extracting texture structures using the Grey Level Co-occurrence Matrix (GLCM) and shape features via Histogram of Oriented Gradients (HOG). A Support Vector Machine (SVM) classifier, trained on these features, demonstrated efficacy. Similarly, [17] embarked on preprocessing a cotton leaf image dataset through resizing and grayscale conversion, leveraging the VGG-16 architecture coupled with transfer learning and data augmentation. The ensuing model's performance was assessed vis-à-vis metrics and contrasted with other deep learning models. Importantly, their model exhibited superior performance, further substantiated through sensitivity analysis across diverse disease categories. This research yielded a methodological blueprint for robustly identifying and classifying cotton leaf diseases, buttressed by deep learning models. Presently, the predominant landscape of CLCuD detection accentuates the scrutiny of macroscopic leaf attributes, while the nuanced micro and nano scales of disease progression often remain sidelined. The intrinsic disposition of our self-curated dataset culminates in 252, 132, 350, 285, and 330 instances for the respective classes of Fully Resistant (FR), Partially Resistance (PR), Healthy (H), Partially Susceptible (PS), and Fully Susceptible (FS) as depicted in Fig. 1. The assemblage of our self-collected dataset encompasses 1349 images spanning five distinct classes (FS, PS, H, PR, FR), where 1052 images are earmarked for training and 297 for testing. The dataset available to the public is characterized by a binary distinction, with one class comprising 418 instances associated with the Curl-Virus category, and the other class containing 426 instances denoting the Healthy category. Transitioning into the subsequent phase, we adopt an exhaustive array of preprocessing techniques, orchestrated to extract multifarious features from the encompassing image dataset. This elaborate feature extraction process serves as a precursor to our utilization of two distinct Deep Learning (DL) models, specifically CNN, for the fundamental task of classifying the Cotton Leaf Curl Disease (CLCuD).

Fig. 1
figure 1

CLCuV susceptibility scale level

In the culminating phase of our endeavor, the image dataset undergoes a classification operation, facilitated by the CNN model, which is driven by meticulously calculated weights. The CNN paradigm is renowned and extensively employed in the realm of image recognition and classification, a fact well-established within the scholarly domain [4,5,6,7, 18]. The ambit of this research encompasses the design and instantiation of a Vision model characterized by a stratified architecture comprising seven distinct layers: conv2d_2, max_pooling2d_2, conv2d_3, max_pooling2d_2, flatten_1, dense_2, and dense_3. This dual-pronged model fabric has been meticulously conceived to effectively disentangle the nuanced domains of susceptibility and resistance pertaining to the Cotton Leaf Curl Disease (CLCuD), predicated on the discernible symptomatic attributes. This study focuses on exploring the complexities of building a robust dataset of cotton leaf images, investigating feature extraction techniques, investigating preprocessing techniques, and thinking about the potential uses of such datasets in disease diagnosis and agricultural research.

The holistic appraisal encompasses quintessential metrices, viz. training accuracy, testing accuracy, training loss, and testing loss. Notably, our proposed model’s performance acumen on our in-house dataset is conspicuously prominent, characterized by an illustrious training accuracy index of 94.57%, complemented by a compelling testing accuracy record of 99%.

The present research makes a series of notable contributions, including but not limited to:

  • Recognize the many kinds of cotton leaf diseases and their frequency.

  • Recognize the connection between environmental elements and cotton leaf diseases.

  • Propose an automated CLCuD detection method that accurately classifies susceptibility scale levels by cotton leaf images.

  • Capture multiple images to build a self-collected dataset that classifies CLCuD based on visual symptoms.

  • Pre-processing steps are applied to the captured images, resizing images, normalizing image, augmentation, and the cleaning step start in which blur images are removed and the background of images is replaced with a neutral color for better visibility.

  • To extract features from images, apply preprocessing approaches.

  • To implement an optimized deep learning model for predicting susceptibility levels of the CLCuD self-collected dataset as well as the downloaded dataset.

Researchers may utilize this knowledge to create new strategies for avoiding and controlling cotton leaf diseases, and farmers can use it to recognize and manage diseases in their farms. The step-by-step process of how our models is to get input image, perform pre-processing steps, segmentation, features extractions, and classification in Fig. 2.

Fig. 2
figure 2

Flow chart of this study

Literature review

Numerous studies have been done to develop methods that can help identify crops in an agricultural setting to find the best answer to the issue of crop disease detection.

Abade et al., akin to vigilant gardeners, meticulously examined CNN algorithms tailored for the detection of plant diseases. Their journey through 121 papers spanning the decade from 2010 to 2019 unveiled Plant Village as the quintessential dataset, while the resounding echo of Tensor Flow emerged as the most frequented framework in this botanical symphony [19]. Dhaka et al., with the precision of botanists, sketched the blueprint of CNN models deployed in the identification of plant diseases through leaf images. They embarked on a comparative expedition, scrutinizing CNN models, pre-processing techniques, and the frameworks that cradle them. Their exploration extended to the fertile terrain of datasets and performance metrics, vital signposts in the quest for model prowess [20]. Meanwhile, Nagaraju et al., akin to data archaeologists, embarked on a journey to unearth the gems of datasets, pre-processing techniques, and the mystical arts of Deep Learning in the realm of plant disease diagnosis [21]. With an analytical lens, they dissected 84 papers, discovering that many DL methods yearned for the keen insight of suitable pre-processing techniques to unlock their full potential. Kamilaris et al., the pioneers of agricultural innovation, charted the course of DL approaches as they ventured into solving the multifaceted challenges of agriculture. Their findings, like fertile soil, revealed that DL methods outshone conventional image processing techniques, promising greener pastures in agricultural technology [22].

Fernandez-Quintanilla et al., the sentinels of weed control, focused their gaze on the evergreen battlegrounds of agricultural fields. They surveyed both the remote realms and the grounded territories of weed-monitoring technologies. Their prognosis: Weed monitoring is the linchpin in the battle for weed control, and the future holds the promise of harnessing sensor data, stored in the boundless expanse of the public cloud, to be wielded judiciously in the crucible of crop protection [23]. Lu et al., the maestros orchestrating the symphony of plant disease classification, tuned their instruments to the resonance of CNN. Their composition evaluated the crescendos and diminuendos of CNN in the realm of plant disease classification, identifying the need for more intricate datasets to compose a harmonious melody [24]. Golhani et al., the virtuosos of hyperspectral data, painted a portrait of leaf disease identification. They illuminated the challenges and the shimmering prospects in the realm of hyperspectral data. Their canvas extended to the realm of NN approaches, offering a swift brushstroke on SDI development [25]. Bangari et al., the connoisseurs of disease detection, plucked the strings of CNN in the context of potato leaf diseases. Their review, akin to a masterful composition, resonated with the chorus that CNNs excelled in disease detection, weaving a harmonious tapestry of accuracy in the annals of plant pathology [26].

The landscape of disease detection methodologies has witnessed a dynamic transformation with the emergence of the dynamic weighted layering model, a promising Deep Learning (DL)-based technique geared towards detecting Cotton Leaf Curl Disease (CLCuD). Rooted in the utilization of multiple layers of nodes endowed with adaptable weights, this model is tailored to optimize performance, garnering considerable attention across diverse investigations.

In a related vein, Pechuho, Khan, and Kalwar [27] present an intriguing machine learning-based approach to the identification of cotton crop diseases. Their endeavors encompass dataset compilation, preprocessing, and the employment of CNN fortified by transfer learning, culminating in a superior classification proficiency that surpasses the ambit of alternative Machine Learning (ML) techniques across various evaluation metrics.

Conversely, the study by Tripathy [16] unfolds with the curation of a dataset housing cotton leaf images stratified across three distinct categories. Texture structures extracted via Grey Level Co-occurrence Matrix (GLCM) and shape features derived through Histogram of Oriented Gradients (HOG) serve as pivotal features. In this context, a Support Vector Machine (SVM) assumes the role of classifier, constituting a robust model for discerning and classifying instances within the dataset. A compelling manifestation of the nexus between DL and image analysis is encapsulated within Magsi, Shaikh, Shar, Arain, and Soomro’s [28] research, wherein a dataset comprising 1600 images of CLCuD-infected cotton plants is harnessed. The model construction hinges on Information Processing (IP) techniques, facilitating the extraction of color and textual attributes. Propelled by a deep CNN, this approach navigates decision-making and classification tasks with finesse, yielding an efficacious framework capable of identifying the severity of CLCuD in a swift fashion. This outcome culminates in timely intervention strategies, instrumental in abating potential losses.

Zhu et al. [29] put forth an innovative methodology to decipher crop diseases, predicated upon the synergy of a Transformer Encoder and Centerloss optimization. This amalgamation fosters heightened feature analysis and accentuates differentiation between distinct diseases, triumphing over impediments like focal disease spot recognition and confusion engendered by resembling diseases. A novel approach surges to the forefront within Li, Wang, and Hu’s [30] study, wherein a CNN architecture with three pivotal modules—feature extraction, classification, and augmentation—takes center stage. Propelled by residual connections and spatial pyramid pooling, this model systematically outpaces conventional ML techniques, an accomplishment vividly illustrated through a comprehensive array of evaluation metrics.

In a similar vein, Amin, Darwish, Hassanien, and Soliman [31] meticulously orchestrate a methodology underscored by feature abstraction and classification. Underpinned by data augmentation to recalibrate input data distribution, this approach employs a pre-trained VGG-16 model for feature extraction. The resultant model is subjected to meticulous hyperparameter tuning, culminating in a classification proficiency marked by accuracy and efficacy.

Equally noteworthy, Naeem et al. [17] delve into dataset compilation, preprocessing, and classification within the domain of cotton leaf disease identification. Leaning on the VGG-16 architecture with transfer learning and augmented data strategies, their model vaults ahead of rival deep learning models. The culmination of their study encompasses a sensitivity analysis tailored to ascertain the model's efficacy across diverse disease categories. Collectively, this panorama of literature unfolds a tapestry of diverse methodologies and approaches, coalescing around the common pursuit of robust and accurate disease detection techniques within the realm of agriculture. In the Table 1 different paper accuracy results are compared.

Table 1 Comparative results of models

Methodology

Our proposed technique offers a systematic method for the rapid identification of the susceptibility scale levels of Cotton Leaf Curl Disease (CLCuD) using picture files. In the process, the Cotton Leaf Curl Virus (CLCuV) exhibits a variety of modifications that are categorized by looking at DNA patterning. Figure 3 provides a schematic representation of the process’s steps. In the first stage of our suggested method, a wide range of pictures of cotton plants are taken, making it easier to create a self-curated dataset. The second phase then involves obtaining a publicly available dataset on cotton leaf disease from the Kaggle platform. The third step entails taking stand-out features from the image dataset. By using a variety of preprocessing approaches, such as image resizing, image normalization, and image augmentation, this extraction process is made easier. In order to classify CLCuD, these processed datasets are then fed into a Deep Learning (DL) model, specifically a CNN.

Fig. 3
figure 3

Proposed framework

The model is expanded in the last stage of our suggested methodology to categorize susceptibility scale levels into five different groups: Fully Resistant (FR), Partially Resistant (PR), Healthy (H), Partially Susceptible (PS), and Fully Susceptible (FS). The weights associated with the used model have been calculated and used to give these categories. Based on measures for both training accuracy and testing accuracy, the CNN model's performance is assessed. This thorough methodology guarantees reliable detection and categorization of CLCuD susceptibility levels and may have implications for improving disease control tactics in cotton farming.

Following steps are exist in the proposed frame work Self-collected dataset description, downloaded dataset description, feature extraction and preprocessing, construction of CNN model and results.

Data collection

The images constituting the cotton leaf disease dataset were gathered from diverse cotton fields located in Multan, Pakistan. Specifically, the following locations were chosen:

  • Muhammad Nawaz Sharif University of Agriculture Multan (35% of the images).

  • Bahauddin Zakariya University Multan (35% of the images).

  • Other cotton-cultivated areas within Multan (20% of the images).

  • Plants that were deliberately inoculated with the Cotton Leaf Curl Virus (CLCuV) (10% of the images).

These locations were carefully selected to ensure that the dataset accurately represented a range of cotton cultivation conditions and various levels of disease severity. To capture these images, a DSLR camera was employed, configured with the following settings:

  • Image formats: JPEG and PNG.

  • Image resolution: 256x256 pixels.

  • ISO: 100

  • Shutter speed: 1/100 second

  • Aperture: f/8

These settings were specifically chosen to yield high-quality images that exhibit a consistent and standardized appearance across the dataset.

Dataset description

The dataset that used in this is given in the following (Tables 2, 3 and 4). Table 2 show the detail self-collected with number of training and testing images with each class. Fully Resistant (FR): This category encompasses images of cotton leaves entirely impervious to CLCuV. These FR cotton leaves exhibit no indications of CLCuV infection. Partially Resistant (PR): Within this class, you'll find images of cotton leaves that possess partial resistance to CLCuV. PR cotton leaves may display mild CLCuV symptoms, but they can recover and yield normally. Healthy(H): This classification comprises images of robust and uninfected cotton leaves. Healthy cotton leaves maintain their natural green color and regular shape. Partially Susceptible (PS): In this category, you'll discover images of cotton leaves that display partial susceptibility to CLCuV. PS cotton leaves manifest moderate CLCuV symptoms, which may reduce their yield potential. Fully Susceptible (FS): This category features images of cotton leaves highly susceptible to CLCuV. FS cotton leaves exhibit severe CLCuV symptoms, often resulting in crop failure.

Table 2 Detail of self-collected dataset
Table 3 Self-collected dataset class symptoms
Table 4 Detail of downloaded dataset

Table 3 shows the symptoms of self-collected dataset and Table 4 shows the number of training and testing images with each belonging class in downloaded dataset.

Table 3 shows that PS class has 233 and 52 images respectively for train and test, in this the leaf color become darker. The complete curl leaf cup shape upward and downward belongs to this FS class contain 277 train and 53 test images.

In the Table 4 a description about downloaded dataset is given in detail. It contains two classes Curl-Virus and healthy.

Preprocessing and feature extraction

After the pictures were shot, they were cleaned to get rid of blurry pictures and change the background color to something neutral. Various image processing methods were chosen to reduce noise and make the cotton leaves more visible in the photos.

Image resizing

To diminish the dimensions of images, a resizing function within the realm of Image Processing is harnessed, employing a computer vision library within the Python programming framework. The entirety of the cotton leaf images undergoes a harmonizing transformation, conformed to a standardized size of 256 by 256 pixels (width * height) as shown in Fig. 4.

Fig. 4
figure 4

Sample images of self _collected dataset

$$R=Re \left(I, Wo, Ho, Wr, Hr\right)$$
(1)

R represents the resized image. Re is the resizing operation applied to the original image. I stand for the original image. WO and HO are the width and height of the original image, respectively. Wr and Hr are the desired width and height for the resized image.

Image normalization

The process of aligning the intensity values of resized cotton leaf images with a predetermined range was successfully achieved through a procedure of normalization. This operation was executed within the context of the Python programming language, facilitated by the Google Colab platform. By subjecting the images to this normalization process, their intensity values were meticulously recalibrated to harmonize with the desired range. This recalibration engendered greater uniformity and coherence among the data, thereby fostering a more conducive environment for subsequent analytical undertakings. The equation encapsulating the essence of normalization is represented as follows:

$${\text{XN}}=\frac{{\text{Xmx}}-\mathrm{Xmn }}{{\text{X}}-{\text{Xmn}}}$$
(2)

X represents the original data value. XN is the normalized value of X. X mn is the minimum value of the data. X mx is the maximum value of the data. This equation transforms the data value X into a normalized value that falls within the range of 0 to 1 based on its relationship with the minimum and maximum values of the dataset as display in Fig. 5.

Fig. 5
figure 5

Sample of original image and resized image

Image augmentation

An amplified quantum of data becomes imperative to augment the efficacy of model training. In response to this requisition, the methodology of data augmentation assumes prominence. This approach entails the fabrication of novel data instances derived from the preexisting repository, thereby engendering a deliberate expansion of the dataset's magnitude.

This augmentation endeavor encompasses a spectrum of operations administered to images, encompassing rotational manipulations, mirror reflections, spatial cropping, and the incorporation of perturbations or distortions. The assimilation of these manifold alterations into the dataset confers upon the model a more comprehensive and diversified pedagogical trajectory, thus amplifying its resilience and refining its performance throughout the training regimen as shown in Fig. 6.

Fig. 6
figure 6

Sample of resized image and normalized image

D represent the original dataset containing N instances. Data augmentation introduces a set of M augmentation operations denoted as A = {A1, A2, A3, …. AM}, each representing a specific data transformation. This results in an augmented dataset ′ D ′ containing N × M instances, where n*m = W:

$$\mathrm{D {\prime}} = \{\mathrm{D }1,\mathrm{ D }2 ...,\mathrm{ D W}\}$$
(3)

where ′ D ′ is the augmented dataset, D i is the i-th instance after augmentation, and M represents the number of augmentation operations. Each augmentation operation A j modifies an instance D i according to the operation’s characteristics, producing a transformed instance D I j:

$${\text{Di}},\mathrm{ j }=\mathrm{A j }(\mathrm{D i })$$
(4)

where Di, j denotes the instance after applying the j-th augmentation operation to the i-th original instance, and A j is the j-th augmentation operation. In this manner, the augmented dataset ′ D ′ is constructed, comprising a diverse collection of transformed instances that enhance the dataset's richness and facilitate improved model training.

Proposed model

Specifically, we employed CNN model. The purpose of employing these models was to assess the outcomes generated from both datasets.

CNN model

A common class of ANN used in image processing and recognition is referred to as CNN. Numerous advances in CNN architectures have been presented since the 1998 release of LeNet-5 [36]. Additionally, learning relied on extracting interesting variables called features prior to the development of DL for computer vision. These techniques, meanwhile, necessitate a good deal of prior knowledge in image processing. With the advent of CNN [37], image processing was revolutionized, and manual feature extraction was done away with. The classification, segmentation, face recognition, and object recognition of images are increasingly frequently performed using CNNs. Many organizations have effectively used them in a variety of fields, including health, the web, postal services, etc. Images, video, sound, speech, and natural language can all be fed into CNN [38, 39]. Convolution, pooling, Relu correction, and fully connected layers are all simply stacked together to form CNN (see Fig. 7), which starts with a convolution layer and progresses through the following layers: pooling, Relu correction, and finally a fully-connected layer [40,41,42,43].

Fig. 7
figure 7

CNN architecture

The proposed CNN model is structured as a composition of seven distinct layers, namely conv2d_2, max_pooling2d_2, conv2d_3, max_pooling2d_2, flatten_1, dense_2, and dense_3. Within the domain of Deep Learning, CNNs have attained a prominent stature due to their efficacy in tasks encompassing the classification of images, detection of objects, and segmentation of images. This model class operates through a process of feature extraction from input images via convolutional layers, subsequently facilitating a process of subsampling through pooling layers, and ultimately culminating in the process of classification through fully connected layers. The inherent architecture of CNNs conventionally integrates convolutive, pooling, and fully connected strata, which are further fortified by the infusion of normalization, activation, and regularization layers. This augmentation serves the dual purpose of optimizing performance and mitigating the challenges of overfitting. These intrinsic architectural attributes find explication within Fig. 7. It is noteworthy that CNNs have engendered remarkable progress within the domain of computer.

Vision, establishing their pervasive utilization across pivotal domains such as autonomous vehicular navigation, the analysis of medical imagery, and the domain of facial recognition. transformations, including the establishment of a consistent image size, potential conversion to grayscale or RGB color formats, and the normalization of pixel values to a standardized range of 0 to 1 as shown in Fig. 8.

Fig. 8
figure 8

Sample of normalized image and augmented image

Input layer

The input configuration of the proposed model encompasses two distinct sets of images, allocated to both the training and testing datasets. These input images undergo a preliminary preprocessing phase, ensuring their compatibility with the CNN architecture. This preprocessing entails several This preparatory stage facilitates optimal data representation and harmonizes the images for subsequent processing within the CNN framework.

Convolutional layer

Our proposed model architecture integrates two essential convolutional layers, specifically referred to as con2d_2 and conv2d_3, as shown in Fig. 9. These layers play a fundamental role in extracting meaningful and pertinent features from the input images during the model's training phase. The initial layer, con2d_2, is responsible for receiving input images and subsequently transmitting the generated output to the subsequent processing stage. Through this process, the input images are transformed, yielding an output matrix with dimensions (64, 64, 32), accompanied by a parameter count of 896. Following con2d_2, the subsequent layer, conv2d_3, becomes operative. This layer takes input from the preceding max_pooling2d_2 layer and contributes to the derivation of output parameters. The input configuration for conv2d_3 is defined as (32, 32, 32), culminating in an output shape of (32, 32, 32). Consequently, the layer encompasses 9248 parameters. The mathematical formulation underlying convolutional layers can be succinctly expressed as follows:

Fig. 9
figure 9

CNN convolutional level function

$${\varvec{Y}} ({\varvec{I}}, {\varvec{J}}, {\varvec{K}})={\varvec{b}}\left({\varvec{K}}\right)+\sum \left({\varvec{Q}}\times {\varvec{F}}{\varvec{m}}\left({\varvec{i}} + 1,{\varvec{j}} + {\varvec{m}}, {\varvec{n}}\right)\right)$$
(5)

Where F[l,m,n,k] = Q Y is the output feature map, b is the bias, F is image features, Fm is input feature map.

Max pooling layer

Max_pooling2d_2 and Max_pooling2d_3, two crucial max pooling layers, are cleverly used to achieve dimensional reduction in image pixel representations. The second layer of our unique CNN model, called Max_pooling2d_2, receives input from the Conv2d_2 output matrix and manages the dimensionality reduction. A reduced output matrix with the dimensions (32, 32, 32) is produced by transforming the initial input matrix, which has (64, 64, 32) dimensions, as shown in Fig. 10.

Fig. 10
figure 10

Pooling function of CNN model

Our models fourth layer, Max_pooling2d_3, acts as a bridge between the Conv2d_3 and Flatten_1 layer. It produces an output matrix with dimensions (16, 16, 32) by operating on an input matrix of shape (32, 32, 32), which is produced from the preceding layer. Following that, this matrix is propagated to the following layer for additional processing. The mathematical model demonstrating the functionality of the Max_pooling layer is expressed concisely as follows:

$${\varvec{F}}{\varvec{t}}({\varvec{i}},\boldsymbol{ }{\varvec{j}},\boldsymbol{ }{\varvec{k}})={\varvec{M}}{\varvec{p}}({\varvec{F}}{\varvec{p}}({\varvec{i}}{\varvec{s}}{\varvec{d}}+{\varvec{l}},\boldsymbol{ }{\varvec{j}}{\varvec{s}}{\varvec{d}}+{\varvec{m}},\boldsymbol{ }{\varvec{k}}))$$
(6)

where Ft is output feature map, Mp max_over _pool, Fp Input feature map, isd is istride, jsd is jstride.

Flatten layer

Flatten_1, situated as the fifth layer in the model, orchestrates a nuanced transition by molding the input matrix sourced from Max_pooling2d_3. This transformation yields a structured array, subsequently routed to the subsequent dense layer for intricate computational analysis. Characterized by dimensions (16, 16, 32) for the input matrix, it engenders an array with dimensions (8192). Crucially, the pivotal role of Flatten_1 resides in its intricate facilitation of neural network architecture—an artful reconstitution of the input tensor into a refined unidimensional vector. The equation for the Flatten_1 layer can be written as:

$${\varvec{Y}}=\boldsymbol{ }{\varvec{i}}.\boldsymbol{ }{\varvec{r}}{\varvec{e}}{\varvec{p}}\boldsymbol{ }\left({\varvec{s}},\boldsymbol{ }-1\right)$$
(7)

where y is output, I is input value, rep is reshape, bs is batch_size.

Dense layer

The bespoke CNN architecture incorporates a tandem of dense layers—Dense_2 and Dense_3. Serving as the sixth stratum, Dense_2 engenders a complexity of 1048704 parameters. It transmutes the input array stemming from Flatten_1, channeling it to the ensuing Dense_3 layer. Imbued with an input array characterized by dimensions (8192), Dense_2 orchestrates a reduction, yielding an output array of dimensions (156). This array, in turn, is propagated to the subsequent layer. In a parallel fashion, the ultimate and terminal layer, Dense_3, concludes the model's architecture, encapsulating 645 parameters. Drawing its input from Dense_2, furnished with an input array bearing dimensions (156), Dense_3 undertakes the task of mapping it onto 5 distinct classes. This predictive mapping serves to ascertain the susceptibility scale level of CLCuD. The mathematical essence of the Dense layer finds articulation in the ensuing equation.

$${\varvec{Y}}=\boldsymbol{ }{\varvec{a}}\boldsymbol{ }\left({\varvec{W}}\boldsymbol{ }\left({\varvec{i}},\boldsymbol{ }{\varvec{w}}{\varvec{m}}\right)+{\varvec{b}}{\varvec{v}}\right)$$
(8)

where Y is output, a is activation, W is dot values, is input, wm is weight _matrix, bv bias vector.

Output

The output is meticulously mapped across susceptibility to resistance levels of CLCuD, defining discrete data classes corresponding to distinct scale levels.

Experiments and result

The experimental configuration entails the deployment of a 238 GB Solid State Disk and a motherboard with 12 GB of RAM. The system is operational with Windows 10 Pro as the operating system, supported by an Intel (R) Core (TM) processor. The experimentation environment further incorporates the utilization of Google Colab platform, Python programming language, and the availability of a Google Colab GPU.

Evaluation matrix

In the context of assessing machine learning models, commonplace performance metrics encompass accuracy and loss. The formulation denoting accuracy finds prevalent application as a quintessential measure for evaluation.

$${\varvec{A}}{\varvec{c}}{\varvec{c}}{\varvec{u}}{\varvec{r}}{\varvec{a}}{\varvec{c}}{\varvec{y}}=\frac{{\varvec{T}}{\varvec{P}}+{\varvec{F}}{\varvec{P}}}{{\varvec{T}}{\varvec{P}}+{\varvec{F}}{\varvec{P}}+{\varvec{T}}{\varvec{N}}+{\varvec{F}}{\varvec{N}}}$$
(9)

The count of accurate predictions represents instances where the model's output coincided with the true anticipated result. Conversely, the aggregate count of predictions pertains to the cumulative instances upon which the model's predictions were applied. The formulation governing the loss metric is contingent upon the intrinsic nature of the problem under consideration. For classification problems, a commonly used loss function is cross-entropy loss, which is defined as:

$${\varvec{L}}{\varvec{o}}{\varvec{s}}{\varvec{s}}\boldsymbol{ }=\boldsymbol{ }-1/{\varvec{N}}\times {\varvec{s}}{\varvec{u}}{\varvec{m}}({\varvec{y}}\times {\varvec{l}}{\varvec{o}}{\varvec{g}}({\varvec{y}}{\varvec{p}})+(1-{\varvec{y}})\times {\varvec{l}}{\varvec{o}}{\varvec{g}}(1-{\varvec{y}}{\varvec{p}}))$$
(10)

where y_pred = yp. Within this context, N signifies the count of instances subjected to analysis. The variable y denotes the veritable label, encompassing binary values of either 0 or 1. Conversely, y_pred encapsulates the predictive probability attributed to the affirmative class.

Comparison of accuracy

The scrutiny encompasses a comparative analysis of accuracy metrics between the self-collected dataset and the downloaded dataset.

Table 5 provides a comprehensive exposition of the training and testing accuracy measurements pertaining to both datasets, namely the self-collected and downloaded datasets. Evidently, the tabulated results underscore that the self-collected dataset yields a training accuracy of 94.57% coupled with an impressive testing accuracy of 99%. Conversely, the downloaded dataset demonstrates commendable training accuracy of 97.49%, juxtaposed with a testing accuracy of 89.71%. Figure 11 serves as an illustrative repository, visually presenting the graphical depictions of training accuracy and testing accuracy for both datasets.

Table 5 Training and testing accuracy
Fig. 11
figure 11

Accuracy graph of both datasets

Within this tabulation, the initial column encapsulates the accuracy manifestations derived from the self-collected dataset, while the subsequent column delineates the accuracy metrics originating from the downloaded dataset. The x-axis of the accuracy graphs delineates a spectrum spanning from 0 to 20 epochs, parameters germane to the custom CNN model's utilization. Concurrently, the y-axis offers a continuum ranging from zero percent to one hundred percent, symbolizing the accuracy spectrum. The graphical representation is characterized by a distinctive blue line, emblematic of the training trajectory, while the orange line, conversely, signifies the testing trajectory of the model.

Comparison of loss

The CNN model which painstakingly compiles the training and testing loss metrics for both the self-collected and downloaded datasets.

Contained within Table 6 is a comprehensive exposition of the training and testing loss metrics, effectively encapsulating the intricacies of both the self-collected and downloaded datasets.

Table 6 Training and testing loss

Noteworthy is the discerning insight provided by the tabulated data, revealing that the self-collected dataset exhibits a training loss of 16.38%, accompanied by a relatively modest testing loss of 6.56%. In stark contrast, the downloaded dataset showcases a more favorable training loss of 8.36%, while the testing loss exhibits a notably higher magnitude of 52.67%. Figure 12 delineates graphs portraying training and testing loss for both datasets. The initial column pertains to the self-collected dataset's loss, while the second column pertains to the downloaded dataset's loss. The x-axis portrays epochs ranging from 0 to 20 for the custom CNN model, and the y-axis signifies loss from 0 to 100 percent. Blue lines represent training loss, while orange lines depict testing loss. Notably, the self-collected dataset demonstrates declining loss as epochs increase, concurrent with an ascending accuracy trend as shown in Fig. 13. The experimental outcomes are as follows:

  • The application of preprocessing procedures encompassing background removal, data augmentation, and image resizing enhances the dataset's quality.

  • Expert annotation of the Self-Collected dataset, guided by meticulous disease symptom assessment, facilitates the establishment of a susceptibility scale level mapping.

  • The utilization of the Self-Collected dataset yields superior model performance outcomes when contrasted with the downloaded dataset.

Fig. 12
figure 12

Loss graph of both datasets

Fig. 13
figure 13

Comparative graph of accuracy and loss of both datasets after augmentation

Recent advancements across various scientific and engineering disciplines are well-captured through a series of innovative studies. Zhou et al. [44] and Qi et al. [45] have made significant strides in remote sensing and image processing. Zhou et al. [44] developed a novel method for LiDAR hidden echo signal decomposition, while Qi et al. [45] enhanced image quality through a brightness correction algorithm. Lin et al. [46] demonstrated the application of AI in practical fields like construction and infrastructure, focusing on pavement anomaly detection. The study by Y. et al. [47] illustrates the potential of data analytics in knowledge graph completion, enhancing information systems.

In the realm of electronics and communication, Jiang and Li [48] improved interference cancellation systems, and Wang et al. [49] developed a biosensor system for explosive detection, showcasing interdisciplinary research. Liu et al. [50] and Dang et al. [51] leveraged AI in creative and analytical applications, with the former creating photo-realistic images from sketches and the latter developing a feature matching method based on convolutional neural networks. Healthcare technology saw advancements with the surgical instrument localization algorithm by Siyu Lu et al. [52], indicating progress in surgical precision. In environmental applications, Cheng et al. [53] utilized machine learning for vegetation mapping, and Zheng et al. [54] in agricultural monitoring. Tao et al. [55] applied AI in defect recognition, while Zhou et al. [56] and another study by Zhou et al. [57] focused on enhancing signal detection in remote sensing. Object and vehicle detection were advanced by Zhang et al. [58] and Li et al. [59], respectively.

Predictive maintenance was addressed by Zhao et al. [60] in aero-engine life prediction. Control systems for multi-agent and autonomous underwater vehicles were explored by Hu et al. [61] and Chen et al. [62]. Liao et al. [63] and Ding et al. [64] applied AI to detect fake news and taxi fraud. Zhang et al. [65] explored cloud management systems, and network resource allocation and wireless communication were the focus of Xuemin et al. [66] and Lyu et al. [67]. Traffic anomaly detection by Xu et al. [68] and transportation detection by Chen et al. [69] show the breadth of technology’s impact, while Ma et al. [70] enhanced pavement assessment techniques. Lastly, Jin et al. [71] created a dataset for image quality assessment, reflecting the diversity and depth of current technological advancements across sectors.

The study presented here, focusing on the use of deep learning models, specifically Convolutional Neural Networks (CNN), for the detection of Cotton Leaf Curl Disease (CLCuD), marks a significant advancement in agricultural technology, particularly in crop disease management. By successfully harnessing deep learning to analyze a comprehensive dataset of cotton leaf images, the research addresses a critical challenge in agriculture, especially in regions with limited resources. The high accuracy rate of 99% achieved by the CNN model underscores the potential of AI in transforming disease detection processes, leading to more efficient and timely interventions. This approach not only contributes to improving cotton yields by mitigating the impact of diseases like CLCuV but also paves the way for integrating advanced technologies in agriculture. The methodology and findings of this study could be instrumental in developing automated, accurate, and accessible disease detection systems, thereby enhancing crop management and supporting sustainable agricultural practices globally.

Conclusion

The symphony of ecological forces is essential to the complex effects of plant growth and development. Stickiness, temperature, water availability, sunlight, and nutrition are just a few of the variables that play a role in the environmental orchestra that determines a plant's journey. The threat of crop diseases lurks menacingly within this happy coexistence, posing a significant threat to agricultural output and global food security. Problems with illness detection are frequently worsened by insufficient infrastructure in many parts of the world. Enter the centerpiece of our project: a carefully prepared dataset on cotton leaf disease. This data collection is a gold mine that unlocks a world of inventive possibilities.

Using AI for Early Detection: Visualize a world in which cutting-edge machine learning models, cultivated on this dataset, serve as field sentinels. They quickly and precisely identify the unmistakable symptoms of cotton leaf diseases. These digital guardians, which are integrated into mobile apps and web services, provide farmers with real-time assistance, reversing the tide in the war against crop diseases. Beyond practical uses, this dataset serves as a fruitful research environment for the advancement of science. Armed with this wealth of information, researchers set out to solve the riddles of cotton leaf diseases. They make clear the complex connections between the frequency of diseases and the dynamic environmental web. Their results open the door for creative ways to safeguard crops and guarantee food security. Education and Empowerment: By illuminating the road for farmers, the dataset serves as a cornerstone of education. Resources start to appear that are vibrant, like interesting pamphlets and posters. By teaching others about cotton leaf diseases and the art of detection and management, they take on the role of teachers. With their newly acquired knowledge, farmers protect their crops, fostering abundant harvests. Switching gears, we go into the world of Cotton Leaf Curl (CLC) Disease, a scourge that plagues cotton plants and is brought on by the Cotton Leaf Curl (CLC) Gemini virus. Deep learning technology is emerging as a strong force that cuts across numerous industries in the large field of agricultural research. Here, the scene is ready for the CNN models to identify pictures of cotton leaves. Two datasets take center stage, one chock full of Cotton Leaf Curl Disease-related photos and the other a self-assembled treasure trove of 842 and 1349 images, respectively. The power of the CNN is amplified by augmentation techniques, providing thorough training. The data, which includes training, testing, loss, and accuracy, paints an interesting picture. Within the self-assembled dataset, CNN achieves an astounding 99% accuracy, demonstrating the power of deep learning. The voyage is still far from finished. The horizons of the future hold out hope for progress. The first steps toward bigger accomplishments include a wide range of ecological circumstances, an expanded dataset including various regions, and model improvements. This dataset is a priceless resource for researchers and programmers exploring the boundaries of cotton leaf disease detection, prevention, and management. It shines as a beacon of information, directing varied minds toward the objective of protecting our essential crops thanks to its balanced distribution, spanning both the sick and the healthy, and its open accessibility.

The voyage is still far from finished. The horizons of the future hold out hope for progress. The first steps toward bigger accomplishments include a wide range of ecological circumstances, an expanded dataset including various regions, and model improvements.

Availability of data and materials

No datasets were generated or analysed during the current study.

References

  1. Altieri MA (2018) Agroecology: the science of sustainable agriculture; CRC Press: Boca Raton. FL, USA

    Book  Google Scholar 

  2. Gebbers R, Adamchuk VI (2010) Precision agriculture and food security. Science 327:828–831

    Article  ADS  CAS  PubMed  Google Scholar 

  3. Carvalho FP (2006) Agriculture, pesticides, food security and food safety. Environ Sci Policy 9:685–692

    Article  Google Scholar 

  4. Bhatti UA, Bazai SU, Hussain S, Fakhar S, Ku CS, Marjan S, Jing L (2023) Deep learning-based trees disease recognition and classification using hyperspectral data. Comput Mater Contin, 77(1)

  5. Bhatti UA, Tang H, Wu S (2023) Mangrove decline puts Pakistan’s coasts at risk. Science 382(6671):654–655

    Article  ADS  CAS  PubMed  Google Scholar 

  6. Bhatti UA, Marjan S, Wahid A, Syam MS, Huang M, Tang H, Hasnain A (2023) The effects of socioeconomic factors on particulate matter concentration in China’s: new evidence from spatial econometric model. J Clean Prod 417:137969

    Article  CAS  Google Scholar 

  7. Bhatti UA, Huang M, Neira-Molina H, Marjan S, Baryalai M, Tang H, Bazai SU (2023) MFFCG–Multi feature fusion for hyperspectral image classification using graph attention network. Expert Syst Appl 229:120496

    Article  Google Scholar 

  8. Mohanty SP, Hughes DP, Salathé M (2016) Using deep learning for image-based plant disease detection. Front Plant Sci 7:1419

    Article  PubMed  PubMed Central  Google Scholar 

  9. Miller SA, Beed FD, Harmon CL (2009) Plant disease diagnostic capabilities and networks. Annu Rev Phytopathol 47:15–38

    Article  CAS  PubMed  Google Scholar 

  10. LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521:436–444

    Article  ADS  CAS  PubMed  Google Scholar 

  11. Najafabadi MM, Villanustre F, Khoshgoftaar TM, Seliya N, Wald R, Muharemagic E (2015) Deep learning applications and challenges in big data analytics. J Big Data 2:1–21

    Article  Google Scholar 

  12. Rasheed R (2021) Socio-economic and environmental impacts of COVID-19 pandemic in Pakistan — an integrated analysis. 19:19926–19943

  13. Afzal M, Saeed S, Riaz H, Ishtiaq M, Rahman MH (2023) A critical review of whitefly (Bemisia tabaci gennadius) cryptic species associated with the cotton leaf curl disease. J Innov Sci 9:24–43

    CAS  Google Scholar 

  14. Kiobia DO, Mwitta CJ, Fue KG, Schmidt JM, Riley DG, Rains GC (2023) A review of successes and impeding challenges of IoT-based insect pest detection systems for estimating agroecosystem health and productivity of cotton. Sensors 23(8):4127

    Article  ADS  PubMed  PubMed Central  Google Scholar 

  15. Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Boston, MA, USA, 7–12; pp. 1–9

  16. Tripathy S (2021) Detection of cotton leaf disease using image processing techniques. J Phys Confer Ser 2062(1):012009

    Article  Google Scholar 

  17. Naeem A, Bin Senapati B, Chauhan AS, Kumar S, Orosco JC, Abdel-rehim WMF (2023) Intelligent systems and applications. In. Engineering deep learning models for cotton leaf disease detection with VGG-16. p. 0–2

  18. Mahmud M, Kaiser MS, McGinnity TM, Hussain A (2021) Deep learning in mining biological data. In Cognitive Computation (Vol. 13). Springer US

  19. Abade A, Ferreira PA, de Barros Vidal F (2021) Plant diseases recognition on images using convolutional neural networks: a systematic review. Comput Electron Agric 185:106125

    Article  Google Scholar 

  20. Dhaka VS, Meena SV, Rani G, Sinwar D, Ijaz MF, Woźniak M (2021) A survey of deep convolutional neural networks applied for prediction of plant leaf diseases. Sensors 21:4749

    Article  ADS  PubMed  PubMed Central  Google Scholar 

  21. Nagaraju M, Chawla P (2020) Systematic review of deep learning techniques in plant disease detection. Int J Syst Assur Eng Manag 11:547–560

    Article  Google Scholar 

  22. Kamilaris A, Prenafeta-Boldú FX (2018) Deep learning in agriculture: a survey. Comput Electron Agric 147:70–90

    Article  Google Scholar 

  23. Fernández-Quintanilla C, Peña J, Andújar D, Dorado J, Ribeiro A, López-Granados F (2018) Is the current state of the art of weed monitoring suitable for site-specific weed management in arable crops? Weed Res 58:259–272

    Article  Google Scholar 

  24. Lu J, Tan L, Jiang H (2021) Review on convolutional neural network (CNN) applied to plant leaf disease classification. Agriculture 11:707

    Article  Google Scholar 

  25. Golhani K, Balasundram SK, Vadamalai G, Pradhan B (2018) A review of neural networks in plant disease detection using hyperspectral data. Inf Process Agric 5:354–371

    Google Scholar 

  26. Bangari S, Rachana P, Gupta N, Sudi PS, Baniya KK (2022) A Survey on Disease Detection of a potato Leaf Using CNN. In Proceedings of the 2nd IEEE International Conference on Artificial Intelligence and Smart Energy (ICAIS), Coimbatore, India, 23–25; pp. 144–149

  27. Pechuho N, Khan Q, Kalwar S (2020) Cotton crop disease detection using machine learning via tensorflow. Pakistan J Eng Technol 3(2):126–130

    Article  Google Scholar 

  28. Magsi A, Shaikh RA, Shar ZA, Arain RH, Soomro AA (2021) A novel framework for disease severity level identification of cotton plant using machine learning techniques. (05). 367–373

  29. Zhu W, Sun J, Wang S, Shen J, Yang K, Zhou X (2022) Identifying Field Crop Diseases Using Transformer-Embedded Convolutional Neural Network

    Book  Google Scholar 

  30. Li L, Wang B, Hu Z (2022) Classification of plant leaf diseases based on improved convolutional neural network Jie. INMATEH Agric Eng 67(2):553–561

    Article  Google Scholar 

  31. Amin H, Darwish A, Hassanien AE, Soliman M (2022) End-to-End Deep Learning Model for Corn Leaf Disease Classification. IEEE Access 10:31103–31115

  32. Kumbhar S, Nilawar A, Patil S, Mahalakshmi B, Nipane M (2019) Farmer buddy-web based cotton leaf disease detection using CNN. Int J Appl Eng Res 14(11):2662–2666

    Google Scholar 

  33. Sarwar R, Aslam M, Khurshid (2021) Detection and Classification of Cotton Leaf Diseases Using Faster R-CNN on Field Condition Images Acta Scientific AGRICULTURE ( ISSN : 2581–365X ) Detection and Classification of Cotton Leaf Diseases Using Faster R-CNN on Field Condition Images. (September)

  34. Dubey YK, Mushrif MM, Tiple S (2018) Superpixel based roughness measure for cotton leaf diseases detection and classification. In: 2018 4th International Conference on Recent Advances in Information Technology (RAIT). IEEE, pp 1–5

  35. Rai CK, Pahuja R (2023) Northern maize leaf blight disease detection and segmentation using deep convolution neural networks. Multimed Tools Appl 1–18

  36. LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86:2278–2324

    Article  Google Scholar 

  37. Le Cun Y, Jackel LD, Boser B, Denker JS, Graf HP, Guyon I, Henderson D, Howard RE, Hubbard W (1989) Handwritten digit recognition: applications of neural network chips and automatic learning. IEEE Commun Mag 27:41–46

    Article  Google Scholar 

  38. Abdel-Hamid O, Mohamed AR, Jiang H, Deng L, Penn G, Yu D (2014) Convolutional neural networks for speech recognition. IEEE/ACM Trans Audio Speech Lang Process 22:1533–1545

    Article  Google Scholar 

  39. Kamilaris A, Prenafeta-Boldú FX (2018) Disaster monitoring using unmanned aerial vehicles and deep learning. arXiv :1807.11805

  40. Canziani A, Paszke A, Culurciello E (2016) An analysis of deep neural network models for practical applications. arXiv :1605.07678

  41. Schmidhuber J (2015) Deep learning in neural networks: an overview. Neural Netw 61:85–117

    Article  PubMed  Google Scholar 

  42. Alom MZ, Taha TM, Yakopcic C, Westberg S, Sidike P, Nasrin MS, Hasan M, Van Essen BC, Awwal AA, Asari VK (2019) A state-of-the-art survey on deep learning theory and architectures. Electronics 8:292

    Article  Google Scholar 

  43. Bahrampour S, Ramakrishnan N, Schott L, Shah M (2015) Comparative study of deep learning software frameworks. arXiv: 1511.06435

  44. Zhou G, Deng R, Zhou X, Long S, Li W, Lin, G, Li X (2021) Gaussian Inflection Point Selection for LiDAR Hidden Echo Signal Decomposition. IEEE geoscience and remote sensing letters, p 1–5. https://doi.org/10.1109/LGRS.2021.3107438

  45. Qi M, Cui S, Chang X, Xu Y, Meng H, Wang Y, Arif M (2022) Multi-region Nonuniform Brightness Correction Algorithm Based on L-Channel Gamma Transform. Security and communication networks, 2022. https://doi.org/10.1155/2022/2675950

  46. Lin Z, Wang H, Li S (2022) Pavement anomaly detection based on transformer and self-supervised learning. Autom Constr 143:104544. https://doi.org/10.1016/j.autcon.2022.104544

    Article  Google Scholar 

  47. Y., S., N., D., H., T. Z., Y., L., & M., Y (2021) Modeling relation paths for knowledge graph completion. IEEE Transact Knowl Data Eng 33(11):3607–3617. https://doi.org/10.1109/TKDE.2020.2970044

    Article  Google Scholar 

  48. Jiang Y, Li X (2022) Broadband cancellation method in an adaptive co-site interference cancellation system. Int J Electron 109(5):854–874. https://doi.org/10.1080/00207217.2021.1941295

    Article  Google Scholar 

  49. Wang Z, Ma R, Chen B, Yu X, Wang X, Zuo X, Yang J (2024) A transcription factor-based bacterial biosensor system and its application for on-site detection of explosives. Biosens Bioelectron 244:115805. https://doi.org/10.1016/j.bios.2023.115805

    Article  CAS  PubMed  Google Scholar 

  50. Liu H, Xu Y, Chen F (2023) Sketch2Photo: Synthesizing photo-realistic images from sketches via global contexts. Eng Appl Artif Intell 117:105608. https://doi.org/10.1016/j.engappai.2022.105608

    Article  Google Scholar 

  51. Dang W, Xiang L, Liu S, Yang B, Liu M, Yin Z, Zheng W (2023) A feature matching method based on the convolutional neural network. J Imaging Sci Technol. https://doi.org/10.2352/J.ImagingSci.Technol.2023.67.3.030402

    Article  Google Scholar 

  52. Lu S, Yang J, Yang B (2023) Analysis and design of surgical instrument localization algorithm. Comput Model Eng Sci 137(1):669–685. https://doi.org/10.32604/cmes.2023.027417

    Article  Google Scholar 

  53. Cheng Y, Lan S, Fan X, Tjahjadi T, Jin S, Cao L (2023) A dual-branch weakly supervised learning based network for accurate mapping of woody vegetation from remote sensing images. Int J Appl Earth Observ Geoinform 124:103499. https://doi.org/10.1016/j.jag.2023.103499

    Article  Google Scholar 

  54. Zheng H, Fan X, Bo W, Yang X, Tjahjadi T, Jin S (2023) A multiscale point-supervised network for counting maize tassels in the wild. Plant Phenomics 5:100. https://doi.org/10.34133/plantphenomics.0100

    Article  Google Scholar 

  55. Tao Y, Shi J, Guo W, Zheng J (2023) Convolutional neural network based defect recognition model for phased array ultrasonic testing images of electrofusion joints. J Pressure Vessel Technology, 145(2). https://doi.org/10.1115/1.4056836

  56. Zhou G, Xu C, Zhang H, Zhou X, Zhao D, Wu G, Zhang L (2022) PMT gain self-adjustment system for high-accuracy echo signal detection. Int J Remote Sensing 43(19–24):7213–7235. https://doi.org/10.1080/01431161.2022.2155089

    Article  ADS  Google Scholar 

  57. Zhou G, Zhao D, Zhou X, Xu C, Liu Z, Wu G, Zou L (2022) An RF amplifier circuit for enhancement of echo signal detection in bathymetric LiDAR. IEEE Sens J 22(21):20612–20625. https://doi.org/10.1109/JSEN.2022.3206763

    Article  ADS  Google Scholar 

  58. Zhang R, Li L, Zhang Q, Zhang J, Xu L, Zhang B, Wang B (2023) Differential feature awareness network within antagonistic learning for infrared-visible object detection. IEEE Transact Circuits Syst Video Technol. https://doi.org/10.1109/TCSVT.2023.3289142

    Article  Google Scholar 

  59. Li S, Chen J, Peng W, Shi X, Bu W (2023) A vehicle detection method based on disparity segmentation. Multimed Tools Appl 82(13):19643–19655. https://doi.org/10.1007/s11042-023-14360-x

    Article  Google Scholar 

  60. Zhao K, Jia Z, Jia F, Shao H (2023) Multi-scale integrated deep self-attention network for predicting remaining useful life of aero-engine. Eng Appl Artif Intell 120:105860. https://doi.org/10.1016/j.engappai.2023.105860

    Article  Google Scholar 

  61. Hu J, Wu Y, Li T, Ghosh BK (2019) Consensus control of general linear multiagent systems with antagonistic interactions and communication noises. IEEE Trans Autom Control 64(5):2122–2127. https://doi.org/10.1109/TAC.2018.2872197

    Article  MathSciNet  Google Scholar 

  62. Chen B, Hu J, Zhao Y, Ghosh BK (2022) Finite-time velocity-free rendezvous control of multiple AUV systems with intermittent communication. IEEE Transact Syst Man Cybernetics 52(10):6618–6629. https://doi.org/10.1109/TSMC.2022.3148295

    Article  Google Scholar 

  63. Liao Q, Chai H, Han H, Zhang X, Wang X, Xia W, Ding Y (2022) An integrated multi-task model for fake news detection. IEEE Transact Knowl Data Eng 34(11):5154–5165. https://doi.org/10.1109/TKDE.2021.3054993

    Article  Google Scholar 

  64. Ding Y, Zhang W, Zhou X, Liao Q, Luo Q, Ni LM (2021) FraudTrip: Taxi Fraudulent Trip Detection From Corresponding Trajectories. IEEE Internet Things J 8(16):12505–12517. https://doi.org/10.1109/JIOT.2020.3019398

    Article  Google Scholar 

  65. Zhang J, Liu Y, Li Z, Lu Y (2023) Forecast-assisted service function chain dynamic deployment for SDN/NFV-enabled cloud management systems. IEEE Syst J. https://doi.org/10.1109/JSYST.2023.3263865

    Article  Google Scholar 

  66. Xuemin Z, Ying R, Zenggang X, Haitao D, Fang X, Yuan L (2023) Resource-constrained and socially selfish-based incentive algorithm for socially aware networks. J Signal Process Syst. https://doi.org/10.1007/s11265-023-01896-2

    Article  Google Scholar 

  67. Lyu T, Xu H, Zhang L, Han Z (2023) Source selection and resource allocation in wireless powered relay networks: an adaptive dynamic programming based approach. IEEE Internet Things J. https://doi.org/10.1109/JIOT.2023.3321673

    Article  Google Scholar 

  68. Xu H, Han S, Li X, Han Z (2023) Anomaly traffic detection based on communication-efficient federated learning in space-air-ground integration network. IEEE Trans Wireless Commun. https://doi.org/10.1109/TWC.2023.3270179

    Article  Google Scholar 

  69. Chen J, Wang Q, Peng W, Xu H, Li X, Xu W (2022) Disparity-based multiscale fusion network for transportation detection. IEEE Trans Intell Transp Syst 23(10):18855–18863. https://doi.org/10.1109/TITS.2022.3161977

    Article  Google Scholar 

  70. Ma X, Dong Z, Quan W, Dong Y, Tan Y (2023) Real-time assessment of asphalt pavement moduli and traffic loads using monitoring data from Built-in Sensors: optimal sensor placement and identification algorithm. Mech Syst Signal Process 187:109930. https://doi.org/10.1016/j.ymssp.2022.109930

    Article  Google Scholar 

  71. Jin K, Gao Z, Jiang X, Wang Y, Ma X, Li Y, Ye J (2023) MSHF: A Multi-Source Heterogeneous Fundus (MSHF) Dataset for Image Quality Assessment. Sci Data 10(1):286. https://doi.org/10.1038/s41597-023-02188-x

    Article  PubMed  PubMed Central  Google Scholar 

Download references

Acknowledgements

The authors express their gratitude to Huanggang Normal University for supporting this research, and to King Saud University for funding it through Researchers Supporting Program number (RSP2024R206), King Saud University, Riyadh, Saudi Arabia.

Funding

This research is supported by the 2022 Hubei Provincial Department of Education Philosophy and Social Science Research Project (22D117).

Author information

Authors and Affiliations

Authors

Contributions

Rubaina Nazeer and Sajid Ali: Led the study’s design, methodology, and manuscript writing. Hu Zhihua: Focused on data analysis and interpretation. Ghulam Jillani Ansari: Handled data collection and preprocessing. Muna Al-Razgan: Contributed to the technological aspects and tools used in the research. Emad Mahrous Awwad: Assisted in literature review and validation of results. Yazeed Yasin Ghadi: Involved in editing and finalizing the manuscript.

Corresponding author

Correspondence to Zhihua Hu.

Ethics declarations

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Nazeer, R., Ali, S., Hu, Z. et al. Detection of cotton leaf curl disease’s susceptibility scale level based on deep learning. J Cloud Comp 13, 50 (2024). https://doi.org/10.1186/s13677-023-00582-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13677-023-00582-9

Keywords