225
Views
0
CrossRef citations to date
0
Altmetric
Food Science & Technology

A multimodal fusion framework to diagnose cotton leaf curl virus using machine vision techniques

, ORCID Icon &
Article: 2339572 | Received 26 Jun 2023, Accepted 02 Apr 2024, Published online: 26 Apr 2024

Abstract

Cotton diseases are disastrous for quality and sustainable production of the yield. Cotton leaf curl virus (CLCuV) is one of the most damaging diseases for cotton crops. Symptoms-based CLCuV identification is tedious, time consuming, error prone and needs exceptional expertise. Sensor-based machine vision approaches have great potential to detect the CLCuV at early stages. This research study proposes a machine vision-based multimodal fusion framework to diagnose various CLCuV severity levels. Our designed model is based on three contrasting datasets: digital photographic, multispectral and fused datasets. A digital camera was used to acquire the digital photographic dataset, the multispectral dataset was obtained by a multispectral radiometer-5 (MSR5), and the two datasets were fused to formulate the third one. From the digital photographic dataset, 269 texture features were extracted and optimized to the most discriminant 30 texture features, the multispectral dataset consisted of 5 spectral features, and the fused dataset was formed by combining the two. The 30 most discriminant features from the digital photographic dataset were selected by incorporating fisher co-efficient, probability of error plus average correlation and mutual information (MI). To diagnose CLCuV, four machine-learning classifiers, namely simple logistics (SL), multilayer perceptron (MLP), sequential minimal optimization (SMO), and random forest (RF), were deployed separately on each dataset. The maximum CLCuV diagnosing accuracies attained from digital photographic, multispectral, and fused datasets were 81.263%, 91.177%, and 96.313%, respectively.

1. Introduction

During the last four decades, machine vision approaches have been applied sizably in many areas, such as automatic inspection, process control, assessing yield production, disease detection, etc. (Radhakrishnan, Citation2012). Disease detection in plants is a crucial process. Detection of plant leaf diseases based on visual assessment is challenging and requires considerable expertise and domain knowledge but is still error-prone. It can be successfully achieved by machine vision (Pantazi et al., Citation2019). Disease detection in crops at an early stage increases advantages and reduces losses. Machine learning and deep learning data analysis methods accurately detect crop diseases (Shahi et al., Citation2023). Bacterial, fungal, and viral infections of insects instigate plant diseases. Viral diseases affect the plant’s natural growth, infections, or mechanical injury to leaves or plants (Kerkech et al. Citation2018). Nevertheless, cotton plants are also affected by diseases in seeding, cotton bolls, cotyledons, and mature cotton leaves (Hassan et al., Citation2016; Prajapati et al., Citation2016). CLCuV is a damaging disease and a serious threat to cotton crops. Plants infected by CLCuV disease exhibit various symptoms due to their severity levels. The symptoms of CLCuV appear in photosynthate blockage, causing tertiary veins to be yellowish and thicken. Later, leaf curling and blockage of secondary veins reduce the leaf surface area for photosynthesis and block the passage of photosynthesis. Consequently, the infected plant seems to be dwarf. Leaves are curled downward or upward; growth of the leaves is also stunted due to inter-nodal distance reduction in severe conditions, and an outgrowth called enation appears on the lower side of a curled leaf (Saeed et al., Citation2018).

The evolutionary story of CLCuV disease in cotton crops is fascinating. The disease was first recorded in 1912 in Nigeria on Gossypium peruvianum, in 1924 in Sudan, in 1926 in Tanzania and in 1959 in the Philippines. CLCuV was reported in Pakistan for the first time in 1967 in District Multan of Punjab. Similar symptoms were reported on tobacco and other plants much earlier than 1967 in Pakistan. CLCuV was not measured as a serious disease until 1987, but in the early 1990s, it was assumed to be an epidemic proportion in Punjab. A huge loss of 1.3 million bales of cotton over 24.28 hectares was recorded in Punjab. Sindh remained a free province till 1997. CLCuV was first observed in Sindh in 1997 and then varied from area to area, year-wise. Therefore, it is very hard to estimate the loss caused by CLCuV overall in Pakistan (Farooq et al., Citation2014).

Cotton is considered a cash and fiber crop in Pakistan and is called ‘White Gold’ or ‘king of fibers’. Pakistan is in 4th position in the world for the production of cotton. Approximately, 55% of Pakistan’s foreign exchange earnings are due to cotton. Two provinces, Punjab and Sindh, are growing nearly 65% of Pakistan’s cotton. More than 15% of the cultivated area is devoted to cotton, and almost 26% of farmers in Pakistan grow a cotton crop. 4.5% of the agriculture gross domestic product (GDP) is added by cotton, which is 0.8 percent of the GDP. The country’s agro-industrial sector provides 17 percent employment opportunities (Rana et al., Citation2020). Cotton production in Pakistan has decreased during recent decades due to several factors, including the CLCuV, insect pests and various abiotic strains. CLCuV is a highly destructive plant disease that significantly reduces Pakistan’s ability to produce cotton (Akhtar et al., Citation2004).

One of the major diseases that cause economic and production loss in Pakistan and India is cotton leaf curl disease (CLCuD). CLCuD belongs to the Geminiviridae and is caused by a monopartite plant virus. Disease detection in plants requires expert opinions, as little variation in symptoms can cause economic and production loss. Based on the inappropriate diagnosis of diseases, farmers must input pesticides, resulting in an incredible amount of economic and production consequences. Plant disease diagnosis at the manual level can be error-prone and costly. The computer is used for automatic disease detection in the current information age (Zubair et al., Citation2017). Symptom-based detection of plant leaf diseases remains a headache due to its complexity. Many experienced plant pathologists and agronomists need help accurately diagnose a specific disease. Detection of plant diseases based on visual assessment is prone to error. Therefore, a farmer has to bear a large economic and production loss (Ferentinos, Citation2018). In the case of large-scale cultivation, it is very hard for farmers and experts to visit the field and observe each infected plant physically. Proper treatment can be made timely for accurate detection and diagnosis of plant diseases at the initial stage by deciding the right pesticide. Ultimately, economic and production loss can be avoided (Ilic et al., Citation2018).

The human eye cannot see within a visual spectral range of 400–700 nm, while remote sensing (RS) devices can measure radiations in numerous parts of the Light spectrum ranging from 10 to 255 nm (Nilsson, Citation1995). RS measures electromagnetic radiations that are emitted or reflected from an object. Numerous non-invasive procedures, like RS, imaging-based and spectroscopy-based, provide precise and reliable support to detect and monitor diseases in the plant in a large-scale and real-time environment. In addition, digital imaging sensors and hyperspectral or multispectral sensors are also used to detect plant diseases. Multispectral sensor normally works on an electromagnetic spectrum range of 400–2500 nm. These wavebands can easily capture a minor change in plant pathology by its diseases at their early stages (Zhang et al., Citation2020). Data Fusion combines data and information from multiple sources to produce more accurate results than separately using data sources (Ouhami et al., Citation2021). Data fusion can be classified into three major categories. 1) Observation level fusion; 2) Feature level fusion; 3) Decision level fusion. Observation-level fusion is used to combine raw data from two sources. Feature level fusion is used to integrate extracted features of different sources, while decision level is used to incorporate classification results achieved by each source alone to reach a common final decision (Schmitt & Zhu, Citation2016).

Researchers used deep learning and support vector machine (SVM) technologies to recognize patterns and detect four types of rice leaf diseases by using healthy and infected images. Features of infected rice leaf images were extracted using a convolutional neural network (CNN). After feature extraction, the SVM model was used to classify and predict the infected rice leaf. Pattern recognition accuracy was improved to 96.8% when both deep learning and SVM were combined (Jiang et al., Citation2020). Machine learning algorithms and hyperspectral imaging (HSI) techniques were used to detect tomato spotted wilt virus (TSWV) infection in tobacco at its early stages. A spectral range from 400 to 1000 nm with 28 bands was used. Four machine learning techniques, RF, SVM, classification and regression tree (CART) and boosted regression tree (BRT), and three wavelength selection methods (BRT, successive projection algorithm (SPA) and genetic algorithm (GA)) were used. BRT algorithm performs best with the SPA wavelength feature selection technique at 10-fold cross-validation and achieves 85.2% accuracy. Results also showed that Healthy and infected tobacco were differentiated by near-infrared (NIR) spectral region (Gu et al., Citation2019).

An approach was demonstrated to identify leaf diseases from different crop species. Various leaf images of species were used. One class classifier was deployed to classify powdery mildew, downy mildew, black rot and healthy images, while local binary patterns (LBPs) were used for feature extraction. The algorithm achieved a 95% success rate when tested on 46 plants while trained on vine leaves only (Pantazi et al., Citation2019). CNN was developed to detect and diagnose plant diseases using healthy and diseased plant images with deep learning models. 87,848 images of 25 different plants with 58 distinct classes from the open database were used for training. Training several architectures with plant disease combinations achieved a high success rate. VGG-CNN achieved 99.53% accuracy. Due to the 99.53% success rate with very deep neural network VGG, the suggested model could be used as a warning and advisory tool to identify diseases (Ferentinos, Citation2018).

Dielectric spectroscopy was suggested to detect basal stem rot (BSR) in oil palm leaves using electric properties like dielectric constant, impedance, capacitance and dissipation factor. Four types of leaf samples (healthy, mild, moderate and severely infected) were collected, and a solid test fixture was used to classify oil palm leaves by BSR. RF, support vector machine-feature selection (SVM-FS) and GA were used to analyze the dataset with significant electrical properties. Significant frequencies 100 kHz–30 MHz with 300 spectral intervals were also selected. Artificial neural network (ANN) and SVM were used to determine the classification accuracy. The SVM-FS model showed 88.64% accuracy, while the ANN classifier produced results with 80% accuracy (Khaled et al., Citation2018).

CLCuV can be predicted and managed through bio-product and resistant germplasm. Every year, there is significantly reduced cotton yield due to CLCuV. Multiple regression analysis evaluated 15 varieties to predict disease incidence abiotic environmental variables like humidity, minimum air temperature, maximum air temperature and rainfall. Two bio-products were used against the whitefly population to control the disease. No one variety was found out of the fifteen highly resistant against CLCuV (Saeed et al., Citation2018). A deep learning approach was developed for detecting vine diseases using unmanned aerial vehicle (UAV) images with vegetation and colorimetric indices. Early detection of symptoms prevents serious disease attacks in grape leaves, and disease management can be addressed better. Visible domain was used in UAV images. Color information and CNN were used to detect disease symptoms in vineyards. Different vegetation indices, color spaces, and combinations of both were used to compare the performance of CNN. The best results were achieved with 95.8% accuracy when CNN was combined with YUV color space, ExG, ExR and ExGR vegetation indices (Kerkech et al., Citation2018).

A deep convolutional neural network (DCNN) model was proposed to diagnose cucumber diseases using leaf symptoms like Powdery mildew, downy mildew, target leaf spots and anthracnose. The segmentation was performed on cucumber leaf images. The data augmentation method increased the dataset to minimize the chance of data overfitting. Fourteen thousand two hundred eight symptoms-based images were used in the dataset. AlexNet, SVM and RF classifiers were used to compare results obtained by DCNN. DCNN has proven to be the best tool for recognizing cucumber diseases, with an accuracy of 93.4% (Ma et al., Citation2018). A technique was developed for detecting citrus diseases based on symptoms by textural descriptors and color histograms. A color difference algorithm ΔE was used to classify the healthy and infected areas. For feature reduction, principal component analysis (PCA) was used, and state-of-the-art classifiers like K-Nearest Neighbor (KNN), SVM and Bagged Tree Ensemble (BTE) classifier were used. BTE outperformed with 99.9% accuracy results (Ali et al., Citation2017).

Bacteria Blight (BB) and CLCuV were detected among healthy and unhealthy leaves by RESNET50, a transfer learning algorithm, and KNN, a machine learning algorithm. RESNT50 obtained 95% accuracy, while KNN outperformed with 98% accuracy (Kotian et al., Citation2022). A framework was implemented to diagnose BB, Ball Rot (BR) and CLCuV using SVM, KNN, Naïve Bayes (NB) and Inception V4 based on CNN. Data was collected from Sindh, Pakistan. Inception V4-based CNN achieved an accuracy of 98.26% (Anwar et al., Citation2022). A deep learning model, MobileNet, Snapshot Ensemble (SE) and Vision Transformers (VT), was deployed to diagnose real-time CLCuV, BB and Fusarium Wilt (FW), three common diseases of cotton leaf. The suggested framework was designed to detect early stage diseases to overcome yield loss. It was observed that VT is less efficient, with an accuracy of 95.29%, MobileNet achieved 96.74% and the SE technique got 97.21% accuracy (Gudela et al., Citation2023).

A machine vision spectral statistical framework was suggested to classify different types of land cover data, such as Sutlej River land, fertile cultivated land, green pasture, desert range land and bare land. Multispectral radiometer-5 (MSR5), a handheld device with 5 spectral bands (red, green and blue [RGB], shortwave infrared and NIR), was used to acquire land cover data, while a digital camera was used to acquire texture data. Digital images were transformed into 229 texture features. Three statistical feature selection methods, MI, Fisher and (PA + ACC), were used to obtain the most discriminant features. Multispectral data was verified by linear discriminant analysis (LDA), while non-linear discriminant analysis (NDA) was applied to selected texture data. An ANN: n-class was used to classify multispectral and texture data. A cross-validation of 80:20 was implemented. 91.32% and 96.40% accuracy for texture and multispectral data were achieved, respectively (Qadri et al., Citation2016). Data fusion is applied to texture and multispectral data. MLP, NB, j48 and RF machine learning classifiers were used to achieve the best results. An accuracy of 99.60% is obtained using an MLP classifier on a fused dataset (Qadri et al., Citation2017).

Machine learning techniques were implemented to diagnose cotton crop diseases like BB, CLCuV, Whitefly infection (WI) and Jassid at early stages. Three rounds were used with different training and testing capacities. A TensorFlow classifier got 86.3%, 88.6% and 91% accuracy in the first, second and third rounds, respectively (Pechuho et al., Citation2020). A DJANGO framework was implemented, using VGG16 and VGG19, to detect multiple plant leaf diseases like Bacterial spot (BS), CLCuV, BB and Tomato mosaic virus (TMV). VGG16 got an accuracy of 86.32%, while VGG19 got an average of 93.61% accuracy (Gelli et al., Citation2020). A Rice-Fusion framework was proposed to diagnose rice diseases. A camera and agrometeorological sensor collected the dataset. The dataset contains 3200 healthy rice images. In the Rice-Fusion framework, numerical features were extracted from the agrometeorological dataset, and visual features were extracted from the image dataset. An accuracy of 95.31% was achieved by fusing numerical data with visual features as opposed to the unimodal framework, which had 91.25% and 82.03% accuracy based on MLP and CNN Architecture (Patil & Kumar, Citation2022).

Data fusion strategies have great potential to detect rice diseases like leaf blight, rice blast and rice sheath blight by using three spectroscopic techniques, including NIR HSI spectra, mid-infrared spectroscopy (MIR) and Laser-induced breakdown spectroscopy (LIBS). Rice diseases were classified by fusing different types of spectral features. Features were extracted by PCA and autoencoder algorithms. Raw data fusion, feature fusion and decision-level fusion were used. MLP and CNN models were used to identify rice diseases. CNN achieved an accuracy of 93%. HSI-based models performed better outcomes than MIR and LIBS (Feng et al., Citation2020).

It has been concluded that most authors identified leaf diseases using machine learning classifiers such as SVM to discriminate healthy and infected leaves with notable accuracy. Others applied multiple classifiers, such as RF, KNN, NB and MLP, to classify the healthy and infected leaves, using CLCuV, BB, FW and BR diseases as the infected leaves dataset. Some researchers also experimented with deep learning models like AlexNet, VGG, RESNET50 and Inception V4, classifying infected and healthy ones and showing suitable results. The survey also revealed that almost all the research was to discriminate between healthy and infected leaves. If some research studies used multiple diseases to represent the infected leaves, they only collect them as a single dataset of unhealthy leaves. It is also to be noted that few research studies used feature optimization, such as LDA, PCA and RDA. Finally, the survey showed that data fusion needed to be improved. The summary of the related work is given in .

Table 1. Summary table of related work.

This research work is instanced and based on three research contributions. First, Smruti, with colleague researchers (Kotian et al., Citation2022), extracted texture features from a digital photographic dataset to identify CLCuV. Second, Qing, with his fellow researchers (Gu et al., Citation2019), used the multispectral dataset to identify the curl virus, and finally, Rutuja, with colleague authors (Patil & Kumar, Citation2022) applied data fusion approach to identify the leaf disease. It is examined that most researchers performed bi-class classification, i.e. just discriminating the infected leaves from healthy ones. Rare work was found on CLCuV, though few efforts were noticed to diagnose CLCuV and a healthy one. No research work was conducted to discriminate the severity levels of CLCuV using multiple and fused datasets and hybrid feature selection in the selected regions. These indicated parameters clearly argue the novelty of this research study.

It is analyzed that cotton production is dominant for the economy of any cotton-yield country, but it is essential for developing countries like Pakistan. Cotton crops may be damaged substantially by many harmful diseases. Late disease identification is disastrous, and we need efficient and timely disease detection to set aside the loss. Human eye observation of diseases needs to be reassessed, whereas machine vision approaches have great potential to diagnose diseases accurately and efficiently. This leads to many practical benefits, like minimizing the loss of yield production and conquering economic damage. CLCuV is a destructive disease for cotton yields, and early and accurate detection of CLCuV is inevitable. This research study proposes a machine-based solution to the problem by developing a multimodal fusion framework to discriminate various CLCuV severity levels by incorporating three different CLCuV datasets, which include digital images, multispectral and fused datasets. The designed framework comprised data acquisition, image preprocessing, feature extraction, feature selection, classification and evaluation.

2. Material and methods

This research study designs an automated system for early detection of the CLCuV using image processing and RS technologies. Digital photographic data was captured for texture features, while radiometric/RS data was acquired for spectral features. The digital Conon camera model IXUS 185, which has 20.00 megapixels, was used for digital photographic data. A CROPSCAN device named MSR5 was used to capture radiometric data. Data was collected from Bahawalpur (29° 23′ 44′′ (North) latitude and 71° 41′ 1′′ (East) longitude) and Multan (30° 15′ 75′′ (North) latitude and 71° 52′ 49′′ (East) longitude) Division of Punjab, Pakistan as Punjab is the province producing more than 70% cotton of the country.

2.1. Proposed methodology

Four levels were used to accomplish this study, out of which one is healthy, and three are maximum, medium and minimum affected by CLCuV. No specific experimental setup was required for this study. A digital photographic dataset was collected from a cotton field using a digital camera. The multispectral dataset was acquired from the same cotton fields used to collect digital photographic datasets. The novel multimodal fusion framework to diagnose the CLCuV is shown in .

Figure 1. Proposed multimodal fusion framework to diagnose CLCuV.

Figure 1. Proposed multimodal fusion framework to diagnose CLCuV.

Image processing software Irfan’s view, CVIP Tools version 5.7e, built at Southern Illinois University Edwardsville, United States, MaZda version 4.6, created at the Institute of Electronics, Technical University of Lodz, Poland, and data mining tool WEKA version 3.8.3, developed at the University of Waikato, New Zealand, were used on Intel® Core i3 2.4 GHz processor, 8 GB RAM with 64-bit Microsoft® Windows version 8.1 Pro operating system to implement the proposed framework (Tariq et al., Citation2020) (Szczypiński et al. Citation2009) (Aher and Lobo Citation2011). The steps used to implement the proposed framework are discussed below.

2.2. Photographic data acquisition

Digital photographic images of different cotton varieties were captured by a digital Canon IXUS 185 model camera with 20.00 Megapixel with four severity levels (healthy, maximum effected, medium effected and minimum effected) of CLCuV, which are shown in . A total of 600 digital photographic images with dimensions 5152 × 3864 and 24-bit depth of joint photographic expert group (jpg) format are captured for each severity level.

Figure 2. Healthy, maximum effected, medium effected and minimum effected leaf images.

Figure 2. Healthy, maximum effected, medium effected and minimum effected leaf images.

All images were acquired in an open environment at 39 °C–44 °C in August and September 2020 from 12:00 PM to 3:00 PM in GMT + 5 Time Zone. There was no shadow at the image-capturing time. For each severity level, 100 digital photographic images were selected from 600 digital images by consulting with domain experts. Luxmeter Extech LT300 is used to measure sunlight intensity levels during data acquisition. Sunlight intensity with time is shown in .

Table 2. Time and sunlight intensity information.

2.3. Remote sensing data acquisition

RS data of cotton plants with four severity levels, healthy, maximum effected, medium effected, and minimum effected, was gathered by MSR5 with serial no. 566 is made by CROPSCAN, Inc. USA. MSR5 has five bands, three visible (blue, green and red). Two are invisible (NIR and shortwave infrared) bands, as shown in . Five hundred ninety-four MSR scans for each severity level were captured at a distance of 5 feet in August and September 2020 at 39 °C–44 °C from 12:00 PM to 3:00 PM in GMT + 5 time zone. Each MSR scan covers a half area 2.5 feet of its height.

Table 3. Multispectral feature table.

2.4. Image preprocessing

Six hundred colored digital photographic images for each category were acquired from twelve different cotton fields, each with an area of 43,560 square feet. The 100 best images were selected by consulting with experts and plant pathologists to obtain better and more accurate results. These selected images contain irrelevant areas, so captured images were cropped to various sizes from 3215 × 2776 to 5152 × 3826 pixels to get prominent leaves only. These cropped images were then resized to 64 × 64, 128 × 128, 256 × 256 and 512 × 512 pixels for each severity level and stored in bitmap (.bmp) image format. This study uses a digital photographic dataset of 400 × 4 = 1600 images to diagnose different severity levels of CLCuV. Each pixel requires 8 × 3 = 24 bits of memory to process each image in the RGB color model. So, each color image is transformed into grayscale to occupy only 8 bits (Yan et al., Citation2016). Grayscale images were enhanced at contrast level, and salient features like shadow and sharpness were restored. MaZda version 4.6, an image texture analysis tool, was used to obtain texture features. It is an efficient and reliable tool for image segmentation, texture feature computation, feature selection and feature extraction (Szczypiński et al., Citation2009).

The region of interest (ROI) of three-pixel dimensions for each image size was used to obtain better results. Digital photographic images were captured from 3 to 4 feet in height, depending upon the cotton plant size so that the whole cotton plant could be observed. Acquired images may contain many irrelevant areas, so large and equal pixel dimensions ROI cannot be used for all images. Therefore, various ROI sizes were used on an image dataset. ROI sizes 8 × 8, 10 × 10, and 12 × 12 were used for 64 × 64 image size. ROI sizes 12 × 12, 14 × 14, and 16 × 16 were used for 128 × 128 image size. ROI sizes 16 × 16, 18 × 18, and 20 × 20 were used for 256 × 256 size image. ROI sizes 20 × 20, 22 × 22, and 24 × 24 were used for 512 × 512 image size. The digital photographic dataset contains 1600 × 3 = 4800 images in this study. ROIs of 512 × 512 image size for various severity levels are shown in .

Figure 3. ROIs of healthy, maximum effected, medium effected and minimum effected leaf.

Figure 3. ROIs of healthy, maximum effected, medium effected and minimum effected leaf.

2.4.1. Feature extraction

In feature extraction, raw data is altered into numerical values while original dataset information is preserved. Extracted features have the potential to produce better results than processing raw data using machine learning classifiers (Kumar & Bhatia, Citation2014). Five non-overlapping sub-images or ROIs from each image were drawn to obtain texture features. The digital photographic dataset used a total of 4800 × 5 = 24,000 ROIs or sub-images/segments. For each ROI, 9 first-order histogram features, 11 second-order co-occurrence features with 5 inter-pixel distances in four directions (0°, 45°, 90° and 135°) (11 × 5 × 4 = 220), 5 second-order run-length features in four directions (0°, 45°, 90° and 135°) (5 × 4 = 20) and 20 wavelet features, a total of 9 + 220 + 20 + 20 = 269 features were extracted. The total feature vector space is 24,000 × 269 = 6,456,000, while 538,000 features were used for the same ROI size in all severity levels. Three-pixel dimensions for each image size were used to obtain better results, and the dataset was also increased.

2.4.1.1. First-order histogram features

The first-order histogram or statistical feature calculates the intensity of individual pixels. Histogram features are popular due to their simplicity and compact performance. Data analysis based on histogram features provides the best solution for image processing problems to recognize and classify objects (Blachnik & Laaksonen, Citation2008). In this study, nine histogram features were calculated for each sub-image or ROI. Probability and various features of first-order histogram are described by EquationEquations (1)–(5). (1)  P(α)=S(α)T(1)

where S(α) is a measure of pixels in gray level image and T is number of pixels in an image or sub image (ROI). First-order histogram features like mean, variance, skewness, kurtosis, etc., are explained below.

Mean is average value of light and dark brightness in an image or sub-image (ROI) (2) Mean=xyZ(c,d)M(2)

Variance is used to measure pixel value to the mean pixel value. (3) Variance=1Spt=0SpGrG¯2(3)

Skewness is the degree of asymmetry about the mean value. (4) Skewness=1Spt=0SpGrG¯31Spt=0SpGrG¯23(4)

Kurtosis is the measure of peakedness of distribution in an image or sub-image (ROI) (5) Kurtosis=1Spt=0SpGrG¯41Spt=0SpGrG¯22(5)

2.4.1.2. Second-order texture features

Second-order texture features calculate the co-occurrence and spatial inter-dependency of two neighboring pixels. Gray-level co-occurrence matrix (GLCM) and run length matrix (RLM) methods commonly calculate second-order texture features. Second-order texture features give more accuracy in the classification process than first-order histogram features (Feng et al., Citation2015). Eleven GLCM features up to the 5-pixel distance, and five RLM features in all four directions (0°, 45°, 90° and 135°) were calculated for each ROI.

GLCM features like contrast, correlation, entropy, entropy difference, sum entropy, etc., are described below and showed by EquationEquations (6)–(15).

Contrast is the measure of variation in local intensity of pixels. (6) Contrast=x=1Zay=1Za(xy)2S(x,y)(6)

Correlations are used to determine similarity among pixels. (7) Correlat=1SxSyxyLωxMωyx,y(7)

Entropy is the degree of randomness and contained the average information of an image or sub-image (ROI). (8) Entrop=xy(x,y)log2(x,y)(8)

Difference entropy is used to compute neighborhood intensity value difference variability. (9) Diff_Entropy=x=0Za1Sab(x)log2(Sab(x)+(9)

Sum entropy is used to measure sum of intensity value difference of neighbor pixels. (10) Sum_Entropy=x=22ZcSa+b(x)log2(Sa+b(x)+(10)

RLM features are used to compute the length or range of run. Gonzales described length or range as a grayscale section or the same color level in a particular direction for a specified θ. RLM features like run length non-uniformity (RLNU), run length non-uniformity normalized (RLNN), long run emphasis (LRE), short run emphasis (SRE) and run percentile (RP)/fraction, etc., are described below.

RLNU measures similarity index among run length of an image, high homogeneity factor contains low RLNU value. (11) RLNU=x=1Zay=1Zbx,yσ2Zbσ(11)

RLNU is normalized to RLNN with substantial enhancement. RLNN also measures the similarity index among run length of an image, high homogeneity factor contains high RLNN value. (12) RLNN=y=1Zax=1Zbx,yσ2Zbσ2(12)

LRE computes the distribution of long run length. High value of LRE indicates the coarser texture. (13) LRE=y=1Zax=0Zbx,yσ2Zbσ(13)

SRE computes the distribution of short run length. High value of SRE indicates finer texture. (14) SRE=x=0Zay=1Zb(x,y |σ)x2Zb(σ)(14)

RP or fraction computes the coarseness of texture. (15) R P=ZzZb(15)

2.4.1.3. Wavelet features

A signal is decomposed into a spatial domain with finite energy called wavelet transform. The wavelet feature is widely used in texture classification to measure energy distribution (Huang & Aviyente, Citation2008). Many researchers use wavelet features to compress and denoise an image, pattern recognition and image analysis (Barbhuiya & Hemachandran, Citation2013). For each ROI, 20 wavelet features were calculated. In discrete wavelet transform 2 M × 2 M matrix is converted into same dimension matrix. Energy can be measured by EquationEquation (16). (16) Energy=φ1a,bϵROIxa,b2(16)

2.4.2. Feature selection

A dataset with hundreds of thousands of features increases time and complexity to compute results. In the machine learning process, too many features cause a dimension disaster. The curse of dimensionality produces a bad impact on classification results. So, high-dimensional data, which may contain many irrelevant and redundant features that overfit in machine learning methods, is reduced to low-dimensional data to obtain efficient and prominent features using feature selection to overcome such a problem (Lutu & Engelbrecht, Citation2010). Feature selection is considered a critical process in machine learning applications. Feature selection methods are classified into filter, embedded and wrapped methods. Filter feature selection methods select features with a high correlation score to other outcome features. No dependency exists between selected features and machine learning classifiers. Filter feature selection methods are low-cost computational methods. Embedded feature selection methods select features during the execution of machine learning classifiers. Such methods are highly dependent on classifiers and have too much computational cost than other methods. Wrapper feature selection methods select features that are dependent on specific classifiers. A dependency exists between selected features and machine learning classifiers. The computational cost of such methods is much higher than filter feature selection methods (Jovic et al., Citation2015).

The digital photographic dataset has 269 × 400 × 5 = 538,000 extracted features for the same ROI size in all severity levels, which is a large-scale dataset and requires high computational power, time, and memory. All 269 extracted features of each ROI are not equally significant. The most significant features are extracted by fisher co-efficient (F), probability of error plus average correlation coefficient (POE + ACC = PA) and mutual information coefficient (MI) filter feature selection techniques to obtain more accurate results. MaZda combines three approaches (F + PA + MI) and gives the thirty most significant features for each ROI, ten features by each method, reducing the dataset to 30 × 400 × 5 = 60,000 features. Most significant thirty features selected by (F + PA + MI) are shown in .

Table 4. Feature selection (F + PA + MI) for ROIs (512 × 512 image size).

Mathematically, F, PA and MI formulation and description are given below.

Fisher coefficient (F) calculates the ratio between inter-class and intra-class variation, is given by EquationEquation (17). (17) Fisher=M2N2=11x=1XOx2  x=1Xy=1XOxOyPxPy2x=1XOxNx2(17)

Probability of error plus average correlation coefficient (POE + ACC = PA) analyze ratio between misclassified samples and total samples in dataset, is described by EquationEquations (18)–(21). (18) POE(Xm)=Mis_classified_samplesTotal samples(18) (19) Xj=Xm:minm [ POE (Xm)+|Corrlt (Xi,Xj)|](19) (20) Xn=Xm:minm [ POE (Xm)+1T1|Corrlt (Xi,Xm)|](20) (21) S=Xl:minl [ Ui POE (Xl)+Uj ACC (Xl)](21)

MI compute dependency between two variables, explained in EquationEquation (22). (22) MI (Si,J)=J=1Can=1DbP (SiJ,Xn)log2[P(SiJ,Xn)P (SiJ)P(Xn)] (22)

2.4.3. Classification

A supervised machine learning process in which input data is characterized into classes based on one or more variables is called classification (Lever et al., Citation2016). Waikato environment for knowledge analysis (WEKA) is used for data analysis and predictive modeling, a machine learning application developed by the University of Waikato, New Zealand. It accepts attribute relation file format (ARFF) and provides many classifiers (Aher & Lobo, Citation2011). Many ML classifiers were used to detect CLCuV, but SL, MLP, SMO and RF outperformed for digital photographic, spectral, and fused datasets. SL is the most common and fundamental ML classifier, which is easy to implement (Ray, Citation2019). MLP is a powerful multi-layered neural network, an effective classifier for complex and large datasets (Feng et al., Citation2020). RF is another widely used ML classifier in classification problems, which computes the final result by combining the results from multiple sub-decision trees. It is a flexible and easy-to-implement classifier (Ma et al., Citation2018). SMO is one of the compelling ML classifiers designed particularly for sparse datasets (Wang et al., Citation2018). It is based on the SVM model but also differs concerning the requirement of quadratic programming. The hyperparameters of deployed ML classifiers are shown in .

Table 5. Hyperparameters of ML classifiers.

2.4.4. Performance evaluation

Various evaluation parameters are used to identify the performance of proposed model on given dataset. These performance evaluation parameters are explained by EquationEquations (23)–(29) below.

Kappa statistics (KS) is a measure of agreement between two or multiple classes. (23) KS=So Se 1Se  (23)

Mean absolute error (MAE), is an absolute difference, which is a positive value, between actual and predicted value measured by a model. (24) MAE=1SGG¯(24)

Root mean squared error (RMSE) measures the average difference between actual and predicted value and calculates the square root of that average. (25) RMSE=t=1sytyt¯2S(25)

True positive (TP) rate or sensitivity, is a measure of perceived positive prediction with total positive prediction. (26) TP rate=TPTP+FN(26)

False positive (FP) rate, is a ratio of negative predictions that are wrongly categorized as positive predictions. (27) FP rate=FPFP+TN(27)

F-measure/F-score measures the accuracy of classification model on a given dataset. (28) FScore=2×precision×recallprecision+recall(28)

Mathew’s correlation coefficient (MCC) is used to measure correlation co-efficient between predicted and actual value. (29) MCC=TN×TPFN×FP(TP+FP)(TP+FN)(TN+FP)(TN+FN) (29)

Receiver-operating characteristics (ROC), are used to measure the performance of each class in classification model.

Precision recall curve (PRC), is used to summarize the trade-off between TP rate and FP rate.

3. Results and discussion

Four supervised machine learning classifiers named SL, MLP, SMO and RF were used to detect various severity levels (healthy, maximum effected, medium effected and minimum effected) of CLCuV at digital photographic, spectral, and fused dataset with k-fold cross validation where k is set to 10. All employed classifiers compute the average result of all turns by splitting each dataset into ten folds. In each turn, one is used for testing, and the remaining nine are used for training. All deployed ML classifiers give unsatisfactory accuracy results for less than 512 × 512 images with various ROI sizes. As the image size increased to 512 × 512 with ROI size of 24 × 24, more promising results were achieved. Surprisingly, Results accuracy decreased as the image size increased to 1024 × 1024 with different ROI sizes. Performance evaluation parameters like KS, MAE, RMSE, TP rate, FP rate, F-measure/F-Score, MCC, ROC and PRC were calculated to achieve accuracy. All accuracy tables and confusion matrices (CM) were computed based on averages of 10-fold results.

In the first step, ML classifiers SL, MLP, SMO and RF were deployed on a digital photographic dataset. ML classifiers were implemented on 64 × 64, 128 × 128, and 256 × 256 image sizes with various ROIs feature datasets, but promising results were not found and have result accuracy less than 80%. The same ML classifiers were deployed on 512 × 512 image size with 24 × 24 ROI size feature dataset. It was observed that SL outperformed with an accuracy of 81.263%, while classification results accuracy for MLP, SMO and RF are 80.051%, 80.707% and 79.647%, respectively, which are shown in .

Table 6. ML classifiers detailed accuracy on 512 × 512 image size for digital photographic dataset.

shows the comparative result accuracy of SL, MLP, SMO, and RF classifiers to diagnose CLCuV on a digital photographic dataset for 512 × 512 image size. illustrates the CM prediction result summary of 512 × 512 image size with 24 × 24 ROI size for healthy, maximum effected, medium effected and minimum effected severity levels of CLCuV.

Figure 4. Accuracy results for CLCuV detection on digital photographic dataset.

Figure 4. Accuracy results for CLCuV detection on digital photographic dataset.

Table 7. CM showing result of SL classifier for 512 × 512 image size of digital photographic dataset.

The same ML classifiers were deployed on a multispectral dataset in the next step. MLP classifier has outperformed on a multispectral dataset with 91.177% accuracy, while the classification accuracy results for SL, SMO, and RF are 88.202%, 82.023% and 88.483%, respectively, as shown in . Comparative result accuracy of SL, MLP, SMO, and RF classifiers for CLCuV is shown in . The prediction result summary is shown in CM for the MLP classifier for all severity levels of CLCuV.

Figure 5. Accuracy results for CLCuV detection on multispectral dataset.

Figure 5. Accuracy results for CLCuV detection on multispectral dataset.

Table 8. ML classifiers detailed accuracy on multispectral dataset.

Table 9. CM showing results of multispectral dataset using MLP.

Finally, data fusion was performed by combining multispectral and digital photographic datasets to obtain more powerful results. Fusion-based dataset have 35 (5 multispectral and 30 digital photographic) features. SL, MLP, SMO, and RF ML classifiers also classified the fused dataset. SL ML Classifier was the best classifier with 96.313% result accuracy. MLP, SMO, and RF classifiers performed well with accuracy results of 96.061%, 93.586%, and 90.354%, respectively, which is better accuracy than individual datasets, as shown in . ML classifier’s detailed accuracy results for the fused dataset are shown in , while the comparative analysis of all ML classifiers is presented in . shows the CM prediction analysis summary for a fused dataset. A detailed comparison between current state-of-the-art technologies and the proposed framework is shown in .

Figure 6. Accuracy results for CLCuV detection on fused dataset.

Figure 6. Accuracy results for CLCuV detection on fused dataset.

Figure 7. Comparative results graph for spectral, digital, and fused dataset.

Figure 7. Comparative results graph for spectral, digital, and fused dataset.

Table 10. ML classifiers detailed accuracy for fused dataset.

Table 11. CM showing results of SL classifier for fused dataset.

Table 12. Comparison between state-of-the-art current technologies and proposed multimodal fusion framework.

It is perceived that most of the research studies applied machine learning models, such as SVM, RF, CART, and BRT to distinguish healthy and CLCuV-infected leaves, but their classification accuracies remained less than 90%. In some other research studies, deep learning techniques like CNN, VGG, TensorFlow, and MLP were used to identify CLCuV with healthy leaves, but the result accuracy of most of these models was below 95%. In comparison with other methodologies, discrimination of CLCuV severity levels, multiple and fused datasets and hybrid feature selection are the parameters that make this research study exclusive. Additionally, our designed framework achieved a high accuracy of 96.313%.

4. Conclusion

This research study has designed a novel multimodal fusion framework to diagnose different severity levels of CLCuV. For the research experiment, three distinct datasets have been conceived. These are the digital photographic dataset, multispectral dataset and their fused composition. After preprocessing, the feature extraction process extracts relevant features from the designated datasets. The hybrid feature optimization (F + PA + MI) gave the thirty most optimized features from the digital photographic dataset, whereas the five spectral features were extracted from the multispectral dataset. The fused dataset has been assembled by integrating the features extracted from the digital photographic and multispectral datasets. Each contrived dataset is input to the four designated machine vision classifiers, SL, MLP, SMO, and RF, to diagnose the CLCuV. SL is the best classifier for the digital photographic dataset, with a diagnosing accuracy of 81.263%. Next, the MLP classifier gained a considerably good accuracy of 91.177% on the multispectral dataset. Ultimately, the SL classifier again outperformed the fused dataset and achieved 96.313% diagnosing accuracy. Even though digital photographic and multispectral datasets have also given good results, it has been observed that the CLCuV diagnosing accuracy results were significantly improved for the fused dataset.

Future work

The proposed framework can also be used in decision-making, defect detection, yield assessment, water quality analysis, verification process, water stress in crops, seed quality classification and bioinformatics in the future.

Disclosure statement

The authors declare no conflitcs of interest.

Additional information

Notes on contributors

Nazir Ahmad

Mr. Nazir Ahmad is Assistant Professor and Ph.D research scholar in the Department of Information Technology, The Islamia University of Bahawalpur. He is an esteemed teacher, having good knowledge in pattern recognition, data mining, machine learning and artificial intelligence. His research focuses on machine learning application in agriculture and health care. Mr. Nazir Ahmad has worked as a co-author in many publications, and supervised almost 10 MPhil students. He also has supervised more than 100 undergraduate students. He is recognized for his commitment to education. He fosters student’s growth in technology innovations.

References

  • Aher, S. B., & Lobo, L. M. R. J. (2011). Data mining in educational system using WEKA. International Conference on Emerging Technology Trends (ICETT), Kerala, India. (pp. 1–17).
  • Akhtar, K. P., Hussain, M., Khan, A. I., Haq, M. A., & Iqbal, M. M. (2004). Influence of plant age, whitefly population and cultivar resistance on infection of cotton plants by cotton leaf curl virus (CLCuV) in Pakistan. Field Crops Research, 86(1), 15–21. https://doi.org/10.1016/S0378-4290(03)00166-7
  • Ali, H., Lali, M. I., Nawaz, M. Z., Sharif, M., & Saleem, B. A. (2017). Symptom based automated detection of citrus diseases using color histogram and textural descriptors. Computers and Electronics in Agriculture, 138, 92–104. https://doi.org/10.1016/j.compag.2017.04.008
  • Anwar, S., Kolachi, A. R., Baloch, S. K., & Soomro, S. R. (2022). Bacterial blight and cotton leaf curl virus detection using inception V4 based CNN model for cotton crops . IEEE 5th International Conference on Image Processing Applications and Systems (IPAS). Genova, Italy. https://doi.org/10.1109/ipas55744.2022.10052835
  • Barbhuiya, A. J., & Hemachandran, K. (2013). Wavelet transformations & its major applications in digital image processing. International Journal of Engineering Research & Technology, 2(3), 1–5. https://doi.org/10.17577/IJERTV2IS3538
  • Blachnik, M., & Laaksonen, J. (2008). Image classification by histogram features created with learning vector quantization. Artificial Neural Networks – ICANN (pp. 827–836). https://doi.org/10.1007/978-3-540-87536-9_85
  • Farooq, J., Farooq, A., Riaz, M., Shahid, M. R., Saeed, F., Iqbal, M. S., & Mahmood, A. (2014). Cotton leaf curl virus disease a principle cause of decline in cotton productivity in Pakistan. Canadian Journal of Plant Protection, 2, 9–16. https://repo.mel.cgiar.org/handle/20.500.11766/564
  • Feng, L., Wu, B., Zhu, S., Wang, J., Su, Z., Liu, F., He, Y., & Zhang, C. (2020). Investigation on data fusion of multisource spectral data for rice leaf diseases identification using machine learning Methods. Frontiers in Plant Science, 11, 577063. https://doi.org/10.3389/fpls.2020.577063
  • Feng, Q., Liu, J., & Gong, J. (2015). UAV remote sensing for urban vegetation mapping using random forest and texture analysis. Remote Sensing, 7(1), 1074–1094. https://doi.org/10.3390/rs70101074
  • Ferentinos, K. P. (2018). Deep learning models for plant disease detection and diagnosis. Computers and Electronics in Agriculture, 145, 311–318. https://doi.org/10.1016/j.compag.2018.01.009
  • Gelli, J. S., Madduri, L. A., Tanveer, R., Bhanu, U., & Kishore, G. K. (2020). Plant disease detection using VGG and DJANGO. International Research Journal of Engineering and Technology, (IRJET ), 7(8), 4369–4373. https://irjet.net/archives/V7/i8/IRJET-V7I8753.pdf
  • Gu, Q., Sheng, L., Zhang, T., Lu, Y., Zhang, Z., Zheng, K., Hu, H., & Zhou, H. (2019). Early detection of tomato spotted wilt virus infection in tobacco using the hyperspectral imaging technique and machine learning algorithms. Computers and Electronics in Agriculture, 167, 105066. https://doi.org/10.1016/j.compag.2019.105066
  • Gudela, Y. S., Jangam, E., & Mogili, S. (2023). Cotton crop disease detection using mobilenet snapshot ensemble technique and vision transformers. 8th International Conference on Communication and Electronics Systems (ICCES). Coimbatore, India. https://doi.org/10.1109/icces57224.2023.10192615
  • Hassan, F., Qayyum, A., Malik, W., Maqbool, A., Rehman, A. M., Shoaib, M., & Arshad, M. (2016). Cotton leaf curl virus (CLCuV) disease in Pakistan: A critical review. Applied Sciences and Business Economics, 3, 8–14. https://www.researchgate.net/publication/305994089_cotton_leaf_curl_virus_clcuv_disease_in_pakistan_a_critical_review
  • Huang, K., & Aviyente, S. (2008). Wavelet feature selection for image classification. IEEE Transactions on Image Processing: A Publication of the Ieee Signal Processing Society, 17(9), 1709–1720. https://doi.org/10.1109/TIP.2008.2001050
  • Ilic, M., Ilic, S., Jovic, S., & Panic, S. (2018). Early cherry fruit pathogen disease detection based on data mining prediction. Computers and Electronics in Agriculture, 150, 418–425. https://doi.org/10.1016/j.compag.2018.05.008
  • Jiang, F., Lu, Y., Chen, Y., Cai, D., & Li, G. (2020). Image recognition of four rice leaf diseases based on deep learning and support vector machine. Computers and Electronics in Agriculture, 179, 105824. https://doi.org/10.1016/j.compag.2020.105824
  • Jovic, A., Brkic, K., & Bogunovic, N. (2015). A review of feature selection methods with applications. IEEE 38th International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO), Opatija, Croatia. (pp. 1200–1205). https://doi.org/10.1109/mipro.2015.7160458
  • Kerkech, M., Hafiane, A., & Canals, R. (2018). Deep learning approach with colorimetric spaces and vegetation indices for vine diseases detection in UAV images. Computers and Electronics in Agriculture, 155, 237–243. https://doi.org/10.1016/j.compag.2018.10.006
  • Khaled, A. Y., Abd Aziz, S., Bejo, S. K., Nawi, N. M., & Abu Seman, I. (2018). Spectral features selection and classification of oil palm leaves infected by Basal stem rot (BSR) disease using dielectric spectroscopy. Computers and Electronics in Agriculture, 144, 297–309. https://doi.org/10.1016/j.compag.2017.11.012
  • Kotian, S., Ettan, P., Kharche, S., Saravanan, K., & Kumar, K. A. (2022). Cotton leaf disease detection using machine learning. 2nd International Conference on Advancement in Electronics & Communication Engineering, Ghaziabad, India. https://doi.org/10.2139/ssrn.4159108
  • Kumar, G., & Bhatia, P. K. (2014). A detailed review of feature extraction in image processing systems. International Conference on Advanced Computing & Communication Technologies (ACCT). Rohtak, India. (pp. 5–12.). https://doi.org/10.1109/acct.2014.74
  • Lever, J., Krzywinski, M., & Altman, N. (2016). Classification evaluation. Nature Methods, 13(8), 603–604. https://doi.org/10.1038/nmeth.3945
  • Lutu, P. E., & Engelbrecht, A. P. (2010). A decision rule-based method for feature selection in predictive data mining. Expert Systems with Applications, 37(1), 602–609. https://doi.org/10.1016/j.eswa.2009.06.031
  • Ma, J., Du, K., Zhen, F., Zhang, L., Gong, Z., & Sun, Z. (2018). A recognition method for cucumber diseases using leaf symptom images base on deep convolutional nerual network. Computers and Electronics in Agriculture, 154, 18–24. https://doi.org/10.1016/j.compag.2018.08.048
  • Nilsson, H. E. (1995). Remote sensing and image analysis in plant pathology. Annual Review of Phytopathology, 33(1), 489–528. https://doi.org/10.1146/annurev.py.33.090195.002421
  • Ouhami, M., Hafiane, A., Es-Saad, Y., Hajji, M. E., & Canals, R. (2021). Computer vision, IoT and data fusion for crop disease detection using machine learning: A survey and ongoing research. Remote Sensing, 13(13), 2486. https://doi.org/10.3390/rs13132486
  • Pantazi, X. E., Moshou, D., & Tamouridou, A. A. (2019). Automated leaf disease detection in different crop species through image features analysis and One Class Classifiers. Computers and Electronics in Agriculture, 156, 96–104. https://doi.org/10.1016/j.compag.2018.11.005
  • Patil, R. R., & Kumar, S. (2022). Rice-fusion: A multimodality data fusion framework for rice disease diagnosis. IEEE Access, 10, 5207–5222. https://doi.org/10.1109/ACCESS.2022.3140815
  • Pechuho, N., Khan, Q., & Kalwar, S. (2020). Cotton crop disease detection using machine learning via tensor flow. Pakistan Journal of Engineering and Technology, 3(2), 126–130. https://doi.org/10.51846/vol3iss2pp126-130
  • Prajapati, B. S., Dabhi, V. K., & Prajapati, H. B. (2016). A survey on detection and classification of cotton leaf diseases. IEEE International Conference on Electrical, Electronics, and Optimization Techniques (ICEEOT). Chennai, India. (pp. 2499–2506). https://doi.org/10.1109/iceeot.2016.7755143
  • Qadri, S., Khan, D. M., Ahmad, F., Qadri, S. F., Babar, M. E., Shahid, M., Ul-Rehman, M., Razzaq, A., Shah Muhammad, S., Fahad, M., Ahmad, S., Pervez, M. T., Naveed, N., Aslam, N., Jamil, M., Rehmani, E. A., Ahmad, N., & Akhtar Khan, N. (2016). A comparative study of land cover classification by using multispectral and texture data. BioMed Research International, 2016, 8797438–12. https://doi.org/10.1155/2016/8797438
  • Qadri, S., Khan, D. M., Qadri, S. F., Razzaq, A., Ahmad, N., Jamil, M., Nawaz Shah, A., Shah Muhammad, S., Saleem, K., & Awan, S. A. (2017). Multisource data fusion framework for land use/land cover classification using machine vision. Journal of Sensors, 2017, 1–8. https://doi.org/10.1155/2017/3515418
  • Radhakrishnan, M. (2012). Application of computer vision technique on sorting and grading of fruits and vegetables. Food Processing & Technology, 10, 2157–7110. https://doi.org/10.4172/2157-7110.s1-001
  • Rana, A. W., Ejaz, A., & Shikoh, S. H. (2020). Cotton crop: A situational analysis of Pakistan. International Food Policy Research Institute. https://doi.org/10.2499/p15738coll2.133702
  • Ray, S. (2019). A quick review of machine learning algorithms. International Conference on Machine Learning, Big Data, Cloud and Parallel Computing (COMITCon). Faridabad, India. https://doi.org/10.1109/comitcon.2019.8862451
  • Saeed, H., Ehetisham-Ul-Haq, M., Atiq, M., Kamran, M., Idrees, M., Ali, S., Burhan, M., Mohsan, M., Iqbal, M., Nazir, S., & Il Yasin, S. (2018). Prediction of cotton leaf curl virus disease and its management through resistant germplasm and bio-products. Archives of Phytopathology and Plant Protection, 51(3–4), 170–186. https://doi.org/10.1080/03235408.2018.1443602
  • Schmitt, M., & Zhu, X. X. (2016). Data fusion and remote sensing: An ever-growing relationship. IEEE Geoscience and Remote Sensing Magazine, 4(4), 6–23. https://doi.org/10.1109/MGRS.2016.2561021
  • Shahi, T. B., Xu, C. Y., Neupane, A., & Guo, W. (2023). Recent advances in crop disease detection using UAV and deep learning techniques. Remote Sensing, 15(9), 2450. https://doi.org/10.3390/rs15092450
  • Szczypiński, P. M., Strzelecki, M., Materka, A., & Klepaczko, A. (2009). MaZda- A software package for image texture analysis. Computer Methods and Programs in Biomedicine, 94(1), 66–76. https://doi.org/10.1016/j.cmpb.2008.08.005
  • Tariq, M.,Akhtar, M. S., &Anjum, M. S. (2020). Identification of Rice Quality Through Pattern Classification Using Computer Vision Image Processing. Computer Engineering and Intelligent Systems, 11(2), 1–10. 2222–2863.
  • Wang, Z., Hu, M., & Zhai, G. (2018). Application of deep learning architectures for accurate and rapid detection of internal mechanical damage of blueberry using ­hyperspectral transmittance data. Sensors (Basel), 18(4), 1–17. https://doi.org/10.3390/s18041126
  • Yan, J. J., Kuo, H. H., Lin, Y. F., & Liao, T. L. (2016). Real-time driver drowsiness detection system based on PERCLOS and grayscale image processing. International Symposium on Computer, Consumer and Control.Xi’an, China. (pp. 243–246). https://doi.org/10.1109/is3c.2016.72
  • Zhang, N., Yang, G., Pan, Y., Yang, X., Chen, L., & Zhao, C. (2020). A review of advanced technologies and development for hyperspectral-based plant disease detection in the past three decades. Remote Sensing, 12(19), 3188. https://doi.org/10.3390/rs12193188
  • Zubair, M., Zaidi, S. S. A., Shakir, S., Amin, I., & Mansoor, S. (2017). An insight into cotton leaf curl Multan betasatellite, the most important component of cotton leaf curl disease complex. Viruses, 9(10), 280. https://doi.org/10.3390/v9100280