ArticlePDF Available

A Systematic Literature Review of Machine Learning and Deep Learning Approaches for Spectral Image Classification in Agricultural Applications Using Aerial Photography

Tech Science Press
Computers, Materials & Continua
Authors:
  • KIET Karachi Institute of Economics & Technology

Abstract and Figures

Recently, there has been a notable surge of interest in scientific research regarding spectral images. The potential of these images to revolutionize the digital photography industry, like aerial photography through Unmanned Aerial Vehicles (UAVs), has captured considerable attention. One encouraging aspect is their combination with machine learning and deep learning algorithms, which have demonstrated remarkable outcomes in image classification. As a result of this powerful amalgamation, the adoption of spectral images has experienced exponential growth across various domains, with agriculture being one of the prominent beneficiaries. This paper presents an extensive survey encompassing multispectral and hyperspectral images, focusing on their applications for classification challenges in diverse agricultural areas, including plants, grains, fruits, and vegetables. By meticulously examining primary studies, we delve into the specific agricultural domains where multispectral and hyperspectral images have found practical use. Additionally, our attention is directed towards utilizing machine learning techniques for effectively classifying hyperspectral images within the agricultural context. The findings of our investigation reveal that deep learning and support vector machines have emerged as widely employed methods for hyperspectral image classification in agriculture. Nevertheless, we also shed light on the various issues and limitations of working with spectral images. This comprehensive analysis aims to provide valuable insights into the current state of spectral imaging in agriculture and its potential for future advancements.
This content is subject to copyright. Terms and conditions apply.
This work is licensed under a Creative Commons Attribution 4.0 International License,
which permits unrestricted use, distribution, and reproduction in any medium, provided the
original work is properly cited.
ech
T
PressScience
DOI: 10.32604/cmc.2024.045101
REVIEW
A Systematic Literature Review of Machine Learning and Deep Learning
Approaches for Spectral Image Classification in Agricultural Applications
Using Aerial Photography
Usman Khan1, Muhammad Khalid Khan1, Muhammad Ayub Latif1, Muhammad Naveed1,2,*,
Muhammad Mansoor Alam2,3,4,SalmanA.Khan
1and Mazliham Mohd Su’ud2,*
1College of Computing and Information Sciences, Karachi Institute of Economics and Technology, Karachi, 75190, Pakistan
2Faculty of Computing and Informatics (FCI), Multimedia University, Cyberjaya, 63100, Malaysia
3Faculty of Computing, Riphah International University, Islamabad, 46000, Pakistan
4FacultyofEngineeringandInformationTechnology,SchoolofComputerScience,UniversityofTechnologySydney,Sydney,
Australia
*Corresponding Authors: Muhammad Naveed. Email: naveed@kiet.edu.pk; Mazliham Mohd Su’ud.
Email: mazliham@mmu.edu.my
Received: 17 August 2023 Accepted: 10 November 2023 Published: 26 March 2024
ABSTRACT
Recently, there has been a notable surge of interest in scientific research regarding spectral images. The potential of
these images to revolutionize the digital photography industry, like aerial photography through Unmanned Aerial
Vehicles (UAVs), has captured considerable attention. One encouraging aspect is their combination with machine
learning and deep learning algorithms, which have demonstrated remarkable outcomes in image classification. As
a result of this powerful amalgamation, the adoption of spectral images has experienced exponential growth across
various domains, with agriculture being one of the prominent beneficiaries. This paper presents an extensive survey
encompassing multispectral and hyperspectral images, focusing on their applications for classification challenges
in diverse agricultural areas, including plants, grains, fruits, and vegetables. By meticulously examining primary
studies, we delve into the specific agricultural domains where multispectral and hyperspectral images have found
practical use. Additionally, our attention is directed towards utilizing machine learning techniques for effectively
classifying hyperspectral images within the agricultural context. The findings of our investigation reveal that
deep learning and support vector machines have emerged as widely employed methods for hyperspectral image
classification in agriculture. Nevertheless, we also shed light on the various issues and limitations of working with
spectral images. This comprehensive analysis aims to provide valuable insights into the current state of spectral
imaging in agriculture and its potential for future advancements.
KEYWORDS
Machine learning; deep learning; unmanned aerial vehicles; multi-spectral images; image recognition; object
detection; hyperspectral images; aerial photography
2968 CMC, 2024, vol.78, no.3
1Introduction
An image constitutes an array of pixels originating from diverse sources, such as standard
specialized cameras or mobile phones. The significance of imaging extends across various domains,
particularly in detection and recognition tasks. Image processing techniques are the initial step for
computational methods by extracting valuable information from the provided image. By utilizing
this extracted information, computers can attain heightened intelligence in tackling real-world and
intricate challenges [1]. One of the limitations commonly found in traditional image processing tech-
niques is the inability to acquire spatial and spectral information for various objects [2]. Spectroscopy
investigates the behavior of light within an object, enabling the identification of materials through their
distinct spectral signatures.
Nevertheless, traditional imaging methods face limitations in covering a large volume since they
cannot extract spectral information from such images. As a result, there is an ongoing demand for
novel technologies and sensors to facilitate the automatic detection and recognition of diverse objects
[3]. Over the past two decades, notable progress has been made in the application and enhancement
of image sensors and imaging techniques, with particular emphasis on the fields of agriculture and
food [4]. Technological advancements have enabled researchers to classify food products based on
color, quality, size, and weight [1]. An interesting feature of spectral cameras is that they are capable of
seeing more than “just” colors. They capture the image within the selective wavelength ranges through
the electromagnetic spectrum.
Furthermore, they can also measure light in a small number of spectral bands and a large
number of spectral bands, typically between three and several hundred. The separation of wavelengths
can be done by filters or using instruments that measure specific wavelengths, including light from
different frequencies above or belowthe visible light range, such as infrared. Some cameras can capture
hundreds of spectral bands, and this phenomenon is known as hyper-spectral imaging [5,6].
The human eye can perceive electromagnetic waves with wavelengths ranging from 380 to 780
nanometers in the visible spectrum. Any electromagnetic waves with wavelengths beyond this range,
such as infrared, remain invisible to humans [7,8]. Therefore, spectral imaging enables acquiring
additional information that surpasses the human eye’s capabilities. There are three feasible methods
for collecting spectral image information: (i) A camera with optical elements, (ii) A continuous filter
like the filter wheel, and (iii) A complementary metal oxide (CMOS) sensor with a filter. The spectrum
essentially characterizes the amount of light in each wavelength, providing valuable insights into the
emission, transmission, or reflection of light from an object. Over the past two decades, various image-
sensing technologies have emerged, with hyperspectral and multispectral imaging being two highly
effective methods [9,10]. The primary distinction between hyperspectral and multispectral imaging is
the number of distinct wavebands they capture. Multispectral imaging typically involves fewer than 15
bands, while hyperspectral imaging can encompass hundreds of bands. Additionally, hyperspectral
imaging offers a complete spectrum for each pixel, whereas multispectral imaging provides only
isolated data points [1115]. Spectral cameras with optical elements consist of a prism, a sensor,
gratings, and lenses. The light enters the camera through a slit. The prism and gratings refract the
light. The line scan sensor can then generate a multispectral image, line by line. The filter wheel must
be moved to the desired filter. This results in a reduced scanning speed, so UAVs are not usable for
moving targets. Moreover, there are also CMOS sensors with band filters integrated into the sensor’s
layout. Compared with the other scanning types, cameras with spectral filters on CMOS sensors
have no optical parts that require alignment. They can capture objects with different wavelengths
in one single shot. A snapshot spectral camera captures multiple ultra-violet, visible, and infrared
CMC, 2024, vol.78, no.3 2969
images. Each image with a multispectral camera is passed through a filter to keep light to a specific
wavelength or colour. Spectral imaging can extract features for additional information that the human
eye cannot recognize with its receptors for RGB. Spectral images are one of the most critical images
attained by remote sensing. Remote sensing radiometers categorized the spectrum into various bands.
Recent research trends with the help of Multispectral Images are in the fields of Printed Circuit
Board (PCB), counterfeit detection on banknotes [1214], skin characterization in dermatology [15],
food inspection in the agriculture sector, etc. [1620]. The existing systematic literature review on
multispectral and hyperspectral imagery provides the details related to the specific domain. A review is
done by Recetin et al. [21] on anomaly detection-related studies using remote sensing applications with
the help of hyperspectral while Aloupogianni et al. [22] focused on the main area of tumor diseases
using both images, hyper and multispectral imagery. The health sector is again the point of attention by
Ortega et al. [23], where digital and computation pathology and hyper and multispectral imagery are
reviewed. The spectral resolution can be increased with the help of a new pan-sharpening technique
for multispectral images. Different pan-sharpening techniques are reviewed for multispectral images
in [20]. The comparison of SVM and convolutional neural networks is studied by Kaul et al. [24].
They also reviewed the most popular dataset for the same type of problems. Datta et al. [25], discussed
the challenges and future scope using hyperspectral along with the benefits of using machine learning
and deep learning. Deep learning using hyperspectral is studied by Ozdemir et al. [26] with future
directions. This study offers a wealth of detailed information, encompassing comprehensive answers
to all the research questions. For a summarized overview of the relevant systematic literature review,
please refer to Table 1.
Table 1: Comparison of the proposed research with the existing literature
Studies Trends Health
sectors
ML&DL
analysis
Popular data
sets
Identifying
agriculture sec
Future
directions
[21]XX X X
[22]XX X
[23]XXX X
[20]XX X X X
[24]XX X
[25]XX X
[26]XXX XX
Our paper X
The research questions formulated for this study are structured to comprehensively address
all inquiries about multispectral and hyperspectral images and their applications in the agriculture
industry. Moreover, the study concentrates on elucidating the methodologies and various techniques
employed to address agricultural challenges effectively through the utilization of hyperspectral or
multispectral images. The addressed research questions (RQ) for this study are:
RQ1 .In what diverse research areas within agriculture have Multispectral or Hyperspectral images
found applications?
RQ2 .Which Machine Learning and Deep Learning Techniques have been employed to tackle
agricultural problems using Multispectral and Hyperspectral images?
2970 CMC, 2024, vol.78, no.3
RQ3 .What are the prevalent datasets extensively utilized for spectral classification purposes?
RQ4 .What are the challenges, limitations, and potential future developments concerning Hyperspec-
tral and Multispectral images in conjunction with Deep Learning and Machine Learning Models?
The systematic literature review methodology was selected to categorize studies that explore
the convergence of spectral images and agriculture fields. This study comprehensively classifies the
research questions concerning multispectral and hyperspectral images. The exponential advancement
of spectral images has significantly enhanced the accuracy of classification problems. The primary
goal of this paper is to review the wide-ranging applications of spectral images and the frequently
employed techniques alongside them. Additionally, we focus on the commonly used datasets in spectral
applications, their limitations, and potential areas for future research.
The paper is structured as follows: Section 2 delves into the background of Multispectral imaging,
Hyperspectral imaging, Machine Learning, and Deep Learning. In Section 3, we explain the materials
and methods used for the systematic literature review. In Section 4, we present the results, offering
comprehensive and in-depth answers to all the research questions posed, and Section 5 presents
the challenges and future directions of our study. Finally, in Section 6, we conclude the paper by
summarizing our findings and suggesting potential areas for future research.
2Background
2.1 Multispectral Imaging
Multispectral imaging is a new and emerging area with diverse application domains such as
agriculture, health and sciences, geo monitoring, and environmental changes. Recent research trends
indicate that the synergy of multispectral images with other fields has generally proven effective
in urban and regional planning [27]. This section discusses some of the recent research around
multispectral images. One crucial problem is soil analysis, particularly soil salinization, where soil
degradation is a prime issue. Hu et al. [6] attempted to solve this problem with an unmanned aerial
vehicle (UAV) borne multispectral imager. Remote sensing is very useful for collecting data for
agriculture applications. Zheng et al. [3] used different modeling algorithms, including parametric and
non-parametric algorithms, for estimating leaf nitrogen content (LNC) in winter wheat. This was done
with the help of multispectral images using UAVs. With the help of classification algorithms, including
linear regression (LR), random forests (RF), and support vector machine (SVM) alongside UAV
multispectral images, the prediction of canopy nitrogen weight in Corn was discussed by Lee et al. [28].
They reported that RF and SVM are better than LR for the problem. In weed monitoring, the
main issue is determining weed amount and location. For this problem, three methods are proposed
by Osorio et al. [29] using deep learning and multispectral images with the comparison of visual
estimations from the experts. The proposed methods are SVM, You Only Look Once Version 3
(YOLOV3), and Region-based convolutional neural network (R-CNN). The F1 scores were 88%, 94%,
and 94%, respectively. Another area where multispectral images are used is related to oil pollution,
a severe environmental issue. Ozigis et al. [30] worked on detecting oil pollution and its impact
on biological elements. They used multispectral and multi-frequency images, fuzzy logic, and RF
methods. An overall accuracy of 75% for dense areas of vegetation was obtained. Furthermore,
the Cropland and grassland areas had an accuracy of 59.4% and 65% respectively. Lima et al. [31]
developed a system for early-stage supervision of the status of organic fertilization on tomato plants.
The proposed system was developed using a multispectral camera with five lenses including green,
red, red edge, near-infrared, and RGB alongside with computation image processing mechanism.
One of the successful mechanisms of monitoring the characteristics of vegetation by exploiting the
CMC, 2024, vol.78, no.3 2971
Normalized Difference Vegetation Index (NDVI). With the help of multispectral images, Cao et al. [32]
identified that using wide-dynamic-range vegetation index (WDRVI) derived from UAV multispectral
image monitoring of sugar beet growth indicator is very useful as sugar beet has very large ground
biomass.
Monitoring several crops (cultivated areas) is very challenging since identifying different tops
is very costly. Sicre et al. [33] attempted to find different ways to use satellite images (both optical
and radar) to classify the land covers. Their suggested method was dependent on the wavelength and
penetration depth of the signal providing the images. Lei Ma et al. [23] reviewed the land cover object-
based classification in different research publications. The findings indicate that remote sensing-based
imagery was the dataset most frequently used.
UAV can monitor leaf area index (LAI). Qi et al. [34] proposed new ways to find the high-precision
LAI prediction system with the help of multispectral images. Multispectral images are also useful in
classifying potato defects [3537]. In addition, detecting bruises on apples is also a very challenging
classification task when using multispectral images. Additionally, classifying bruises is very important
when discussing the automatic apple classification system. With the help of principal component
analysis (PCA), hyperspectral imaging and multispectral imaging systems for selective bands were
developed by Huang et al. [16], which achieved 90% accuracy. Powdery mildew (PM) is a fungal
disease that can cause powdery growth on the surface of plants, leaves, and fruits. Early detection of
powdery mildew is very crucial. Chandel et al. [17] studied using multispectral images for PM detection
and mapping, achieving 77% accuracy. Bhargava et al. proposed a binary classifier [18] for sorting
mono and bi-color apples. They achieved different accuracy on different datasets using multispectral
imaging. Feng et al. showed the results in [19] that indicate that fruit recognition using multispectral
images can be very effective. The current state of the art and new applications of remote sensing in
agriculture were discussed by Khanal et al. [38]. Applications related to plant disease detection, yield
estimation, water stress monitoring, etc., were discussed.
2.2 Hyperspectral Imaging
The hyperspectral image consists of several bands that have unique classifications. Each segment
of the pixel can be treated as a unique label that identifies a specific target class. After image acquisition,
a hypercube is formed, which contains information regarding both spectral and spatial data. As the
hypercube has three dimensions, the image’s resolution helps in two dimensions, while several bands
assist in-depth, which is the 3rd dimension. Ranjan et al. [39] used principal component analysis (PCA)
for feature extraction for hyperspectral image classification, and with the help of K-means clustering,
they formed different clusters. After that, all clusters were trained using SVM. The accuracy of this
approach was recorded to be 90%, 77%, and 88% for three different types of classes. They concluded
that accuracy was better when data was initially divided into different groups before classification.
Research on crop height for maize biomass by Zhang [40] used UAV hyperspectral imagery. Prediction
of Aboveground biomass (ABG) using stepwise regression and XGBoost regression model with
the highest accuracy of R-Squared (R2) of 0.81 and Root Mean Square Error (RMSE) of 0.27
was obtained, showing that hyperspectral imagery can play a vital role in the estimation of maize
aboveground biomass with better accuracy. Rubio-Delgado et al. [41] proved that hyperspectral images
effectively estimate the determination coefficient between spectral data and leaf nitrogen content. They
used a full spectrum range for vegetation indices (VI) and Partial Least Square Regression (PSLR)
models and achieved a large R2value. Results can be improved by using temporal information on
leaf nitrogen content. Messina et al. provided a comprehensive research summary of remote sensing’s
contribution to the growth of olive farm management in the last two decades [42].
2972 CMC, 2024, vol.78, no.3
Yao et al. provided a complete review of remote sensing applications using UAVs [43]. Selected
applications were from precision agriculture and vegetation, urban environment and management,
and disaster hazards and rescue. The advantages and selected areas were also highlighted in the
conclusion. Gautam et al. studied the water status of horticultural crops using different remote sensing
applications [44]. Remote measurements of water status, such as soil moisture, canopy 3D structure,
etc., remote sensing is the best tool for planning and managing different applications for larger areas.
Yan g et a l. [ 45] found that airborne hyperspectral imagery is used in mapping cotton yield. With
stepwise regressions help, 61% and 69% variability for two yield fields were observed. They concluded
that due to narrow bands, hyperspectral imagery is much better than multispectral imagery, which uses
broad bands. Zhang et al. [46] used hyperspectral images to observe water quality and proposed that
a deep learning model could monitor water quality concentration. Mean absolute percentage error
(MAPE) was 8.78% and 12.36%, and R2was 0.81 to 0.93. They suggested using cloud stations for
a reduction in time and costs. Coverage using UAVs can also be much broader so that more data
can be collected quickly. It is very important to use an optical sensor to detect crop stress, which can
specifically find spectral wavelength affected by stress factors. It is also essential to consider which
spectral resolution is required to prevent it. Mewes et al. [47] discovered that the problem is reducing
the spectral resolution and spectrum angle mapper. The proposed method for this problem was SVM.
The results show that only a few hyperspectral bands are enough to detect fungal infection in wheat.
In contrast, the authors also suggested using the derived result for feature selection in agriculture.
Cen et al. [48] examined the internal defects of pickling cucumbers using hyperspectral images. The
performance measure was minimum redundancy–maximum relevance (MRMR) with 95% and 94%
accuracy on two different conveyor belt speeds. They primarily worked on identifying several optimal
wavebands for PCA for fast and efficient algorithms and light sources. Further study can be done in
discrimination between slightly and severely defective labels.
2.3 Machine Learning
Machine Learning is a branch of Artificial Intelligence that deals with problems related to predic-
tion and classification. The phrase ‘learning’ relates to the amount of data given to an algorithm, and
more data would mean more learning. For any task (T), if the algorithm’s performance (P) increases
with respect to experience (E), then the algorithm is said to be a machine learning algorithm. Machine
Learning is fundamentally classified into three types: (1) Supervised Learning, (2) Unsupervised
Learning, and (3) Reinforcement Learning. In supervised learning, the model is trained with labelled
data and predicts the labels from new data. In supervised learning, learning is done from a dataset,
making predictions iteratively, and with each iteration, the parameters of the algorithms are adjusted
accordingly. Supervised learning includes LR, Logistic Regression, Nearest Neighbors, Decision Tree,
RF, SVM, etc. In unsupervised learning, the model is trained with unlabeled data and groups the data
based on feature similarity. Unsupervised learning tries to find the patterns in the dataset without
any pre-existing target labels. It helps in finding insights into data. Sometimes, unsupervised learning
performs complex tasks better than supervised learning. Clustering, association, and dimensional
reduction are the main approaches that use unsupervised learning [49]. In reinforcement learning,
the algorithm performs actions that maximize performance in particular conditions. Fig. 1 shows the
three types of machine learning and their differences.
Fig. 2 explains the overall flow of a machine-learning algorithm. In the first phase, the training
phase, the data or training examples go to the feature extractor for feature extraction.
CMC, 2024, vol.78, no.3 2973
Figure 1: An overview of types of machine learning algorithms
Figure 2: An overview of machine learning model flow
After getting features from training examples, the model for machine learning is trained; once the
training model is ready for the prediction, the system takes a new test example and classifies/predicts
the output.
2.4 Deep Learning
Deep learning is a subset of machine learning inspired by the human brain, consisting of billions of
neurons. Deep learning uses Artificial Neural Network algorithms. The word deep refers to the number
of layers in a network. If the number of layers is unknown, the system automatically adjusts the weights
using an optimization function. The layer consists of several neurons. Since it mimics the human brain,
deep learning also requires much data for better accuracy [50]. Although deep learning evolved from
2974 CMC, 2024, vol.78, no.3
machine learning, the two have some differences. The main difference is that deep learning requires
large amounts of data, and features are not extracted before applying the deep learning model. Deep
learning can automatically extract the features that further process the data for classification. The
concept of deep learning is illustrated in Fig. 3 for further clarification.
Figure 3: Deep learning model flow showing feature extraction and model prediction
3Materials and Methods
A systematic literature review (SLR) was carried out according to the approach suggested by
Kitchenham [51]. We performed the following steps to get the answers to our research questions:
(1) Search, (2) Exclusion and Inclusion Criteria, (3) Publisher Selection, (4) Study Selection, and (5)
Search Strategy.
3.1 Search
The databases used in the search were Springer, Science Direct, IEEE Xplore, and ACM. The
timeline for the manuscripts was between 2017 and 2022 because of the most recent publication in
multispectral and hyperspectral images. We have used four search strings for the survey.
Search String 1. (Multispectral Images) AND (Agriculture) AND (Deep Learning)
Search String 2. (Multispectral Images) AND (Agriculture) AND (Machine Learning)
Search String 3. (Hyperspectral Images) AND (Agriculture) AND (Deep Learning)
Search String 4. (Hyperspectral Images) AND (Agriculture) AND (Machine Learning)
3.2 Exclusion and Inclusion Criteria
The exclusion and inclusion criteria were straightforward. We selected the research from 2017
onwards and did not include any articles before 2017 in our primary studies. The reason for selecting
the last five years is to provide the recent developments in hyperspectral and multispectral images. All
CMC, 2024, vol.78, no.3 2975
the selected articles are peer-reviewed, and all articles are searched using the mentioned string. All
articles were written in English, and articles in other languages that appeared during the search were
not considered. Fig. 4 depicts the inclusion and exclusion criteria.
Figure 4: Exclusion and inclusion criteria for the selection of primary studies
3.3 Publisher Selection
We search the articles from the following databases: SPRINGER, SCIENCEDIRECT, IEEE
XPLORE, and ACM. Surprisingly, we did not find papers in IEEE Xplore despite trying the search
with each of the four strings. Fig. 5 shows the well-known publishers from where articles are selected.
Figure 5: Selected databases for this research
2976 CMC, 2024, vol.78, no.3
Fig. 6 shows the initial frequency comparison of all the research articles from selective publishers.
We found that Google scholars have comparatively large numbers of papers compared to other
databases.
Figure 6: Search string result comparison
3.4 Search Strategy
The search strategy involved an automated search, as depicted in Fig. 7. The automated search
proved instrumental in locating and providing us with 30,100 research articles for the study, out of
which 27,900 papers were obtained from Google Scholar. Consequently, we excluded the articles
sourced from Google Scholar to avoid redundancy.
Figure 7: Preferred reporting items for systematic reviews and meta-analysis (PRISMA)
CMC, 2024, vol.78, no.3 2977
Following the exclusion process, we extracted 2,200 articles for this study. Subsequently, we applied
inclusion and exclusion criteria based on titles, abstracts, year of publication, and keywords, resulting
in 185 articles remaining after the screening process. In the next phase, we applied the selection criteria
to the full-text studies, and we were left with 44 primary studies. Our selection criteria aim to include
multispectral or hyperspectral images or images with deep learning or machine learning in agriculture.
Fig. 8 shows the distribution of selected/primary studies for selected years after applying all the
mentioned criteria and selection mechanisms.
Figure 8: Articles year-wise number count
4Results
This section provides a concise and precise description of the experimental results, their inter-
pretation, and the experimental conclusions that can be drawn. The distribution of all the number
of primary studies is given in Tabl e 2 . We can see that in 2020, the number of articles increased
comparatively compared to others. The overall number of articles was found to be 44. We have
presented the analysis of results corresponding to the research questions outlined in the Introduction
section at the beginning of this paper. In Fig. 9, we have provided a detailed analysis of the articles,
including their respective publishers, to enhance clarity and understanding. Additionally, we present
a comparison of yearly articles based on publishers.
Table 2: Number of articles by year
Year Number of studies Articles
2017 1 [52]
2018 5 [5357]
2019 9 [5866]
2020 16 [6783]
2021 10 [8493]
2022 3 [9496]
2978 CMC, 2024, vol.78, no.3
Figure 9: Yearly distribution of articles categorized by publishers
4.1 Analytical Findings Corresponding to RQ1
RQ1. In what diverse research areas within agriculture have Multispectral or Hyperspectral images
found applications?
The selection of primary studies focused on hyperspectral and multispectral applications in
agriculture. To address our first research question (RQ1), we categorized all relevant agricultural works
into four main classes: grains, fruits, vegetables, and plant issues. The distribution of these classes is
illustrated in Fig. 10 using a Pie chart, revealing that 43% of the selected studies were related to fruits.
Figure 10: Classification of studies for major agriculture classes
These classes encompass a range of diverse subclasses. Under the grains class, we included research
studies on rice and various types of seeds. In the plant class, we focused on issues such as citrus greening
detection, coffee ripeness, and yield monitoring, identification of apple and rose flowers, detection of
different leaf diseases and deficiencies, crop water detection, mapping of agriculture, and vine disease
detection. The summarized structure is presented in Table 3. The vegetable class comprised primary
studies on defect detection and grading of potatoes and tomatoes and yield forecasting of carrots and
cucumbers. However, our primary emphasis was on fruits, which involved the detection and grading
of apples, estimation of mangoes’ maturity, identification of strawberry diseases and ripeness, early
detection of decayed oranges, assessment of pineapple quality, and detection of pears crown bruises,
along with bruise detection of various fruits and recognition of on-branch fruits.
CMC, 2024, vol.78, no.3 2979
Table 3: Summarized structure of RQ1
S. no. Research area Detail Referenced article
1 Grains Rice lodging resistance, identification of
single herd seed, discrimination of
high-quality watermelon seed
[71,80,87,96]
2 Plant Coffee ripeness monitoring, apple flower
detection, coffee yield prediction, plant
leaf disease detection, nutrition
deficiencies in plants, crop water stress,
citrus fruits detection, agriculture
mapping, citrus greening detection, vine
disease detection, rose plant identification
[52,57,58,60,61,65,67,70,74,77,
83,85,92,95,97]
3 Vegetables Potato genotype, external defects on
tomatoes, bruise detection on different
vegetables, tomato discrimination and
grading, plant level tomato biomass and
yield, carrot yield forecasting, internal
defect detection of pickling cucumbers
[48,54,56,64,75,81,88,94]
4 Fruits Apple fruit sorting, maturity estimation of
mangoes, fruit detection in apple orchard,
quality control of apples, bruise detection
on different fruits, apple leaf condition,
strawberry disease, extraction of apple tree
crown information, strawberry ripeness,
fuji apple detection, blossom detection in
apple, defective apple detection, on branch
fruit recognition green mangoes detection,
strawberry maturity classification, early
decayed oranges, pineapple quality, apple
fruit disease, early bruise detection on the
surface of crown pears
[53,55,59,62,63,68,70,72,73,78,
79,82,84,86,89,90,91,93]
4.2 Analytical Findings Corresponding to RQ2
RQ2.Which Machine Learning and Deep Learning Techniques have been employed to tackle
agricultural problems using Multispectral and Hyperspectral images? To address RQ2, we thoroughly
examined all the selected articles, which involved 17 distinct algorithms. Notably, the most frequently
employed techniques were from deep learning, utilized in 27 instances, followed by SVM, used 13
times.
Fig. 11 illustrates all the machine learning and deep learning techniques employed with multi-
spectral and hyperspectral images. Certain techniques were applied to address specific agricultural
problems, and researchers recommended their continued use for similar challenges due to algorithmic
and data limitations. We have categorized these techniques into five groups: Deep Learning techniques,
SVM, K Nearest Neighbors, and Logistic Regression, and the remaining methods were grouped
2980 CMC, 2024, vol.78, no.3
under “Other.” Fig. 12 illustrates the occurrence frequency of each technique in our selected studies.
Additionally, Tabl e 4 provides a study-wise classification of Machine and Deep Learning algorithms.
Figure 11: An overview and classification of machine and deep learning techniques
Figure 12: Machine learning and deep learning techniques for RQ2
Deep Learning, particularly Convolutional Neural Networks, has shown remarkable performance
in various spectral classification problems [53,55,61,62]. Studies suggested that Convolutional Neural
Networks are very well-suited for spectral image classification because they can automatically learn
discriminative features directly from the input data without the need for explicit feature engineering
[7476,83,8892]. They can capture complex patterns and relationships in spectral images, leading to
accurate classification results. Deep Learning models can handle large amounts of data and effectively
learn from diverse spectral bands, making them suitable for hyperspectral and multispectral image
analysis [52,58].
Conversely, in chosen studies, SVMs are extensively employed owing to their capacity to manage
high-dimensional feature spaces and effectively handle nonlinear data [57]. SVMs seek to find the
optimal hyperplane that maximally separates different classes in the feature space [69,70]. They can
handle the curse of dimensionality by utilizing the kernel trick, which implicitly maps the input data
to a higher-dimensional feature space, allowing for better separation of classes. SVMs have a strong
theoretical foundation, work well with small to medium-sized datasets, and have been successfully
applied to spectral image classification tasks [93]. Deep Learning was the most preferred choice of
researchers when working with multispectral and hyperspectral images in the agriculture domain.
CMC, 2024, vol.78, no.3 2981
Table 4: A task-based algorithmic summary of RQ2 based on individual studies
Ti Article Technique
Detection
[64] Decision tree
[61,66,69,80,93] Support vector machine
[53,55,61,62,69,72,7476,83,91], Deep neural network
[93] Random forest
[61,69,90] Logistic regression
[71] You only look once
Detection and classification
[52] Decision tree
[52,58,65,67] Support vector machine
[52,58] Deep neural network
[33] K nearest neighbors
Classification and regression [56,67] Deep neural network
[67,70] Support vector machine
Classification
[57,86] Support vector machine
[54,63,78,86,88,89,94] Deep neural network
[58]K-means
[59,70] K nearest neighbors
[69] Adaboost
[60] Watershed segmentation
[85] You only look once
[61] Genetic algorithm
[61] Multiple linear regression
[57] Random forest
[80] Principal component analysis
[57] Back propagation neural network
[75] Naive bayes
Detection and segmentation [68,84] Deep neural network
Regression [70] Partial least square
[92] Ridge regression
Recognition [73,87] Deep neural network
[37] K nearest neighbors
Regarding different types of tasks, classification is the most widespread problem while working
with spectral images. We find that there are several reasons for the popularity. One of the primary
objectives in agricultural applications is to identify and classify different plants, diseases, or crops.
Classification provides specific interpretable results so agriculture experts can easily understand and
treat them accordingly. While regression techniques can be used for crop yield estimation, classification
2982 CMC, 2024, vol.78, no.3
can be a precursor to yield prediction. Identification of weeds and their management are essential parts
of agriculture. Weed detection can reduce its competition with crops. Thus, different classification
models on spectral images can be trained to assess the health of crops. The health of crops can include
stress signs, different deficiencies, or any disease that can affect crop growth. We have already stated
that SVMs and deep learning models have proven more effective while working in spectral image
analysis in the agriculture sector.
4.3 Analytical Findings Corresponding to RQ3
RQ3: What are the prevalent datasets extensively utilized for spectral classification purposes?
Since multispectral and hyperspectral imagery is an emerging field, many datasets can be found in
the literature. Here are some of the most extensively utilized datasets. The Xuzhou dataset was collected
in 2014 at the Xuzhou peri-urban site. This dataset consists of 500 ×260 pixels with 9 available classes.
The Salinas dataset is a widely used benchmark dataset for hyperspectral image analysis. It represents
an agricultural area in Salinas Valley, California, and contains hyperspectral imagery acquired by an
airborne sensor. The dataset consists of 512 ×217 pixels with 224 spectral bands of 16 classes. The
Pavia Centre imagery dataset is collected at the University of Pavia, Italy, using a reflective optics
system imaging spectrometer (ROSIS). Nine distinct land cover classes are accessible, with a spatial
resolution of 1096 ×1096. The Pavia University dataset was also captured in the University of Pavia
in 2001, having a spatial resolution of 610 ×640 and a total of 9 classes of urban environmental
construction. Houston dataset consists of 380–1050 nm spectral wavelength with 48 bands. This data
was provided by the IEEE GRSS data fusion contest and captured by the University of National
Center of Airborne Laser mapping on the premises of the University of Houston. The Kennedy Space
Centre dataset was collected in 1996 at Kennedy Space Center, Florida. In Tabl e 5 , we have compiled
a summary of the dataset, including the number of available classes, spectral size, and spatial size.
Table 5: Synopsis of the dataset, encompassing spatial resolution, spectral size, and the number of
classes
Name of dataset Spatial resolution Spectral size Number of classes
1Xuzhou 500 ×260 436 9
2Salinas 512 ×217 224 16
3Pavia Centre 1096 ×1096 102 9
4Pavia University 610 ×640 103 9
5Houston 380 ×1050 191 15
6Kennedy Space Centre 512 ×614 224 13
7Hekla 2084 ×614 157 22
8Trento 600 ×166 63 6
9Botswana 1476 ×256 145 14
10 Indian Pines 145 ×145 220 16
This data is collected with the help of JPL’s Airborne imaging spectrometer, which has 13 classes
and a spatial resolution of 512 ×614. The Hekla dataset has a spatial resolution of 2084 ×614 and
a spectral size of 157. There are t22 classes available in this dataset. The Trento dataset was collected
with the help of the AISA eagle sensor. The total number of classes is six, and the spatial resolution
CMC, 2024, vol.78, no.3 2983
is 600 ×166. Botswana data was collected using a Hyperion sensor over the Okavango Delta in
Botswana, South Africa, in 2001. This consists of 14 land cover labels with a spatial resolution of
1476 ×256. Indian Pines dataset was obtained by AVIRIS sensor in northwest Indiana, United States
of America. This dataset has 16 labels of land covers and 220 spectral bands in the range of 0.2 to
2.4 μm, having a spatial resolution of 145 ×145 pixels [9597].
Indian Pines,Salinas,andHouston datasets are commonly utilized in various studies. Additionally,
the Pavia University dataset and the Pavia Centre and Kennedy Centre datasets hold a prominent
position among the most widely employed datasets in spectral imagery [98100].
4.4 Analytical Findings Corresponding to RQ4
RQ4: What are the challenges, limitations, and potential future developments concerning Hyperspec-
tral and Multispectral images in conjunction with Deep Learning and Machine Learning Models?
Some primary studies suggest employing the same images with alternative classification algo-
rithms to compare results. Within our selected primary studies, various issues were addressed using
limited data, prompting the recommendation for larger datasets to achieve improved accuracy. How-
ever, due to environmental and camera bandwidth constraints, certain limitations persist. Interestingly,
some research asserts that multispectral images are suitable for classifying various types of fruits and
vegetables, while others advocate their use for flowers. To provide a comprehensive overview, we have
summarized these findings, including limitations and challenges, in Tabl e 6 . These findings pave the
way for future research in the expansive domain of multispectral and hyperspectral imagery.
Table 6: Limitations of the work, datasets description and working details
Article Issues/Challenges Data set Worked on Image type
[59] Comparative analysis of
other classification
algorithm for the same
problem
Leaves images of four
categories orange, pink,
red, white, 25 each with
a23MPcamera
Evaluate the
performance of k
nearest neighbor for
rose classification using
five features each from
histogram and texture
related characteristics
Other
[83] The learning process is
slow and can be separated
the for feature selection
process and this method
can be applied to other
fruits as well
Multispectral images of
(450,500,750,800) nm of
infected and healthy
apples, bi-color images
total of 3279 images
Deep neural networks
with a high number of
layers can improve the
accuracy of the apple
fruit disease diagnosis
Multi
(Continued)
2984 CMC, 2024, vol.78, no.3
Table 6 (continued)
Article Issues/Challenges Data set Worked on Image type
[68] Small size of training
samples can be improved
by enriching the dataset
with more disease samples
and can be tested on other
deep learning architecture
Visible and infrared
images 850 nm four
classes 70,560 patches
of size 32 ×32
Studying imaging
modalities in the visible
andinfraredspectral
domains, how can we
combine two types of
images to delineate
symptomatic areas
Hyper
[69] Application scenario is
restricted, can be tested on
different stages of growth,
hard to fit for every growth
stage, different
multispectral camera has
different bandwidth
7695324 samples of
citrus greening infected
and healthy images
Feature pre-processing,
feature extraction,
machine learning
models and accuracy
improvement on
unmanned aerial vehicle
multispectral images
Hyper
[70] Possibility of testing new
algorithm can lead to
better results; also, further
work can be done
including feature selection
and data fusion techniques
Multispectral images in
18 different wavelengths
from 450–970 nm
Mathematical model
basedondifferent
analytical instruments,
comparison of model
performance and
explore the capabilities
and limitations of tools.
Multi
[60] Can be improved with zero
tolerance, image processing
algorithm can be more
optimized by using a large
number of samples,
proposed method can be
applied to other fruits
Multispectral images of
seven different
wavelengths from
575–960 nm
Watershed
segmentation method
with multispectral
principal component
and image
reconstruction
Multi
[84] Accuracy can be
maximized by considering
the early growth of crop
yield
PASCAL visual object
class challenge dataset
for object class
recognition, raster
dataset using extract,
transform, and load
script, point cloud
dataset
Automated all the tasks
like high resolution,
multispectral imagery,
labeling, input dataset
preparation
Other
(Continued)
CMC, 2024, vol.78, no.3 2985
Table 6 (continued)
Article Issues/Challenges Data set Worked on Image type
[85] Method can be proposed
for flowers and fruits
which are hidden under the
leaves, better accuracy can
be achieved with a
high-resolution camera,
more images can be used
with a high number of
cameras
Ten rows of
strawberries, 3875
images with 70% frontal
overlap and 60% side
overlap. All images from
the Plant science
research and education
unit in the USA
Automated
classification of the
strawberry maturity
system was developed
Other
[71] Detection of different
cultivars by retraining the
model. And in different
orchards, different
cultivars, age, pruning style
Images were 544 ×544
pixels; batch size was 64
Unmanned aerial
vehicle images instead
of ground-based
photography for the
same cultivars
Other
[72] Using thermal time to
improve accuracy,
investigated the
performance of shape
model fitting based
method in the absence of
high temporal, and
monitoring the accuracy of
convolutional neural
network-based approach in
mono temporal imagery
63360 training,
validation 13580 testing,
20440 images
Mono-temporal images
of unmanned aerial
vehicles to estimate crop
phenology for a large
area after training
Other
[73] The proposed method can
be trained with large data
set, can use transfer
learning
RGB images of six
different categories
from local orchards of
Iran, different batch
sizes of 16, 32, and 64
with 50 epochs. Total
739 and new dataset
after augmentation
12443
Data augmentation for
creating more data with
original data by
applying some
transformation. New
images with old features
with certain changes,
because deep learning
needs a large number of
data
Other
(Continued)
2986 CMC, 2024, vol.78, no.3
Table 6 (continued)
Article Issues/Challenges Data set Worked on Image type
[74] Convolutional neural
networks can be combined
with mass spectrometry
imaging for the
illumination and detection
improvement
Fuji apples total 0f 300
after initial image
acquisition total of 3300
images and after
rotation with 3 different
angles total of 79200
images was formed
In contrast to other
convolutional neural
network models for the
same problem the speed
of the proposed model
is better (5 fruits/sec)
Other
[75] Results of the detection of
blossoms can be used with
depth information which
can further allow spatial
distribution of blossoms
and new rules can be
developed for image-based
removal systems.
Multiclass object detection
for canopy parameters
Data collection was
done in Washington
USA orchard in April
2018 and 2019 selective
period
Deep learning based
blossom detection
method
Other
[87] The same model can be
tested on a new test dataset
for enhancement and will
update the dataset with a
new model and different
sensors with different
combinations can also be
used for monitoring
500 RGB images which
further disease diversity
resulting in 2100 images
with 450 epochs for
training
Effective optimization
method for counting
lesions at canopy level
using majority voting
Alsoworkedonthe
relationship between the
actual score and the
estimated score
Other
[76] The model can be tested on
other fruits as there is no
specific information
relatedtotomatoesgiven
to the model during the
training phase
43843 images using
different sensors were
used
Only defect detection,
differentiate in defect
detection is not done
Other
(Continued)
CMC, 2024, vol.78, no.3 2987
Table 6 (continued)
Article Issues/Challenges Data set Worked on Image type
[62] Less amount of data is a
drawback when using any
neural network, also
results were affected due to
the RGB-D depth sensor
under sunlight. In the
future 3D fruit localization
can be done with the help
of 2D
Fuji database 967
multimodal images and
a total of 12839 images
of Fuji apples
Different channel
registration including 3
modalities RGB color,
depth, and intensity of
range corrected
Other
[77] More sample images with
the internal and external
quality of strawberries at
different stages of ripeness
Total of 500 images 336
and 144 grayscale
images 227 ×227 pixel
Strawberry fruits
ripeness at two different
stages (early ripe and
ripe) with hyperspectral
imaging data
Hyper
[90] Some models have the
same type of errors for all
features and need
improvement for feature
selection, also difference in
mean absolute percentage
error for a top month and
three months is very little
so only one month of data
can be used. In future
multispectral and
hyperspectral data can be
used for further
improvement
Data collection from
June 2017 and May
2018 for 144 trees in
Minas Geris Brazil
Seven different
regression methods were
used for the selection of
the best features
Other
[78] The model can be deployed
in an embedded system
using unmanned aerial
vehicle, also tree detection
and segmentation datasets
should be enriched before
training and different
methods for fruit tree
information can be used
50 images resolution of
5472 ×3648 unmanned
aerial vehicle images
Tree data acquisition
model for apple tree
data which consists of
counting, detecting, and
segmentation apple
trees and detection of
crown parameters of
apples
Other
(Continued)
2988 CMC, 2024, vol.78, no.3
Table 6 (continued)
Article Issues/Challenges Data set Worked on Image type
[94] The same system can be
agricultural Unmanned
Aerial Vehicle and
agricultural robots etc., for
more automation and
precision
480 images of
strawberry diseases,
2400 total images
Cloud service, mini
program development,
and deep learning
technology with the
help of a
self-supervision
mechanism
Other
[63] The same method can be
used for the classification
of other types of spectral
data
A total of 2700 images
of fruits and vegetables
of 13 classes, a total of
16 spectral bands
ranging from 470 to 630
nm
Added one or two layers
for compression of
hyperspectral
information down to 3
layers
Hyper
[54] The same minimum
redundancy and maximum
relevance can be used to
find which band is more
essential and effective in
monitoring after
man-made infection.
Training patches were
3000 validation 6000
testing 6000 (700 per
class overall)
Reduced unnecessary
bands from
hyperspectral sensor,
data abstraction due to
hierarchy in
convolutional neural
network and Fully
convolutional network.
Hyper
[91] To avoid false positive
results different
convolutional neural
networks can be trained
and combined for majority
voting
A total of 30 apples of
different grades were
used for RGB and
near-infrared cameras
and the same apples
were used for bruise
detection by doing some
damaging
Several convolutional
network models were
trained
Other
[64] A small dataset is a big
problem, with the help of a
large dataset the learning
process can be done
automatically instead of
manual selection. In the
future different ranges of
wavelength can be
considered for fruit
reflectance
3 trees from biologische
bundesanstalt,
bundessortenamt und
chemische industrie two
for training and one for
testing
Illumination condition
does not affect the
measurement and 3D
fruit local information
provided by the
proposed method
Other
(Continued)
CMC, 2024, vol.78, no.3 2989
Table 6 (continued)
Article Issues/Challenges Data set Worked on Image type
[65] The same method can be
used to detect defects in
cherries and heirloom
tomatoes. And also, can be
tested on other varieties of
tomatoes. The accuracy is
inversely proportional to
number of target classes
Total of 2000 images
which 1400 training and
600 testing
Calyx and stalk scar
detection technique
with an overall average
accuracy of 0.9515
Other
[55] The same method can be
used for other types of
fruits
Total of 147 images
using Canon 60D under
normal light conditions
Convolutional neural
network-based
classification algorithm
with the uncontrolled
environment and also
tested on the unseen
dataset
Hyper
[56] Other factors can be
considered while using the
same proposed method like
different orchards
comparison, seasons or
time of detection, and also
the limitation of
illumination Hyperspectral
camera is expensive so can
we do the same proposed
method with RGB or
multispectral camera
The data were collected
using three cultivars
B74 mango orchards
block A 10 rows of 494
treesblockB5rowsof
121 trees and block C 4
rows of 266 trees, A
total of 78 fruit were
selected
Using hyperspectral
sensing for the detection
of fruits and averaging
dry matter per tree
estimation
Other
[92] Transfer learning technique
can be applied for feature
extraction for different
domain phenotypes and
can be compared with the
proposed method result
80 RGB images and 400
spectral images of
different dates and
specific time slot
Feature extraction from
plant height and
vegetation index
calculation prediction
Other
[95] Vegetation indices
performance is dependent
on canopy volume and
crop yield
Data was collected on
four different dates, and
90 images were collected
for each flight
Unmanned aerial
vehicle mounted
cost-effective
multispectral camera
for coffee ripeness
Multi
(Continued)
2990 CMC, 2024, vol.78, no.3
Table 6 (continued)
Article Issues/Challenges Data set Worked on Image type
[66] Feature extraction can be
improved
15 citrus trees Low-altitude unmanned
aerial vehicle using a
multispectral camera
with five bands RGB,
Red Edge, and
near-infrared
Multi
[57] The proposed method can
be applied on other seeds
as well
500 Julong and 500
Xiali seeds were used
Uses of multispectral
imagery combined with
chemometrics
Both
[81] The generic algorithm can
be implemented using the
proposed method
414 samples from 24
different fields
performance
comparison of
vegetation indices
derived from both multi
and hyperspectral
imagery
Hyper
Some models display consistent errors across all features, emphasizing the necessity for improving
feature selection techniques. When conducting a comparative analysis of classification algorithms for
a particular problem, it becomes evident that the learning process tends to be sluggish. However,
this can be accelerated by isolating the feature selection process and applying the method to other
fruits. Enriching the dataset with additional disease samples and testing it on various deep-learning
architectures can help overcome the challenge of limited training samples.
5Challenges and Future Directions
Multispectral and hyperspectral images offer data spanning multiple bands or narrow spectral
channels, enabling the capture of extensive spectral information about the Earth’s surface or other
observed targets. However, the analysis and interpretation of such data can be daunting due to its high
dimensionality noise and artifacts. Machine learning and deep learning algorithms have demonstrated
promising outcomes in addressing these challenges. They efficiently tackle spectral data’s compu-
tational complexity and high dimensionality, facilitating automated analysis and interpretation. A
crucial aspect of hyperspectral image analysis is feature extraction and machine learning algorithms
excel in automatically learning and performing this task. This ability becomes particularly significant
due to the large number of spectral bands in hyperspectral data. Here are some challenges that we find
in our primary studies.
5.1 Small Datasets
The small size of datasets is indeed a standard limitation when working with spectral images in
the context of agricultural fruits [68,77,87]. Obtaining and annotating large-scale datasets specific to
agricultural fruits can be challenging due to various factors [73]. Here are some of the few reasons why
the datasets related to spectral images were often small.
CMC, 2024, vol.78, no.3 2991
5.1.1 Data Collection and Annotation
Acquiring high-quality spectral images of agriculture requires specialized equipment, such as
hyperspectral sensors or multispectral cameras. Conducting field data collection campaigns can be
time-consuming and resource-intensive [69]. Additionally, accurately label and annotate the collected
images with ground truth information, such as fruit type, ripeness, or disease status.
5.1.2 Variability and Diversity
Spectral images exhibit significant variability in size, shape, colour, and texture. Capturing this
variability across different cultivars, growing conditions, and stages of development requires a diverse
dataset [84,85]. However, collecting a representative sample of this variability is challenging, leading
to relatively small datasets that may not fully capture the entire agricultural population’s variation.
5.1.3 Privacy and Proprietary Concerns
Agricultural datasets are sometimes limited due to privacy concerns or proprietary reasons.
Agricultural companies or research institutions may hesitate to share their proprietary datasets due
to intellectual property protection or competitive advantages [62,64]. This can limit the availability of
large-scale datasets for public use. Cost constraints: Collecting and maintaining large-scale datasets
can be costly. The equipment, human resources, and infrastructure required for data collection,
storage, and annotation can impose financial constraints, particularly for research projects or organi-
zations with limited resources. We can employ techniques such as semi-supervised learning or transfer
learning. Semi-supervised learning leverages both labelled and unlabeled data for training. Transfer
learning involves pre-training a model on a related dataset with more data and then fine-tuning it on
the target dataset with limited labelled samples. We can also implement data augmentation techniques
to artificially increase the training dataset’s diversity. This can involve applying transformations such
as rotation, scaling, and brightness adjustment to account for variability.
5.2 Limited Environment and Camera Bandwidth
These factors can impact the quality and representativeness of the datasets. Here is a closer look
at these limitations.
5.2.1 Limited Environment of Data Collection
Spectral imaging in agricultural settings is often conducted in specific environments, such as
controlled greenhouse conditions or selected field sites [72]. This limited data collection environment
can affect the generalizability of the models trained on these datasets. Agricultural conditions, such as
lighting, soil types, weather patterns, and cultivation practices, can vary across different regions and
seasons, making it challenging to capture the full range of environmental factors that influence fruit
characteristics.
5.2.2 Camera Bandwidth Limitations
Spectral cameras for capturing multispectral or hyperspectral images have specific bandwidth
ranges for each spectral band. These bandwidth limitations can affect the spectral resolution and
accuracy of the captured images [78]. Narrow bandwidths may limit the ability to capture subtle
spectral variations and nuances related to fruit properties, such as ripeness, disease, or nutritional
content. This can lead to incomplete or less precise spectral information, which may impact the
performance of analysis and classification algorithms. Researchers often employ strategies such as
2992 CMC, 2024, vol.78, no.3
data augmentation to mitigate these limitations. By applying transformations or modifications to the
available dataset, data augmentation techniques can help increase the diversity and size of the training
data. This can help to some extent in compensating for the limited environment of data collection by
introducing variations that mimic different environmental conditions or scenarios.
Cross-validation and external validation can be employed to assess the generalizability of models
trained on limited datasets. This involves partitioning the dataset into subsets for training and
validation, ensuring that the models are evaluated on data they have not seen during training.
Additionally, external validation on independent datasets collected in different environments or
by different research groups can provide a better understanding of the model’s performance and
robustness.
Collaborative data collection among researchers, institutions, and stakeholders in the agricultural
domain can help overcome the limitations of limited data collection environments. By pooling
resources, sharing expertise, and conducting multi-site data collection campaigns, more extensive and
diverse datasets can be obtained, capturing a wider range of environmental conditions and fruit
characteristics. Camera and sensor advancements in spectral imaging technology, including camera
bandwidth and resolution improvements, can help mitigate the limitations imposed by limited camera
capabilities. Higher spectral resolution and broader bandwidths can provide more detailed and
accurate spectral information, enabling better discrimination of fruit properties and improving the
overall performance of analysis [56].
To overcome these, we can consider fusing spectral data with higher-resolution spatial data, such
as RGB or multispectral imagery, to combine the spectral richness of the data with the spatial detail
of higher-resolution imagery. Alternatively, explore super-resolution techniques to enhance spatial
resolution.
5.3 Data Scarcity
Data scarcity can pose significant challenges, particularly in machine learning. These are some
major aspects and issues that arise while facing data scarcity.
5.3.1 Limited Model Training
Data scarcity can make it challenging to train robust models [74], leading to overfitting (where the
model fits the training data too closely [87] and performs poorly on new data) or underfitting (where
the model cannot capture the underlying patterns in the data).
5.3.2 Reduced Model Performance
With limited data, it is often challenging to achieve high model performance. Models trained on
small datasets may not capture the complexities of the underlying problem, resulting in suboptimal
results [55,95].
5.3.3 Privacy and Security
Data scarcity can arise due to privacy concerns and regulations in sensitive domains, such as
healthcare or finance. Access to sufficient data for research or analysis may be restricted to protect
individuals’ privacy.
CMC, 2024, vol.78, no.3 2993
5.3.4 Transfer Learning Challenges
Machine learning techniques like transfer learning and domain adaptation are often used to
leverage knowledge from related domains. Data scarcity can limit the effectiveness of these techniques,
as they rely on having at least some data available in both the source and target domains [73,92].
Different strategies can be implemented to avoid data scarcity issues.
Data Augmentation: We can implement different data augmentation techniques to expand the
dataset through translation and rotation by adding some noise to the existing data.
Transfer Learning: Adapt models on larger, related datasets to the target task with limited data.
Active Learning: Choose those queries and label the most informative data.
Semi-Supervised Learning: Combine labelled data with unlabeled data to make the most available
information.
Domain Adaptation: Adapt models to perform well in the target domain by mitigating domain
shift issues.
Data Synthesis: Use generative models to create synthetic data resembling the target domain.
Ultimately, addressing data scarcity requires careful consideration of the specific domain and
problem at hand and creative solutions to make the most of the available data and resources.
5.4 Machine Learning and Deep Learning: Promises and Limitations
KNN was used for the classification of the rose plant in [59] and citrus greening detection in [69]
and in [70] for the regression assessment of ready-to-eat pineapple quality. The classification accuracy
was 65% and 81%, and the R-square value was 0.42, respectively. The challenge is to test different
algorithms on the same type of problems at different stages of growth of the particular plant.
Citrus greening detection [6669], pineapple quality [70], strawberry ripeness [77], tomato grading
[65], and high-quality watermelon seeds classification were done by SVM with accuracy ranging from
79% to 98%. Testing new algorithms, refining feature selection, and more diverse sample images at
various ripeness stages can enhance results. It is noted that accuracy may decrease as the number
of target classes increases. Nevertheless, SVM holds potential for broader applications, including
analyzing different seeds and crops, making it a promising avenue for future research and development.
Deep Learning and different types of neural networks were used for the detection of fruit diseases
[83], vine disease detection [68], detection of green vegetation cover [84], detection of growth stages
of rice [72], fruit recognition [73], apple defects [74], blossom detection in apple trees [75], tomato’s
external defects [76] and strawberry disease classification [94]. The accuracy achieved was in the range
of 65% to 99%. One of the main issues was the slow learning process, which can be enhanced by
choosing the meta-heuristic techniques for feature selection. The other main challenge is adaptability.
We can test the same model on the new dataset. We can enhance the adaptability by updating the
dataset with new models on different agriculture problems.
6Conclusion
This paper investigates a wide range of applications based on spectral images. It focuses on relevant
studies from reputable databases, which were carefully chosen to address the four main research
questions of the study. The study examined various applications utilizing spectral images, particularly
agriculture. The findings reveal that Deep Learning is the most prevalent technique for spectral image
2994 CMC, 2024, vol.78, no.3
classification, followed by SVM as the second optimal approach. Furthermore, the study identified
popular datasets relevant to agricultural fruits, with Indian pines, Salinas, and Houston being the most
widely used.
Additionally, the Pavia University and Kennedy Centre datasets were frequently employed
in certain research. The study also acknowledged the limitations of the restricted data collection
environment and camera bandwidth. Despite the constraints of small datasets, researchers have made
notable strides in developing effective machine learning and deep learning models for agricultural
fruit analysis. As the availability and quality of spectral image datasets increase, the potential for
more accurate and robust analysis and applications in the agricultural sector also improves. Indeed,
the limited environment of data collection and camera bandwidth represents additional challenges
when working with spectral images in agricultural fruit analysis. Nevertheless, ongoing research and
technological advancements in spectral imaging and agricultural fruit analysis aim to address these
issues and enhance the dataset’s quality, diversity, and representativeness for machine learning and
deep learning applications.
Acknowledgement: We express our gratitude to the upper management of the Karachi Institute of
Economics and Technology (KIET) for furnishing the technical resources necessary for the execution
of this research. Additionally, we extend our appreciation to Multimedia University for their generous
financial backing of this research.
Funding Statement: The authors received no specific funding for this study.
Author Contributions: The authors confirm their contribution to the paper as follows: study conception
and design: Usman Khan, Muhammad Khalid Khan; data collection: Muhammad Ayub Latif;
analysis and interpretation of results: Usman Khan, Muhammad Khalid Khan; draft manuscript
preparation: Usman Khan, Muhammad Khalid Khan, Muhammad Ayub Latif, Muhammad Naveed,
Salman A. Khan, Muhammad Mansoor Alam, Mazliham Mohd Su’ud. All authors reviewed the
results and approved the final version of the manuscript.
Availability of Data and Materials: Data available within the article. The authors confirm that the data
supporting the findings of this study are available within the article.
Conflicts of Interest: The authors declare that they have no conf licts of interest to report regarding the
present study.
References
[1] A. Vibhute and S. K. Bodhe, “Applications of image processing in agriculture: A survey,”Int. J. Comput.
Appl., vol. 52, no. 2, pp. 34–40, Aug. 2012. doi: 10.5120/8176-1495.
[2] W. Gong et al., “Investigating the potential of using the spatial and spectral information of mul-
tispectral LiDAR for object classification,” Sens., vol. 15, no. 9, pp. 21989–22002, Sep. 2015. doi:
10.3390/s150921989.
[3] H. Zheng et al., “A comparative assessment of different modeling algorithms for estimating leaf nitrogen
content in winter wheat using multispectral images from an unmanned aerial vehicle,” Remote Sens.,vol.
10, no. 12, pp. 2026, 2018. doi: 10.3390/rs10122026.
[4] H. Tian, T. Wang, Y. Liu, X. Qiao, and Y. Li, “Computer vision technology in agricultural automation—A
review,” Inf. Process. Agric., vol. 7, no. 1, pp. 1–19, Mar. 2020. doi: 10.1016/j.inpa.2019.09.006.
CMC, 2024, vol.78, no.3 2995
[5] T. Adão et al., “Hyperspectral imaging: A review on UAV-based sensors, data processing and applications
for agriculture and forestry,” Remote Sens., vol. 9, no. 11, pp. 1110, 2017. doi: 10.3390/rs9111110.
[6] J. Hu et al., “Quantitative estimation of soil salinity using UAV-borne hyperspectral and satellite multi-
spectral images,” Remote Sens., vol. 11, no. 7, pp. 736, 2019. doi: 10.3390/rs11070736.
[7] J. Vargas, S. Alsweiss, O. Toker, R. Razdan, and J. Santos, “An overview of autonomous vehicles
sensors and their vulnerability to weather conditions,” Sens., vol. 21, no. 16, pp. 5397, Aug. 2021. doi:
10.3390/s21165397.
[8] M. Himmler, D. W. Schubert, and T. A. Fuchsluger, “Examining the transmission of visible light through
electrospun nanofibrous PCL scaffolds for corneal tissue engineering,” Nanomaterials, vol. 11, no. 12, pp.
3191, Nov. 2021. doi: 10.3390/nano11123191.
[9] K. J. Fan, and W. H. Su, Applications of fluorescence spectroscopy, RGB- and multispectral imaging
for quality determinations of white meat: A review,” Biosens., vol. 12, no. 2, pp. 76, Jan. 2022. doi:
10.3390/bios12020076.
[10] J. Nataprawira, Y. Gu, I. Goncharenko, and S. Kamijo, “Pedestrian detection using multispectral images
and a deep neural network,” Sens., vol. 21, no. 7, pp. 1–17, 2021. doi: 10.3390/s21072536.
[11] M. D. Iordache, V. Mantas, E. Baltazar, K. Pauly, and N. Lewyckyj, “A machine learning approach to
detecting pine wilt disease using airborne spectral imagery,” Remote Sens., vol. 12, no. 14, pp. 2280, Jul.
2020. doi: 10.3390/rs12142280.
[12] S. Baek, E. Choi, Y. Baek, and C. Lee, “Detection of counterfeit banknotes using multispectral images,”
Digit Signal Process., vol. 78, no. 2, pp. 294–304, Jul. 2018. doi: 10.1016/j.dsp.2018.03.015.
[13] A. Mukundan, Y. M. Tsao, W. M. Cheng, F. C. Lin, and H. C. Wang, “Automatic counterfeit currency
detection using a novel snapshot hyperspectral imaging algorithm,” Sens., vol. 23, no. 4, pp. 2026, Feb.
2023. doi: 10.3390/s23042026.
[14] K. Kang and C. Lee, “Fake banknote detection using multispectral images,” in 2016 7th Int. Conf. Inf.,
Intelli., Syst. Appli. (IISA), Chalkidiki, Greece, IEEE, Jul. 2016, pp. 1–3. doi: 10.1109/IISA.2016.7785338.
[15] S. L. P. Aggarwal and F. A. Papay, Applications of multispectral and hyperspectral imaging in dermatol-
ogy,” Exp. Dermatol., vol. 31, no. 8, pp. 1128–1135, Aug. 2022. doi: 10.1111/exd.14624.
[16] W. Huang, C. Zhao, Q. Wang, J. Li, and C. Zhang, “Development of a multi-spectral imaging system for
the detection of bruises on apples,” Sens. Agric. Food Qual. Saf., vol. 8721, pp. 872105, May 2013. doi:
10.1117/12.2015579.
[17] A. K. Chandel, L. R. Khot, and C. S. Bernardita, “Apple powdery mildew infestation detection and
mapping using high-resolution visible and multispectral aerial imaging technique,” Sci Hortic, vol. 287,
no. 6, pp. 110228, Sep. 2021. doi: 10.1016/j.scienta.2021.110228.
[18] A. Bhargava and A. Bansal, “Quality evaluation of mono & bi-colored apples with computer vision
and multispectral imaging,” Multimed. Tools Appl., vol. 79, no. 11–12, pp. 7857–7874, 2020. doi:
10.1007/s11042-019-08564-3.
[19] J. Feng, L. Zeng, and L. He, “Apple fruit recognition algorithm based on multi-spectral dynamic image
analysis,” Sens., vol. 19, no. 4, pp. 949, Feb. 2019. doi: 10.3390/s19040949.
[20] I. Amro, J. Mateos, M. Vega, R. Molina, and A. K. Katsaggelos, “A survey of classical methods and new
trends in pansharpening of multispectral images,” Eurasip. J. Adv. Sig. Process., vol. 2011, no. 1, pp. 1–22,
2011. doi: 10.1186/1687-6180-2011-79.
[21] I. Racetin and A. Krtali´
c, “Systematic review of anomaly detection in hyperspectral remote sensing
applications,” Applied Sciences, vol. 11, no. 11, pp. 4878, 2021. doi: 10.3390/app11114878.
[22] E. Aloupogianni, M. Ishikawa, N. Kobayashi, and T. Obi, “Hyperspectral and multispectral image
processing for gross-level tumor detection in skin lesions: A systematic review,” J. Biomed. Opt., vol. 27,
no. 6, pp. 1–28, 2022. doi: 10.1117/1.JBO.27.6.060901.
[23] S. Ortega, M. Halicek, H. Fabelo, G. M. Callico, and B. Fei, “Hyperspectral and multispectral imaging
in digital and computational pathology: A systematic review [Invited],” Biomed. Opt. Exp., vol. 11, no. 6,
pp. 3195, 2020. doi: 10.1364/BOE.386338.
2996 CMC, 2024, vol.78, no.3
[24] A. Kaul and S. Raina, “Support vector machine versus convolutional neural network for hyperspectral
image classification: A systematic review,” Concurr. Comput.: Pract. Exp., vol. 34, no. 15, pp. e6945, 2022.
doi: 10.1002/cpe.6945.
[25] D. Datta et al., “Hyperspectral image classification: Potentials, challenges, and future directions,” Comput.
Intelli. Neurosci., vol. 2022, no. 11, pp. 1–36, 2022. doi: 10.1155/2022/3854635.
[26] A. Ozdemir and K. Polat, “Deep learning applications for hyperspectral imaging: A systematic review,”
J. Inst. Electron. Comput., vol. 2, no. 1, pp. 39–56, 2020. doi: 10.33969/JIEC.2020.21004.
[27] D. Lu and Q. Weng, “A survey of image classification methods and techniques for improving classification
performance,” Int. J. Remote Sens., vol. 28, no. 5, pp. 823–870, 2007. doi: 10.1080/01431160600746456.
[28] H. Lee, J. Wang, and B. Leblon, “Using linear regression, random forests, and support vector machine
with unmanned aerial vehicle multispectral images to predict canopy nitrogen weight in corn,” Remote
Sens., vol. 12, no. 13, pp. 2071, 2020. doi: 10.3390/rs12132071.
[29] K. Osorio, A. Puerto, C. Pedraza, D. Jamaica, and L. Rodríguez, “A deep learning approach for weed
detection in lettuce crops using multispectral images,” AgriEng., vol. 2, no. 3, pp. 471–488, Sep. 2020. doi:
10.3390/agriengineering2030032.
[30] M. S. Ozigis, J. D. Kaduk, C. H. Jarvis, P. da Conceição Bispo, and H. Balzter, “Detection of oil pollution
impacts on vegetation using multi-frequency SAR, multispectral images with fuzzy forest and random
forest methods,” Environ. Pollut., vol. 256, no. 2, pp. 113360, Jan. 2020. doi: 10.1016/j.envpol.2019.113360.
[31] M. C. F. Lima, A. Krus, C. Valero, A. Barrientos, J. Del Cerro and J. J. Roldán-Gómez, “Monitoring
plant status and fertilization strategy through multispectral images,” Sens., vol. 20, no. 2, pp. 435, 2020.
doi: 10.3390/s20020435.
[32] Y. Cao, G. L. Li, Y. K. Luo, Q. Pan, and S. Y. Zhang, “Monitoring of sugar beet growth indicators
using wide-dynamic-range vegetation index (WDRVI) derived from UAV multispectral images,” Comput.
Electron. Agric., vol. 171, pp. 105331, 2020. doi: 10.1016/j.compag.2020.105331.
[33] C. M. Sicre, R. Fieuzal, and F. Baup, “Contribution of multispectral (optical and radar) satellite images
to the classification of agricultural surfaces,” Int. J. Appl. Earth Obs., vol. 84, pp. 101972, Jul. 2019. doi:
10.1016/j.jag.2019.101972.
[34] H. Qi et al., “Estimation of peanut leaf area index from unmanned aerial vehicle multispectral images,”
Sens., vol. 20, no. 23, pp. 1–15, 2020. doi: 10.3390/s20236732.
[35] W. Zhang, Q. Zhu, M. Huang, Y. Guo, and J. Qin, “Detection and classification of potato defects using
multispectral imaging system based on single shot method,”Food Anal. Method, vol. 12, no. 12, pp. 2920–
2929, 2019. doi: 10.1007/s12161-019-01654-w.
[36] J. C. Noordam, W. H. van den Broek, and L. M. Buydens, “Perspective of inline control of latent defects
and diseases on French fries with multispectral imaging,” Monitoring Food Safety, Agric. Plant Health,
vol. 5271, pp. 85, 2004. doi: 10.1117/12.515371.
[37] Y. Yang, X. Zhao, M. Huang, X. Wang, and Q. Zhu, “Multispectral image based germination detection
of potato by using supervised multiple threshold segmentation model and Canny edge detector,”Comput.
Electron. Agric., vol. 182, no. 7, pp. 106041, 2021. doi: 10.1016/j.compag.2021.106041.
[38] S. Khanal, J. Fulton, and S. Shearer, “An overview of current and potential applications of thermal
remote sensing in precision agriculture,” Comput. Electron. Agric., vol. 139, pp. 22–32, Jun. 2017. doi:
10.1016/j.compag.2017.05.001.
[39] S. Ranjan, D. R. Nayak, K. S. Kumar, R. Dash, and B. Majhi, “Hyperspectral image classification: A
k-means clustering based approach,” in 2017 4th Int. Conf. Adv. Comput. Commun. Syst., ICACCS 2017,
2017, pp. 6. doi: 10.1109/ICACCS.2017.8014707.
[40] Y. Zhang et al., “Estimating the maize biomass by crop height and narrowband vegetation indices
derived from UAV-based hyperspectral images,” Ecol. Indic., vol. 129, no. 12, pp. 107985, 2021. doi:
10.1016/j.ecolind.2021.107985.
[41] J. Rubio-Delgado, C. J. Pérez, and M. A. Vega-Rodríguez, “Predicting leaf nitrogen content in olive trees
using hyperspectral data for precision agriculture,” Precis. Agric., vol. 22, no. 1, pp. 1–21, 2021. doi:
10.1007/s11119-020-09727-1.
CMC, 2024, vol.78, no.3 2997
[42] G. Messina and G. Modica, “The role of remote sensing in olive growing farm management: A research
outlook from 2000 to the present in the framework of precision agriculture applications,” Remote.,vol.
14, no. 23, pp. 5951, Nov. 2022. doi: 10.3390/rs14235951.
[43] H. Yao, R. Qin, and X. Chen, “Unmanned aerial vehicle for remote sensing applications—A review,”
Remote Sens., vol. 11, no. 12, pp. 1443, Jun. 2019. doi: 10.3390/rs11121443.
[44] D. Gautam and V. Pagay, “A review of current and potential applications of remote sensing to study
the water status of horticultural crops,” Agron., vol. 10, no. 1, pp. 140, Jan. 2020. doi: 10.3390/agron-
omy10010140.
[45] C. Yang, J. H. Everitt, J. M. Bradford, and D. Murden, Airborne hyperspectral imagery and yield
monitor data for mapping cotton yield variability,” Precis. Agric., vol. 5, no. 5, pp. 445–461, 2004. doi:
10.1007/s11119-004-5319-8.
[46] Y. Zhang, L. Wu, L. Deng, and B. Ouyang, “Retrieval of water quality parameters from hyperspectral
images using a hybrid feedback deep factorization machine model,”Wat er Re s. , vol. 204, no. 3, pp. 117618,
2021. doi: 10.1016/j.watres.2021.117618.
[47] T. Mewes, J. Franke, and G. Menz, “Spectral requirements on airborne hyperspectral remote sens-
ing data for wheat disease detection,” Precis. Agric., vol. 12, no. 6, pp. 795–812, 2011. doi:
10.1007/s11119-011-9222-9.
[48] H. Cen, R. Lu, D. P. Ariana, and F. Mendoza, “Hyperspectral imaging-based classification and wavebands
selection for internal defect detection of pickling cucumbers,”Food Bioprocess Tech., vol. 7, no. 6, pp. 1689–
1700, 2014. doi: 10.1007/s11947-013-1177-6.
[49] P. Langley,“Machine learning and concept formation,” Mach. Learn., vol. 2, no. 2, pp. 99–102, Sep. 1987.
doi: 10.1007/BF00114263.
[50] M. Cui and D. Y. Zhang, “Artificial intelligence and computational pathology,” Lab. Invest., vol. 101, no.
4, pp. 412–422, Apr. 2021. doi: 10.1038/s41374-020-00514-0.
[51] B. Kitchenham et al., “Systematic literature reviews in software engineering—A tertiary study,” Inf.
Software Tech., vol. 52, no. 8, pp. 792–805, 2010. doi: 10.1016/j.infsof.2010.03.006.
[52] A. Lowe, N. Harrison, and A. P. French, “Hyperspectral image analysis techniques for the detection and
classification of the early onset of plant disease and stress,” Plant Methods, vol. 13, no. 1, pp. 1–12, 2017.
doi: 10.1186/s13007-017-0233-z.
[53] H. Gan, W. S. Lee, and V. Alchanatis, “An active thermography method for immature citrus fruit
detection,” Biosyst. Eng., vol. 198, pp. 291–303, 2018.
[54] K. Park, Y. K. Hong, G. H. Kim, and J. Lee, “Classification of apple leaf conditions in hyperspectral
images for diagnosis of Marssonina blotch using mRMR and deep neural network,” Comput. Electron.
Agric., vol. 148, no. 3, pp. 179–187, 2018. doi: 10.1016/j.compag.2018.02.025.
[55] P. A. Dias, A. Tabb, and H. Medeiros, “Apple flower detection using deep convolutional networks,”
Comput. Ind., vol. 99, no. 1, pp. 17–28, 2018. doi: 10.1016/j.compind.2018.03.010.
[56] A. Wendel, J. Underwood, and K. Walsh, “Maturity estimation of mangoes using hyperspectral imaging
from a ground based mobile platform,” Comput. Electron. Agric., vol. 155, no. 1, pp. 298–313, 2018. doi:
10.1016/j.compag.2018.10.021.
[57] W. Liu, X. Xu, C. Liu, and L. Zheng, “Rapid discrimination of high-quality watermelon seeds by
multispectral imaging combined with chemometric methods,” J. Appl. Spectrosc., vol. 85, no. 6, pp. 1044–
1049, 2019. doi: 10.1007/s10812-019-00757-w.
[58] S. Kaur, S. Pandey, and S. Goel, “Plants disease identification and classification through leaf images: A
survey,” Arch. Comput. Method Eng., vol. 26, no. 2, pp. 507–530, 2019. doi: 10.1007/s11831-018-9255-6.
[59] M. Malik, A. Ikram, S. N. Batool, and W. Aslam, “A performance assessment of rose plant classification
using machine learning,” in INTAP 2018: Intelligent Technologies and Applications, 2019, pp. 745–756.
doi: 10.1007/978-981-13-6052-7_64.
[60] J. Li et al., “Detection of early decayed oranges based on multispectral principal component image
combining both bi-dimensional empirical mode decomposition and watershed segmentation method,”
Postharvest Biol. Technol., vol. 158, no. 4, pp. 110986, 2019. doi: 10.1016/j.postharvbio.2019.110986.
2998 CMC, 2024, vol.78, no.3
[61] J. G. A. Barbedo, “Detection of nutrition deficiencies in plants using proximal images and machine
learning: A review,” Comput. Electron. Agric., vol. 162, no. 7, pp. 482–492, 2019. doi: 10.1016/j.compag.
2019.04.035.
[62] J. Gené-Mola, V. Vilaplana, J. R. Rosell-Polo, J. R. Morros, J. Ruiz-Hidalgo and E. Gregorio, “Multi-
modal deep learning for Fuji apple detection using RGB-D cameras and their radiometric capabilities,”
Comput. Electron. Agric., vol. 162, pp. 689–698, 2019. doi: 10.1016/j.compag.2019.05.016.
[63] J. Steinbrener, K. Posch, and R. Leitner, “Hyperspectral fruit and vegetable classification using con-
volutional neural networks,” Comput. Electron. Agric., vol. 162, no. 6, pp. 364–372, Oct. 2018. doi:
10.1016/j.compag.2019.04.019.
[64] J. Gené-Mola et al., “Fruit detection in an apple orchard using a mobile terrestrial laser scanner,”Biosyst.
Eng., vol. 187, no. 34, pp. 171–184, 2019. doi: 10.1016/j.biosystemseng.2019.08.017.
[65] D. Ireri, E. Belal, C. Okinda, N. Makange, and C. Ji, “A computer vision system for defect discrimination
and grading in tomatoes using machine learning and image processing,” Artif. Intelli. Agric.,vol.2,no.1,
pp. 28–37, 2019. doi: 10.1016/j.aiia.2019.06.001.
[66] F. DadrasJavan, F. Samadzadegan, S. H. S. Pourazar, and H. Fazeli, “UAV-based multispectral imagery
for fast citrus greening detection,” J. Plant Dis. Protect., vol. 126, no. 4, pp. 307–318, 2019. doi: 10.1007/
s41348-019-00234-8.
[67] S. S. Chouhan, U. P. Singh, and S. Jain, “Applications of computer vision in plant pathology: A survey,”
Arch. Comput. Method Eng., vol. 27, no. 2, pp. 611–632, 2020. doi: 10.1007/s11831-019-09324-0.
[68] M. Kerkech, A. Hafiane, and R. Canals, “Vine disease detection in UAV multispectral images using
optimized image registration and deep learning segmentation approach,” Comput. Electron. Agric.,vol.
174, no. 1, pp. 105446, 2020. doi: 10.1016/j.compag.2020.105446.
[69] Y. Lan et al., “Comparison of machine learning methods for citrus greening detection on UAV mul-
tispectral images,” Comput. Electron. Agric., vol. 171, no. 5, pp. 105234, 2020. doi: 10.1016/j.compag.
2020.105234.
[70] E. Manthou et al., “Application of spectroscopic and multispectral imaging technologies on the assessment
of ready-to-eat pineapple quality: A performance evaluation study of machine learning models generated
from two commercial data analytics tools,” Comput. Electron. Agric., vol. 175, pp. 105529, 2020. doi:
10.1016/j.compag.2020.105529.
[71] J. Xiong et al., “Visual detection of green mangoes by an unmanned aerial vehicle in orchards
based on a deep learning method,” Biosyst. Eng., vol. 194, no. 4, pp. 261–272, 2020. doi: 10.1016/j.
biosystemseng.2020.04.006.
[72] Q. Yang, L. Shi, J. Han, J. Yu, and K. Huang, “A near real-time deep learning approach for detect-
ing rice phenology based on UAV images,” Agric. Forest Meteorol., vol. 287, pp. 107938, 2020. doi:
10.1016/j.agrformet.2020.107938.
[73] S. I. Saedi and H. Khosravi, “A deep neural network approach towards real-time on-branch fruit recog-
nition for precision horticulture,” Exp. Syst. Appl., vol. 159, no. 4, pp. 113594, 2020. doi: 10.1016/j.eswa.
2020.113594.
[74] S. Fan et al., “On line detection of defective apples using computer vision system combined with deep
learning methods,” J. Food Eng., vol. 286, no. 5, pp. 110102, 2020. doi: 10.1016/j.jfoodeng.2020.110102.
[75] U. Bhattarai, S. Bhusal, Y. Majeed, and M. Karkee, “Automatic blossom detection in apple trees using
deep learning,” Int. Fed. Automatic Control-PapersOnLine, vol. 53, no. 2, pp. 15810–15815, 2020. doi:
10.1016/j.ifacol.2020.12.216.
[76] A. Z. da Costa, H. E. H. Figueroa, and J. A. Fracarolli, “Computer vision based detection of exter-
nal defects on tomatoes using deep learning,” Biosyst. Eng., vol. 190, no. 3, pp. 131–144, 2020. doi:
10.1016/j.biosystemseng.2019.12.003.
[77] Z. Gao, Y. Shao, G. Xuan, Y. Wang, Y. Liu and X. Han, “Real-time hyperspectral imaging for the in-field
estimation of strawberry ripeness with deep learning,” Artif. Intelli. Agric., vol. 4, no. 6, pp. 31–38, 2020.
doi: 10.1016/j.aiia.2020.04.003.
CMC, 2024, vol.78, no.3 2999
[78] J. Wu et al., “Extracting apple tree crown information from remote imagery using deep learning,” Comput.
Electron. Agric., vol. 174, no. 4, pp. 105504, 2020. doi: 10.1016/j.compag.2020.105504.
[79] Z. Du, X. Zeng, X. Li, X. Ding, J. Cao and W. Jiang, “Recent advances in imaging techniques for bruise
detection in fruits and vegetables,” Trends Food Sci. Technol., vol. 99, no. 1, pp. 133–141, 2020. doi:
10.1016/j.tifs.2020.02.024.
[80] X. Hu, L. Yang, and Z. Zhang, “Non-destructive identification of single hard seed via multispec-
tral imaging analysis in six legume species,” Plant Methods, vol. 16, no. 1, pp. 1–13, 2020. doi:
10.1186/s13007-020-00659-5.
[81] L. A. Suarez, A. Robson, J. McPhee, J. O’Halloran, and C. van Sprang, Accuracy of carrot yield
forecasting using proximal hyperspectral and satellite multispectral data,” Precis. Agric., vol. 21, no. 6,
pp. 1304–1326, 2020. doi: 10.1007/s11119-020-09722-6.
[82] P. Chithra and M. Henila, “Apple fruit sorting using novel thresholding and area calculation algorithms,”
Soft Comput., vol. 25, no. 1, pp. 431–445, 2021. doi: 10.1007/s00500-020-05158-2.
[83] H. Jiang, X. Li, and F. Safara, “IoT-based agriculture: Deep learning in detecting apple fruit diseases,”
Microprocess. Microsyst., vol. 143, no. 2, pp. 104321, 2021. doi: 10.1016/j.micpro.2021.104321.
[84] L. El Hoummaidi, A. Larabi, and K. Alam, “Using unmanned aerial systems and deep learning for agri-
culture mapping in Dubai,” Heliyon, vol. 7, no. 10, pp. e08154, 2021. doi: 10.1016/j.heliyon.2021.e08154.
[85] X. Zhou, W. S. Lee, Y. Ampatzidis, Y. Chen, N. Peres and C. Fraisse, “Strawberry maturity classification
from UAV and near-ground imaging using deep learning,”Smart Agric. Technol., vol. 1, no. 16, pp. 100001,
2021. doi: 10.1016/j.atech.2021.100001.
[86] Z. Zhou, Y. Majeed, G. D. Naranjo, and E. M. T. Gambacorta, Assessment for crop water stress with
infrared thermal imagery in precision agriculture: A review and future prospects for deep learning appli-
cations,” Comput. Electron. Agric., vol. 182, no. 5, pp. 106019, 2021. doi: 10.1016/j.compag.2021.106019.
[87] J. Gao, J. C. Westergaard, E. H. R. Sundmark, M. Bagge, E. Liljeroth and E. Alexandersson, “Automatic
late blight lesion recognition and severity quantification based on field imagery of diverse potato genotypes
by deep learning,” Knowl-Based Syst., vol. 214, pp. 106723, 2021. doi: 10.1016/j.knosys.2020.106723.
[88] D. Wu et al., “A deep learning-integrated micro-CT image analysis pipeline for quantifying rice lodging
resistance-related traits,” Plant Commun., vol. 2, no. 2, pp. 100165, 2021. doi: 10.1016/j.xplc.2021.100165.
[89] G. Sachdeva, P. Singh, and P. Kaur, “Plant leaf disease classification using deep convolutional
neural network with Bayesian learning,” Mater. Today: Proc., vol. 45, pp. 5584–5590, 2021. doi:
10.1016/j.matpr.2021.02.312.
[90] B. D. S. Barbosa et al., “UAV-based coffee yield prediction utilizing feature selection and deep learning,”
Smart Agric. Technol., vol. 1, no. 17, pp. 100010, 2021. doi: 10.1016/j.atech.2021.100010.
[91] J. D. Arango, B. Staar, A. M. Baig, and M. Freitag, “Quality control of apples by means of convolutional
neural networks—Comparison of bruise detection by color images and near-infrared images,” Proc.
CIRP, vol. 99, no. 7, pp. 290–294, 2021. doi: 10.1016/j.procir.2021.03.043.
[92] K. Tatsumi, N. Igarashi, and X. Mengxue, “Prediction of plant-level tomato biomass and yield using
machine learning with unmanned aerial vehicle imagery,” Plant Methods, vol. 17, no. 1, pp. 1–17, 2021.
doi: 10.1186/s13007-021-00761-2.
[93] X. Fu and M. Wang, “Detection of early bruises on pears using fluorescence hyperspectral imaging
technique,” Food Anal. Method, vol. 15, no. 1, pp. 115–123, 2022. doi: 10.1007/s12161-021-02092-3.
[94] G. F. Yang, Y. Yang, Z. K. He, X. Y. Zhang, and Y. He, “A rapid, low-cost deep learning system to
classify strawberry disease based on cloud service,”J. Integr. Agric., vol. 21, no. 2, pp. 460–473, 2022. doi:
10.1016/S2095-3119(21)63604-3.
[95] J. T. F. Rosas, F. de A. de Carvalho Pinto, D. M. de Queiroz, F. M. de Melo Villar, D. S. Magalhães Valente
and R. Nogueira Martins, “Coffee ripeness monitoring using a UAV-mounted low-cost multispectral
camera,” Precis. Agric., vol. 23, no. 1, pp. 300–318, 2022. doi: 10.1007/s11119-021-09838-3.
[96] F. DadrasJavan, F. Samadzadegan, S. H. S. Pourazar, and H. Fazeli, “UAV-based multispectral imagery
for fast citrus greening detection,” J. Plant Dis. Prot., vol. 126, no. 4, pp. 307–318, Aug. 2019. doi:
10.1007/s41348-019-00234-8.
3000 CMC, 2024, vol.78, no.3
[97] Q. Gao, S. Lim, and X. Jia, “Hyperspectral image classification using convolutional neural networks and
multiple feature learning,” Remote Sens., vol. 10, no. 2, pp. 299, Feb. 2018. doi: 10.3390/rs10020299.
[98] A. Samat, P. Du, S. Liu, J. Li, and L. Cheng, “Ensemble extreme learning machines for hyperspectral
image classification,” IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., vol. 7, no. 4, pp. 1060–1069, Apr.
2014. doi: 10.1109/JSTARS.2014.2301775.
[99] F. Luo, Y. Huang, W. Tu, and J. Liu, “Local manifold sparse model for image classification,” Neurocom-
put., vol. 382, no. 6, pp. 162–173, Mar. 2020. doi: 10.1016/j.neucom.2019.11.084.
[100] Z. Chen, J. Jiang, X. Jiang, X. Fang, and Z. Cai, “Spectral-spatial feature extraction of hyperspectral
images based on propagation filter,” Sens., vol. 18, no. 6, pp. 1978, Jun. 2018. doi: 10.3390/s18061978.
ResearchGate has not been able to resolve any citations for this publication.
Article
Full-text available
In this study, a snapshot-based hyperspectral imaging (HSI) algorithm that converts RGB images to HSI images is designed using the Raspberry Pi environment. A Windows-based Python application is also developed to control the Raspberry Pi camera and processor. The mean gray values (MGVs) of two distinct regions of interest (ROIs) are selected from three samples of 100 NTD Taiwanese currency notes and compared with three samples of counterfeit 100 NTD notes. Results suggest that the currency notes can be easily differentiated on the basis of MGV values within shorter wavelengths, between 400 nm and 500 nm. However, the MGV values are similar in longer wavelengths. Moreover, if an ROI has a security feature, then the classification method is considerably more efficient. The key features of the module include portability, lower cost, a lack of moving parts, and no processing of images required.
Article
Full-text available
Given the importance of olive growing, especially in Mediterranean countries, it is crucial that there is a constant process of modernization aimed at both environmental sustainability and the maintenance of high standards of production. The use of remote sensing (RS) allows intervention in a specific and differentiated way in olive groves, depending on their variability, in managing different agronomic aspects. The potentialities of the application of RS in olive growing are topics of great agronomic interest to olive growers. Using the tools provided by RS and the modernization of the olive sector can bring great future prospects by reducing costs, optimizing agronomic management, and improving production quantity and quality. This article is part of a review that aims to cover the past, from the 2000s onwards, and the most recent applications of aerial RS in olive growing in order to be able to include research and all topics related to the use of RS on olive trees. As far as the use of RS platforms such as satellites, aircraft, and unmanned aerial vehicles (UAVs) as olive growing is concerned, a literature review showed the presence of several works devoted to this topic. This article covers purely agronomic matters of interest to olive farms (and related research that includes the application of RS), such as yielding and managing diseases and pests, and detection and counting of olive trees. In addition to these topics, there are other relevant aspects concerning the characterization of the canopy structure of olive trees which is particularly interesting for mechanized pruning management and phenotyping.
Article
Full-text available
Significance: Skin cancer is one of the most prevalent cancers worldwide. In the advent of medical digitization and telepathology, hyper/multispectral imaging (HMSI) allows for noninvasive, nonionizing tissue evaluation at a macroscopic level. Aim: We aim to summarize proposed frameworks and recent trends in HMSI-based classification and segmentation of gross-level skin tissue. Approach: A systematic review was performed, targeting HMSI-based systems for the classification and segmentation of skin lesions during gross pathology, including melanoma, pigmented lesions, and bruises. The review adhered to the 2020 Preferred Reporting Items for Systematic reviews and Meta-Analyses (PRISMA) guidelines. For eligible reports published from 2010 to 2020, trends in HMSI acquisition, preprocessing, and analysis were identified. Results: HMSI-based frameworks for skin tissue classification and segmentation vary greatly. Most reports implemented simple image processing or machine learning, due to small training datasets. Methodologies were evaluated on heavily curated datasets, with the majority targeting melanoma detection. The choice of preprocessing scheme influenced the performance of the system. Some form of dimension reduction is commonly applied to avoid redundancies that are inherent in HMSI systems. Conclusions: To use HMSI for tumor margin detection in practice, the focus of system evaluation should shift toward the explainability and robustness of the decision-making process.
Article
Full-text available
Recent imaging science and technology discoveries have considered hyperspectral imagery and remote sensing. The current intelligent technologies, such as support vector machines, sparse representations, active learning, extreme learning machines, transfer learning, and deep learning, are typically based on the learning of the machines. These techniques enrich the processing of such three-dimensional, multiple bands, and high-resolution images with their precision and fidelity. This article presents an extensive survey depicting machine-dependent technologies’ contributions and deep learning on landcover classification based on hyperspectral images. The objective of this study is three-fold. First, after reading a large pool of Web of Science (WoS), Scopus, SCI, and SCIE-indexed and SCIE-related articles, we provide a novel approach for review work that is entirely systematic and aids in the inspiration of finding research gaps and developing embedded questions. Second, we emphasize contemporary advances in machine learning (ML) methods for identifying hyperspectral images, with a brief, organized overview and a thorough assessment of the literature involved. Finally, we draw the conclusions to assist researchers in expanding their understanding of the relationship between machine learning and hyperspectral images for future research.
Article
Full-text available
Various machine learning and deep learning techniques have been proposed for classification purposes in the case of hyperspectral imaging. Among all the machine learning techniques support vector machine (SVM) has been a promising classification algorithm in the case of remote sensing applications, particularly in the field of hyperspectral imaging. In the case of deep learning techniques, the convolutional neural network is gaining much attention from researchers for classification purposes and has given reliable results in this field, thus is being used widely by researchers. So, in this article, a systematic review about the two most used hyperspectral classification techniques that is, SVM and CNN is given. A total of 86 papers from the four well‐known journals belonging to the field of remote sensing have been reviewed where the methodologies used by different authors, datasets used, results acquired, contribution and shortcomings have been put forth. This meta‐analysis generally focuses on the recent SVM and CNN methodologies that are proposed by various authors. A summary of the best results obtained from SVM and CNN in terms of classification accuracies is also provided to help the researchers working in this area achieve better results.
Article
Full-text available
Fluorescence spectroscopy, color imaging and multispectral imaging (MSI) have emerged as effective analytical methods for the non-destructive detection of quality attributes of various white meat products such as fish, shrimp, chicken, duck and goose. Based on machine learning and convolutional neural network, these techniques can not only be used to determine the freshness and category of white meat through imaging and analysis, but can also be used to detect various harmful substances in meat products to prevent stale and spoiled meat from entering the market and causing harm to consumer health and even the ecosystem. The development of quality inspection systems based on such techniques to measure and classify white meat quality parameters will help improve the productivity and economic efficiency of the meat industry, as well as the health of consumers. Herein, a comprehensive review and discussion of the literature on fluorescence spectroscopy, color imaging and MSI is presented. The principles of these three techniques, the quality analysis models selected and the research results of non-destructive determinations of white meat quality over the last decade or so are analyzed and summarized. The review is conducted in this highly practical research field in order to provide information for future research directions. The conclusions detail how these efficient and convenient imaging and analytical techniques can be used for non-destructive quality evaluation of white meat in the laboratory and in industry.
Article
Full-text available
The transparency of nanofibrous scaffolds is of highest interest for potential applications like corneal wound dressings in corneal tissue engineering. In this study, we provide a detailed analysis of light transmission through electrospun polycaprolactone (PCL) scaffolds. PCL scaffolds were produced via electrospinning, with fiber diameters in the range from (35 ± 13) nm to (167 ± 35) nm. Light transmission measurements were conducted using UV–vis spectroscopy in the range of visible light and analyzed with respect to the influence of scaffold thickness, fiber diameter, and surrounding medium. Contour plots were compiled for a straightforward access to light transmission values for arbitrary scaffold thicknesses. Depending on the fiber diameter, transmission values between 15% and 75% were observed for scaffold thicknesses of 10 µm. With a decreasing fiber diameter, light transmission could be improved, as well as with matching refractive indices of fiber material and medium. For corneal tissue engineering, scaffolds should be designed as thin as possible and fabricated from polymers with a matching refractive index to that of the human cornea. Concerning fiber diameter, smaller fiber diameters should be favored for maximizing graft transparency. Finally, a novel, semi-empirical formulation of light transmission through nanofibrous scaffolds is presented.
Article
Full-text available
As part of the sustainable future vision, sustainable agriculture has become an essential pillar of the food security strategies formulated by the Dubai Government. Therefore, the Dubai Emirate began relying on new technology to increase productivity and efficiency. Agriculture applications also depend on accurate land monitoring for timely food security control and support actions. However, traditional monitoring requires field surveys to be performed by experts, which is costly, slow, and rare. Agriculture monitoring systems must be furnished with sustainable land use monitoring solutions, starting with remote sensing using drone surveys for affordable, efficient, and time-sensitive agriculture mapping. Hence, the Dubai Municipality is currently using Unmanned Aerial Vehicles (UAVs) to map the farming areas all over the Emirate, support locating lands conducive to cultivation, and create an accurate agriculture database contributing to the decision-making process in determining areas suitable for crop growth. This study used a novel object detection method coupled with geospatial analysis as an integrated workflow to detect individual crops. The UAV flights were executed using a Trimble UX5 (HP) over twelve communities across the Dubai Emirate for six months. Detection methods were applied to high-resolution drone images, consisting of RGB and near-infrared (NIR) bands. Advanced geoprocessing tools were also used to analyze, evaluate, and enhance the results. The performance of detection of the selected deep learning models are discussed (vegetation cover accuracy = 85.4%, F1-scores for date palms and ghaf trees = 96.03% and 94.54% respectively, with respect to visual interpretation ground truth); moreover, sample images from the datasets are used for demonstrations. The main aim is to offer specialists a solution for measuring and assessing living green vegetation cover derived from the processed images that is integrated. The results provide insight into using UAS and deep learning algorithms as a solution for sustainable agricultural mapping on a large scale.
Article
Vision is essential in the diagnostic capabilities in the specialty of dermatology. However, humans are limited in color vision by the trichromatic visual system that we possess. Multispectral and hyperspectral imaging can overcome this limitation and non‐invasively provide novel information about a skin lesion at the cellular level. A literature review from January 2019 to March 2021 for hyperspectral and multispectral imaging in the field of dermatology was conducted. Multispectral/Hyperspectral imaging continues to generate significant research and interest in dermatology. Much of this research is on distinguishing melanoma from benign nevi as this could allow for a diagnosis without biopsy. In addition, adding multispectral/hyperspectral imaging to smartphones is being researched in order to create a portable and low cost device that can be used in remote areas. One of the limitations in developing devices utilizing hyperspectral imaging has been a sacrifice in specificity in order to maximize sensitivity. Potential solutions to combat this that are being researched include combining multispectral/hyperspectral imaging with other imaging modalities such as photoacoustic imaging in order to overcome the limitations of using each individually. Multispectral/Hyperspectral imaging could be an instrumental aid for clinicians in examining, diagnosing and developing the management plan for patient’s skin lesions.
Article
Accurate and timely classification of diseases during strawberry planting can help growers deal with them in timely manner, thereby reducing losses. However, the classification of strawberry diseases in real planting environments is facing severe challenges, including complex planting environments, multiple disease categories with small differences, and so on. Although recent mobile vision technology based deep learning has achieved some success in overcoming the above problems, a key problem is how to construct a non-destructive, fast and convenient method to improve the efficiency of strawberry disease identification for the multi-region, multi-space and multi-time classification requirements. We develop and evaluate a rapid, low-cost system for classifying diseases in strawberry cultivation. This involves designing an easy-to-use cloud-based strawberry disease identification system, combined with our novel self-supervised multi-network fusion classification model, which consists of a Location network, a Feedback network and a Classification network to identify the categories of common strawberry diseases. With the help of a novel self-supervision mechanism, the model can effectively identify diseased regions of strawberry disease images without the need for annotations such as bounding boxes. Using accuracy, precision, recall and F1 to evaluate the classification effect, the results of the test set are 92.48, 90.68, 86.32 and 88.45%, respectively. Compared with popular Convolutional Neural Networks (CNN) and five other methods, our network achieves better disease classification effect. Currently, the client (mini program) has been released on the WeChat platform. The mini program has perfect classification effect in the actual test, which verifies the feasibility and effectiveness of the system, and can provide a reference for the intelligent research and application of strawberry disease identification.