s42979-024-03380-5
s42979-024-03380-5
s42979-024-03380-5
https://doi.org/10.1007/s42979-024-03380-5
ORIGINAL RESEARCH
Abstract
Multispectral satellite image classification is crucial in remote sensing and image analysis, serving various applications
like water area extraction, urban region classification, and vegetation monitoring. Model-based classification, particularly
supervised learning approaches, heavily relies on manually annotated training samples, which are labor-intensive and subject
to human bias. However, solely depending on features derived from individual pixels may fail to capture spatial contextual
information effectively and manage mixed pixels and various types of noise, resulting in classification inaccuracies. To
address these challenges, this research proposes a novel unsupervised learning approach and automatic labeling technique
(ALT). This approach utilizes collective information from groups of pixels instead of relying solely on individual pixels for
identification and extraction of meaningful unique patterns which serves as the most significant factor for satellite image
classification. A new similarity index, the patch similarity index (PSI), is generated by incorporating the rank correlation
similarity index (RCSI). The PSI is used to determine the similarity factor both between the patches and within the patch.
Subsequently, a clustering algorithm is employed to group pixels into four distinct classes: vegetation, water, built-ups, and
open land. Additionally, ALT streamlines the post-classification process by reducing the need for manual intervention in
mapping clusters to the actual class label. The proposed method offers a more accurate and reliable unsupervised classifica-
tion for multispectral satellite imagery, promising enhanced efficiency and reduced manual labor.
Keywords Rank Correlation · Multispectral · Patch similarity index · Frequent masking · Automatic labeling technique
SN Computer Science
Vol.:(0123456789)
1073 Page 2 of 20 SN Computer Science (2024) 5:1073
SN Computer Science
SN Computer Science (2024) 5:1073 Page 3 of 20 1073
improves the post-classification stage by automatically where CDF is the cumulative distribution function of the
mapping the identified clusters to the corresponding standard normal distribution. |τ | is the absolute value of
labels. This approach streamlines the classification process Kendall’s tau coefficient.
without relying on additional human effort to identify and
map pixels to corresponding classes.
Image Data
Manuscript Structure Overview
The image sets used in this research were collected from
The manuscript is thoughtfully structured into five main sec- the Operational Sentinel-2 Imager on 14th February 2023,
tions, each dedicated to a specific aspect of the research. obtained from the United States Geological Survey (USGS)
Section2 provides a detailed background study and material Earth Resources Observation https://e arthe xplor er.u sgs.g ov/.
used in this research. Section 3 explains the methodology The Sentinel-2B satellite provides both spatial and spec-
employed in our research, offering clarity on the approach tral information. To ensure accurate analysis, images were
taken. Section 4 presents the results of the research and pro- atmospheric corrected and converted to reflectance using
vides a comparative discussion of these findings. Finally, the the SCP plugin of QGIS software, which is commonly used
manuscript concludes with a conclusion section that sum- in atmospheric correction [28]. In this research, six spectral
marizes the main points and conclusions of the research. bands were selected from the Sentinel-2B imagery. These
bands include Band 2 (492.1 nm), Band 3 (559.0 nm), Band
4 (664.9 nm), Band 8 (832.9 nm), Band 11 (1610.4 nm),
Background Study and Band 12 (2186.7 nm). These bands were chosen from
the available 13 spectral bands as they are commonly found
Correlation Coefficient in optical sensors and provide valuable information for our
analysis. To ensure consistent cell resolution across the data,
To measure the association between two measured quanti- the spatial resolution of the shortwave bands (Band 11, and
ties, Kendall’s τ coefficient is commonly used. It suits small Band 12) at 20 m was resampled to 10 m using the Bilinear
sample sizes and data with tied ranks [24]. It considers Resampling technique [29], where weighted average values
pairs of observations and determines the strength of asso- of the four nearest pixels values are used to calculate the
ciation based on the patterns of concordance and (discord- ) new pixel of the output raster. All six selected bands have
ance between the pairs. For a set of observations a1 , b1 , the same spatial resolution but different spectral properties.
( )
…, an , bn of the joint random The maps generated are projected in the World Geodetic
[ variables A and B, the num-
ber of concordant pair Nc = (x, y)|ax < ay and bx < by ] and System 1984 (EPSG:7780).
[
number of discordant pair Nd = (x, y)|ax < ay and bx > by ]
The Kendall τ coefficient is given by Eq. 1
Test Area
N − Nd
𝝉= c (1)
n The test scene for this research was collected from Tumak-
Furthermore, tied pairs occur when ax = ay and bx = by, uru district (TD) and Bengaluru Urban district (BUD) in
and they are neither concordant nor discordant. To handle the state of Karnataka [Fig. 1(a)], India, having a coordinate
tied pairs, Stuart et al. [25, 26] proposed a method by modi- of 13.34°N, 77.1°E [30] and 12.9700°N, 77.6536°E [31]
fying the coefficient as shown in Eq. 2 respectively respectively. Two random regions within these
districts were selected as the test areas, each marked by a
2(Nc − Nd ) red rectangular box. The first test area (TA1) selected from
𝝉= (2)
n2 (m−1) TD, covered an area of 1601 hectares, while the second test
m
area (TA2) from BUD covered an area of 1502 hectares [see
here r and c are the number of rows and columns, respec- Fig. 1(b) and Fig. 1(c)]. These test areas were chosen based
tively, and m = min(r, c). on their complex surface features, encompassing various
To determine the probability p [27], associated with each land cover types such as water, open land, vegetation, and
possible value of 𝜏 is calculated based on an approximation built-up areas. It is worth mentioning that these regions also
using the Gaussian distribution given by Eq. 3. contained certain sources of noise, including high-rise build-
ing shadows, cloud shadows, and organic and inorganic con-
Pvalue = 2 ∗ (1 − CDF(|𝝉|)) (3) stituents in water. The selection of scenes with such noise
elements was intentional to evaluate the robustness of the
proposed methods in handling challenging scenarios.
SN Computer Science
1073 Page 4 of 20 SN Computer Science (2024) 5:1073
Reference data to verify the classification accuracy points, we identified vegetation pixels (V) as label 1, built-
up pixels (BU) as label 2, water pixels (W) as label 3, and
Earlier research studies have revealed the effectiveness of open land pixels (OL) as label 4. Consequently, a refer-
high-resolution images in obtaining the labels of different enced map was generated for TA1 and TA2, corresponding
land covers of satellite imagery [32, 33]. In this research, to their respective spatial extents. The reference image
we employed Google Maps having a resolution of about of TA1 and TA2 is shown in Fig. 2 (a) and Fig. 2 (b)
0.15 m [34] with a similar time frame, to extract the cor- respectively. It is important to note that the labeled maps
responding labels for the test areas. Both TA1 and TA2 obtained are considered ground truth for this research and
were delineated on the Google map. As the target classes are only utilized for validating our classification results.
are visually distinguishable, we used manual digitization The reference map for TA1 consists of a total of 758 poly-
of the four land cover classes referring to high-resolution gons and 150,495 pixels, while for TA2, it comprises 395
Google Earth imagery for generating reference data, for polygons and 21,823 pixels. The distribution of pixels
each of the test sites, corresponding to their respective among the different land cover classes in the reference
spatial extents. To assign labels to the selected sampling data is presented in Table 1.
SN Computer Science
SN Computer Science (2024) 5:1073 Page 5 of 20 1073
Table 1 Number of pixels Class Number of Pixels Table 2 Confusion matrix for four class classification
associated with each class in
reference image of TA1 and TA1 TA2 Predicted Values
TA2 Class 1 Class 2 Class 3 Class 4
Vegetation 63,139 11,853
Built-up 21,657 1992 Actual Class 1 A11 A12 A13 A14
Water 8301 1832 Values Class 2 A21 A22 A23 A24
Open Land 57,398 6146 Class 3 A31 A32 A33 A34
Class 4 A41 A42 A43 A44
Accuracy Assessment
calculations for each class. The TP, FP, FN, and TN values
To assess the accuracy of our proposed method for mapping for the four classes are computed using the confusion matrix,
four classes, we evaluated by counting the number of pixels as illustrated in Table 3. To evaluate the performance of the
that were correctly classified and incorrectly classified. To unsupervised learning-based classification results, several
measure the performance, we utilized two widely used met- other metrics are employed for each class, including overall
rics: overall accuracy (OA) and kappa coefficient (KC) [35]. accuracy, precision, recall, and F1-score, as defined Table 4.
These metrics were calculated based on the confusion matrix Additionally, the output is analyzed using the Kappa coef-
(CM) method, which has been extensively employed in vari- ficient to further assess the classification results.
ous studies [35–38]. The confusion matrix consists of four
classes, namely True Positive (TP), False Negative (FN),
False Positive (FP), and True Negative (TN). The interpreta- Methodology
tion of these classes is as follows:
The methodology proposed in this research paper consists of
• TP: The number of times the classifier correctly predicts three key stages for the classification of multispectral satel-
the positive class. lite images:
• FN: Counts the number of times where the classifier
incorrectly predicts the positive class as negative. Patch Similarity Index (PSI) Generation:
• FP: Counts the number of times the classifier incorrectly
predicts a negative class as positive. • Patch Generation: A set of’N’ group of pixels
• TN: Counts the number of times the classifier correctly referred to as patch is generated from the 3D raster
predicts the negative class. ( Rraster ) representing the multispectral image.
• Rank Correlation Similarity Index (RCSI) Calcula-
The confusion matrix (CM) for mapping the target tion: A RCSI matrix is computed for each patch to
class and reference map in the four-class classification is quantify the correlation between the target pixel and
presented in Table 2. In contrast to binary classification its neighboring pixels.
problems [35], multi-class classification requires separate
SN Computer Science
1073 Page 6 of 20 SN Computer Science (2024) 5:1073
Class 1 A11 A12 + A13 + A14 A21 + A31 + A41 A22 + A23 + A24 + A32 + A33 + A34 + A42 + A43 + A44
Class 2 A22 A21 + A23 + A24 A12 + A32 + A42 A11 + A13 + A14 + A31 + A33 + A34 + A41 + A43 + A44
Class 3 A33 A31 + A32 + A34 A13 + A23 + A43 A11 + A12 + A14 + A21 + A22 + A24 + A41 + A42 + A44
Class 4 A44 A41 + A42 + A43 A14 + A24 + A34 A11 + A12 + A13 + A21 + A22 + A23 + A31 + A32 + A33
Table 4 Metrics
Evaluation Metrics Description Expression
• PSI Generation: The RCSI matrix obtained is used a total of 4500 randomly distributed sample values from each
to generate the PSI matrix for further classification class, based on the reference data. These samples were collected
process. separately from each band of the RSbands. We then calculated
the mean reflectance value for each spectral band from both the
Clustering: A clustering algorithm is employed to group
test scene, corresponding to the target class. To visualize and
the data points based on PSI values. This step aims to
analyze the spectral characteristics, we plotted a spectral curve,
identify distinct clusters representing the four land cover
as shown in Fig. 4. The Y-axis represents the reflectance values
classes.
of the target classes (pixel values), while the X-axis represents
Automatic Labeling: The clusters obtained in the previous
the corresponding spectral bands. The spectral curve highlights
step are assigned to one of the four target classes using an
several key observations-
ALT and Frequent Masking (FM) technique.
• Water pixels primarily absorb energy in the SWIR
The overall architecture of the proposed methodology is
region, but they also exhibit a moderate reflectance in
illustrated in Fig. 3. The input to this approach is a radio-
the GREEN and NIR bands.
metrically corrected multispectral image from the Sentinel
• Vegetation pixels show high reflectance in the NIR wave-
2B satellite. Therefore, no additional radiometric correction
length and moderate reflectance in the GREEN bands,
is applied as part of the methodology.
while they absorb energy in the SWIR wavelengths.
• Built-up features reflect more energy in the visible wave-
Patch Similarity Index (PSI) Generation
length range, whereas open land reflects more in the NIR
and SWIR1 bands. Both built-up and open land pixels
This sub section provides a comprehensive breakdown of
show energy absorption in the SWIR2 wavelength.
the various steps involved in generating the PSI, including
spectral curve generation, deriving indices, spectral analysis,
Deriving indices from RS bands
spectral-spatial feature extraction, and the formulation of the
RCSI and PSI matrices.
Spectral indices have been widely used in previous studies to
investigate various fields such as agriculture, water resources,
Spectral curve generation from the six spectral bands
and urban development [32, 33, 36, 37, 39]. These indices are
computed by combining reflectance values from multiple spec-
Classification of satellite images or separating pixels into clus-
tral bands (RSbands) into a single value. In this research, several
ters commonly relies on the spectral bands available in the satel-
indices were derived based on specific formulas, as presented in
lite imagery. We stack together the six spectral bands to create
Table 5. It is important to note that these indices were selected
a 3D spectral raster called RSbands. This allows us to analyze the
based on their capability to highlight specific land cover classes.
spectral characteristics of the target class pixels. To investigate
All the indices are then stacked together to form a 3D indices
the spectral properties of the target class pixels, we collected
raster called RI bands. Each index is considered as an individual
SN Computer Science
SN Computer Science (2024) 5:1073 Page 7 of 20 1073
band. Following the steps, for selecting target pixel values as dis- Spectral analysis of the RS bands and the derived indices
cussed in section A.a, we plotted another spectral curve shown bands:
in Fig. 5 based on the RI bands. The spectral indices are shown
in the X-axis and the corresponding pixel values for each of Analyzing the spectral curve is a useful approach to gain insights
the target classes are shown in the Y-axis. The spectral curve into the properties of land surface pixels [36, 39]. Different land
provides insights into the behavior of the target classes across cover types exhibit distinct spectral properties. It is observed
the derived indices. In Figs. 6 and 7, we display the output of the that vegetation has high reflectance in the near-infrared (NIR)
derived indices for TA1 and TA2, respectively. The analysis of region and water pixels absorb more energy with low reflectance
the spectral indices reveals interesting patterns. The BSI, NDWI- in NIR and shortwave infrared (SWIR) wavelengths. As each
B, AEWInsh, MNDWI, BU, and NBI indices exhibit positive class label interacts with light energy differently, noticeable vari-
values for water regions and negative values for other classes. ations are observed in the normalized difference trends shown
On the other hand, indices such as BU, NDBI, and NBI indices in Fig. 4 and Fig. 5. Figure 4 illustrates that the spectral curve of
effectively highlight urban or built-up areas. Additionally, the built-up areas and open land share a similar trend in the RSbands.
EVI, NDVI, NDWIns, and NDWI-B indices emphasize vegeta- However, a significant disparity can be observed in the reflec-
tion regions, while the BSI, NDVI, NDBI, and NBI indexes pre- tance curves of indices NDBI, BSI, and NBI, as shown in Fig. 5.
dominantly highlight open land areas, suppressing other indices. Furthermore, the spectral curve of water exhibits low variance
The derived spectral indices provide valuable information for compared to the indices curves, which display high negative and
characterizing different land cover classes and are utilized for positive values for water pixels across different indices. While
further analysis and classification tasks. the vegetation curve in RSbands relies solely on the NIR band for
SN Computer Science
1073 Page 8 of 20 SN Computer Science (2024) 5:1073
SN Computer Science
SN Computer Science (2024) 5:1073 Page 9 of 20 1073
Fig. 6 Spectral indices of TA1: a NDBI-B b NDVI c NDSI d NDBI e NBI f MNDWI g EVI h BU i BSI j AWEInsh k NDWIns
SN Computer Science
1073 Page 10 of 20 SN Computer Science (2024) 5:1073
Fig. 7 Spectral indices of TA2: a NDBI-B b NDVI c NDSI d NDBI e NBI f MNDWI g EVI h BU i BSI j AWEInsh k NDWIns
SN Computer Science
SN Computer Science (2024) 5:1073 Page 11 of 20 1073
correlation within the patch. Secondly, the 𝜏c − statistic of (1 − pvalue ), resulting in the final normalized RCSI matrix for
each cell is divided by the corresponding significance quantity each patch.
Algorithm 1 PatchGeneration ( )
SN Computer Science
1073 Page 12 of 20 SN Computer Science (2024) 5:1073
mean of neighboring pixels alone. Instead, the correlated clusters by minimizing the in-cluster sum of squares. Pixels
influencing factors of the neighboring contribute to the act as with similar properties are clustered together, resulting in a
the centroid. Each center pixel is assigned to one of the four classification map Rm∗n
clustered
containing four distinct clusters.
Algorithm 2 PatchSimilarityIndexGeneration( )
Assigning Labels to Clusters generating Bmask. Threshold values, as suggested in the relevant
literature, are employed to separate target pixels and suppress
Assigning labels to clusters often involves subjective human unwanted pixels for each class label. These threshold values are
observation, leading to potential errors and misleading classifi- presented in Table 6. The mask maps for water, vegetation, open
cations. Traditional methods [6, 57] for creating and validating land, and built-up areas are depicted in Fig. 8 (a–d) for TA1 and
labeled maps are prone to human subjectivity and errors. Gen- 8 (e–h) for TA2, respectively. It is important to emphasize that
erating labeled data for satellite images is particularly challeng- the primary focus is on the major mask, disregarding small clas-
ing due to their complex spectral properties. In this research, we sification errors or misclassified pixels. The concept involves
proposed a method that utilizes spectral indices derived from overlaying individual mask maps of the corresponding target
target image data to generate binary classification maps for spe- pixels onto the clustered image to assign specific target labels to
cific target classes. The MNDWI index [42] is employed to dis- each corresponding clustered pixel. For instance, for water iden-
tinguish water and non-water pixels in the test image using a tification and automatic labeling, a Rm∗n
merged
matrix is generated
threshold, resulting in a binary water mask (Wmask), where water through pixel-wise multiplication of the Rclustered and Wmask. The
m∗n
pixels are assigned a value of 1 and non-water pixels 0. Simi- Rm∗n matrix contains positive pixels likely to be represented
merged
larly, EVI index are used to generate a vegetation mask (Vmask) as water and 0 values for others. But it may have misleading
to separate vegetation and non-vegetation pixels, with values of
Table 6 Threshold value for indices
1 and 0 assigned accordingly. To extract the open land pixels and
suppress non open land pixels in the test image to generate Class Threshold Label 1 Label 0
Omask , the BSI index is utilized, assigning values of 1 and 0,
Water [40] NDWI > 0 Water pixel Other pixels
respectively. Although the NBI index is used to extract built-up
Vegetation [45, NDVI > 0.3 Vegetation pixel Other pixels
regions, it is also associated with open land pixels in the classi- 46]
fication. To address this, we identify positive pixels resulting Open land [47] BSI > 0.2 Open land pixel Other pixels
from the difference between the NBI and BSI, assigning them a Built- ups [44] [(NBI > 0.1) Built-up pixel Other pixels
value of 1 for built-up pixels while assigning 0 to other pixels -(BSI > 0.2)] > 0
SN Computer Science
SN Computer Science (2024) 5:1073 Page 13 of 20 1073
Fig. 8 TA1 [a Water mask, b Vegetation mask, c Open Land mask, d Built-up mask] TA2 [e Water mask, f Vegetation mask, g Open Land
mask, h Built-up mask]
pixels. So, to handle this a new masking technique known as label, and Bmask , for the built-up labels. The pseudocode for
Frequent Masking (FM) method is proposed. In this method the assigning labels to clusters can be found in Algorithm 3. It is
frequency of all unique non-zero pixels in Rm∗n merged
is calculated. worth noting that there can be multiple cluster labels with the
The cluster label (number) associated with the highest frequency same frequency can be found in Rm∗n merged
, indicating ambiguity in
is identified. Then all the pixel associated with the correspond- the assigned labels, in that case, the clustering algorithm is re-
ing cluster number in Rm∗nclustered
is identified and marked as water evaluated to find corresponding clusters, till all the labels are
labels. The same procedure is applied using Vmask, to identify the identified and properly labeled.
vegetation cluster label, Omask, to detect the open land cluster
Algorithm 3 AutomaticLabelingTechnique( )
SN Computer Science
1073 Page 14 of 20 SN Computer Science (2024) 5:1073
Kappa
ficient
Coef-
0.762
0.531
0.468
Comparative analysis of different classification
methods with the proposed method
Overall Accuracy
To conduct a comprehensive comparative analysis, we evaluated
several state-of-the-art classification approaches, focusing on
their effectiveness in classifying land cover types. A commonly
0.856
0.655
employed method is pixel-based classification, widely used in
0.6
numerous studies [4, 5, 58–61]. In this method, raw pixel values
are directly fed into the clustering algorithm to classify the four
F1-Score
land cover classes. This is referred to as Pixel-Based Classifica-
0.705
0.963
0.761
0.637
0.275
0.964
0.84
0.89
0.97
0.82
0.31
0.72
tion (PiBC) in this research. Another approach involves group-
ing pixels into chunks, rather than treating them individually,
to classify different land cover types [62, 63]. It is important
0.7662
Recall
0.563
0.967
0.664
0.486
0.741
0.971
0.613
to note that this approach relies solely on multispectral bands
0.83
0.84
0.97
0.84
from satellite imagery and does not incorporate derived spectral
indices. Additionally, these chunks use the reflectance values of
Precision
the pixels within the group for classification, without consider-
0.2001
0.944
0.959
0.891
0.169
0.957
0.894
ing the influence of neighboring pixels on the target pixels. This
TA2
0.85
0.94
0.97
0.92
0.8
method is distinctly different from the proposed approach and
is referred to as Simple Chunk-Based Classification (CuBC ).
Coefficient
Some researchers have integrated dimensionality reduction
techniques into the CuBC approach to improve classification
Kappa
0.609
0.557
0.64
accuracy [64–66]. This variant is termed Chunk-Based Clas-
sification with PCA (CuBCpca) in this research. The above
approaches were evaluated on two feature sets: RSbands (multi-
Accuracy
0.734
0.698
0.76
the latter representing derived spectral indices) across two test
scenes, TA1 and TA2. The quantitative classification results
for RSbands are presented in Table 7, while those for Rraster are
Table 7 Classification accuracies of different approaches using RSbands feature set
F1-Score
0.75
0.76
0.95
0.76
0.67
0.97
0.88
0.62
0.39
0.97
0.89
0.71
0.97
0.77
0.74
0.38
0.98
0.94
0.71
0.29
0.96
0.93
OL
OL
OL
W
W
W
V
V
V
ally, there was relatively low recall for "Open Land," indicat-
CuBC
PiBC
SN Computer Science
SN Computer Science (2024) 5:1073 Page 15 of 20 1073
Rraster PiBC V 0.94 0.83 0.88 0.707 0.572 0.9 0.65 0.75 0.632 0.504
W 0.98 0.97 0.97 0.96 0.97 0.96
BU 0.34 0.69 0.46 0.18 0.74 0.29
OL 0.71 0.57 0.63 0.93 0.53 0.67
CuBC V 0.93 0.86 0.9 0.722 0.591 0.89 0.66 0.76 0.637 0.51
W 0.96 0.98 0.97 0.96 0.97 0.96
BU 0.34 0.62 0.44 0.19 0.77 0.31
OL 0.73 0.59 0.66 0.94 0.52 0.67
CuBCpca V 0.77 0.82 0.8 0.801 0.695 0.8 0.97 0.87 0.88 0.816
W 0.98 0.98 0.98 0.95 0.98 0.97
BU 0.79 0.87 0.83 0.93 0.84 0.89
OL 0.81 0.74 0.77 0.96 0.8 0.87
0.5575 ( RSbands, TA1), 0.5723 ( Rraster , TA1), 0.468 ( RSbands, Kappa Coefficient also improved, reinforcing the effectiveness
TA2), and 0.504 (Rraster, TA2), indicated moderate agreement of this approach. In contrast, the proposed method, depicted in
between the predicted and true labels. The classification map Figs. 9 (g) and 10(g) for TA1 and TA2 respectively, outper-
for PiBC method is shown in Figs. 9 (a) and (d) and Fig. 10(a) formed all other approaches. It achieved high precision, recall,
and (d) and for CuBC method is depicted in Figs. 9 (b) and and F1-scores across all four classes. The “Water” and “Vegeta-
(e) and Fig. 10(b) and (e) for TA1 and TA2 respectively. The tion” classes were classified with high accuracy, and the method
CuBC method demonstrated improved performance over the significantly improved the classification of the “Built-up” and
pixel-based approach for both RSbands and Rraster feature sets. It “Open Land” classes, which had underperformed in previ-
achieved higher precision, recall, and F1-scores for the “Vegeta- ous approaches. The overall accuracy of the proposed method
tion” and “Water” classes, indicating better classification accu- was the highest, 90.30% (TA1) and 91.79% (TA2), indicating
racy. However, the performance for the “Built-up” and “Open a significant enhancement in classification performance. The
Land” classes remained relatively low. The overall accuracy Cohen’s Kappa Coefficient values were the highest for proposed
increased to 73.44% ( RSbands, TA1), 72.29% ( Rraster , TA1), method, suggesting substantial agreement between the predicted
65.57% (RSbands, TA2), and 63.70% (Rraster , TA2), indicating and actual classifications. The results demonstrate that the pro-
enhanced classification performance. The Cohen’s Kappa Coef- posed method effectively captures the characteristics of the tar-
ficient also improved, with values of 0.6098 ( RSbands, TA1), get classes and significantly improves classification accuracy
0.5917 ( Rraster , TA1), 0.5315 ( RSbands , TA2), and 0.5108 compared to the other approaches.
( Rraster , TA2), suggesting a stronger agreement between pre-
dicted and true labels. However, the lower accuracy of the Analyzing the Effectiveness of Proposed Approach
CuBC method when using Rraster compared to RSbands suggests
that the increase in dimensionality due to the addition of indi- To illustrate the importance and effectiveness of the
ces negatively impacts the clustering algorithm, as k-means proposed approach, we conducted ablation research by
does not perform well in high-dimensional spaces. Therefore, progressively modifying the proposed architecture and
dimensionality reduction techniques like PCA are essential for excluding key components to evaluate their impact on
further improving performance. The CuBCpca method, shown classification performance. The results of this research
in Figs. 9 (c) and (f) and Fig. 10(c) and (f) for TA1 and TA2 are summarized in Table 10, where the performance is
respectively, achieved consistent precision, recall, and F1-scores measured by OA and KC for two test areas, TA1 and TA2.
for the “Vegetation” and “Water” classes. Although the “Built- This analysis provides insights into the role of each com-
up” and “Open Land” classes showed slight improvements, ponent– RSbands (multispectral bands), RI bands (the spec-
they approached near-complete classification. The overall tral indices), and PCA in the proposed approach. In the
accuracy increased to 76% ( RSbands, TA1), 80.17% ( Rraster , first approach, we excluded the RI bands and only retained
TA1), 85.69% (RSbands, TA2), and 88.05% (Rraster, TA2), dem- the RSbands . This configuration achieved an OA of 0.834
onstrating enhanced classification performance. The Cohen’s
SN Computer Science
1073 Page 16 of 20 SN Computer Science (2024) 5:1073
Kappa Coefficient
includes both RI bands and RSbands . These results indicate
that although using RSbands with dimensionality reduc-
tion improves accuracy, the absence of spectral indices
0.871
restricts the approach from achieving optimal perfor-
mance. Therefore, the combination of multispectral bands
Overall Accuracy and spectral indices is essential for optimal classifica-
tion performance. In the second approach, both PCA and
RI bands were excluded, leaving only the RSbands . This led to
a significant drop in performance, with an OA of 66.16%
0.917
0.97
0.99
0.87
0.88
0.93
0.99
0.89
0.88
proposed architecture.
Class Label
BU
OL
W
V
Proposed Method
Classification
SN Computer Science
SN Computer Science (2024) 5:1073 Page 17 of 20 1073
Fig. 9 TA1 classification for 〖RS〗_bands [a PiBC b CuBC c CuBCpca], TA1 classification for R_raster [d PiBC e CuBC f CuBCpca] g Pro-
posed method
Fig. 10 TA2 classification for 〖RS〗_bands [a PiBC b CuBC c CuBCpca] TA2 classification for R_raster [d PiBC e CuBC f CuBCpca] g Pro-
posed method
TA1 0.834 0.738 0.6616 0.5070 0.6464 0.4876 0.8421 0.7225 0.7968 0.6955
TA2 0.855 0.762 0.6183 0.4855 0.6312 0.5016 0.8392 0.7396 0.7993 0.6977
SN Computer Science
1073 Page 18 of 20 SN Computer Science (2024) 5:1073
SN Computer Science
SN Computer Science (2024) 5:1073 Page 19 of 20 1073
22. S. Xiaoxia, Z. Jixian, and Z. Liu. 2005. A comparison of object- sentinel-2a msi imageries a case of addis ababa city, ethiopia.
oriented and pixel-based classification approachs using quick- Environ Chal. 2022;8: 100568.
bird imagery. 01 2005 45. N.E.Observatory. 2023 Measuringvegetation(ndvi&evi),” \
23. M. Oyekola and K. Adewuyi. 2018. Unsupervised classification {https://earthobservatory.nasa.gov/features/Measur ing Vegeta-
in land cover types using remote sensing and gis techniques. tion}, 28 may 2023.
24. Kendall MG. A new measure of rank correlation. Biometrika. 46. USGS, “Ndvi foundation - remote sensing phenology,” \{https://
1938;30(1/2):81–93. www.usgs.gov/special-topics/remote-sensing-phenology/science/
25. Berry KJ, Johnston JE, Zahran S, Mielke PW. Stuart’s tau meas- ndvi-foundation-remote-sensing-phenology}, 28 may 2023.
ure of effect size for ordinal variables: some methodological 47. Mzid N, Pignatti S, Huang W, Casa, and Raffaele. An analysis
considerations. Behav Res Methods. 2009;41:1144–8. of bare soil occurrence in arable croplands for remote sensing
26. Kendall MG. The treatment of ties in ranking problems. Biom- topsoil applications. Remt Sens. 2021. https://doi.org/10.3390/
etrika. 1945;33(3):239–51. rs13030474.
27. H. Abdi. 2006. The kendall rank correlation coefficient herv 48. G. Liu, K. J. Shih, T.-C. Wang, F. A. Reda, K. Sapra, Z. Yu, A.
abdi Tao, and B. Catanzaro. 2018. Partial convolution based padding.
28. Tempa K, Aryal K. Semi-automatic classification for rapid arXiv preprint arXiv:1811.11718.
delineation of the geohazard-prone areas musing sentinel-2 49. K. K. Perikamana, K. Balakrishnan, and P. Tripathy. 2021. A cnn
satellite imagery. SN App Sci. 2022. https://doi.org/10.1007/ based method for sub-pixel urban land cover classification using
s42452-022-05028-6. landsat-5 tm and resourcesat-1 liss-iv imagery,” arXiv preprint
29. M. R. Singh and A. Bhide. 2016. A review of image retrieval arXiv:2112.08841.
using different types of interpolation techniques. International 50. Park JH, Inamori T, Hamaguchi R, Otsuki K, Kim JE, Yamaoka
Research Journal of Engineering and Technology (IRJET), K. Rgb image prioritization using convolutional neural net-
2016. work on a microprocessor for nanosatellites. Remote Sensing.
30. “Tumakuru, karnataka, india - latitude and longitude,” \{https:// 2020;12(23):3941.
www.latlong.net/place/tumakuru-karnataka-india-11565.html}, 51. O. Thunder, “Reflect padding,” {https://m edium.c om/O rca_T
hund
accessed: [28 may 2023]. er/image-padding-techniques-reflect-padding-part-2-5a013cd965
31. “Bengaluru, karnataka, india - latitude and longitude,” \{https:// 37}, 2024, accessed on 06 june.
www.latlong.net/place/bengaluru-karnat aka-india-499.html}, 52. [S. Basu, S. Ganguly, S. Mukhopadhyay, R. DiBiano, M. Karki,
accessed: [28 may 2023]. and R. Nemani. 2015. Deepsat: A learning framework for satellite
32. Feyisa GL, Meilby H, Fensholt R, Proud SR. Automated water imagery. In Proceedings of the 23rd SIGSPATIAL International
extraction index: A new technique for surface water mapping Conference on Advances in Geographic Information Systems.New
using landsat imagery. Remote Sens Environ. 2014;140:23–35. York, NY, USA: Association for Computing Machinery
33. Yan D, Huang C, Ma N, Zhang Y. Improved landsat-based water 53. H. Ferdous, T. Siraj, S. J. Setu, M. M. Anwar, and M. A. Rahman.
and snow indices for extracting lake and snow cover/glacier in the 2021. Machine learning approach towards satellite image classi-
tibetan plateau. Water. 2020;12(5):1339. fication. In Proceedings of International Conference on Trends in
34. Wikipedia, “Google earth,” Accessed 6 january 2024, wikipedia, Computational and Cognitive Engineering. Springer, Singapore
the free encyclopedia. [Online]. https://en.wikipedia.org/wiki/ 54. Ahmed M, Seraj R, Islam SMS. The k-means algorithm: a com-
Google_Earth prehensive survey and performance evaluation. Electronics. 2020.
35. Khalid HW, Khalil RMZ, Qureshi MA. Evaluating spectral indices https://doi.org/10.3390/electronics9081295.
for water bodies extraction in western Tibetan plateau. Egyptian J 55. Napoleon D, Pavalakodi S. A new method for dimensionality
Remote Sens Space Sci. 2021;24(3):619–34. reduction using k-means clustering algorithm for high dimen-
36. Yang J, Du X. An enhanced water index in extracting water bodies sional data set. Intern J Com Applicat. 2011;13(7):41–6.
from landsat tm imagery. Ann GIS. 2017;23(3):141–8. 56. Estornell J, Martí-Gavilá JM, Sebastiá MT, Mengual J. Principal
37. Mondejar JP, Tongco AF. Near infrared band of landsat 8 as water component analysis applied to remote sensing. Model Sci Educat
index: a case study around cordova and lapu-lapu city, cebu, phil- Learn. 2013;6:83–9.
ippines. Sustainable Environment Research. 2019;29:1–15. 57. Abid N, Shahzad M, Malik MI, Schwanecke U, Ulges A, Kovács
38. Wei Q, Jingxuan L, Lin L, Xiao-Wen L. Research on automatic G, Shafait F. Ucl: Unsupervised curriculum learning for water
extraction of water bodies and wetlands on hj satellite ccd images. body classification from remote sensing imagery. Int J Appl Earth
Remote Sens Inform. 2011;4(2011):28–33. Obs Geoinf. 2021;105: 102568.
39. Hidayati IN, Suharyadi R. A comparative study of various indices 58. Li D, Wang S. Cost-effective land cover classification for remote
for extraction urban impervious surface of landsat 8 oli. Forum sensing images. J Cloud Comp. 2022. https://doi.org/10.1186/
Geografi. 2019;33(2):162–72. s13677-022-00335-0.
40. EOS, “Ndwi analysis,” \{https://e os.c om/m ake-a n-a nalys is/ 59. Adejoke AO, Badaru YU. Accuracy assessment of pixel-based
ndwi/}, 28 may 2023. image classification of kwali council area, abuja, nigeria. J Nat
41. Chen G, Jin R, Ye Z, Li Q, Gu J, Luo M, Luo Y, Christakos G, Sci Res. 2014;4:133–40.
Morris J, He J, et al. Spatiotemporal Mapping of Salt Marshes in 60. Sibaruddin HI, Shafri H, Pradhan B, Haron N. Comparison of
the Intertidal Zone of China during 1985–2019. J Remote Sens. pixel-based and object-based image classification techniques in
2022;2022:9793626. extracting information from uav imagery data. IOP Conf Ser Earth
42. Xu H. A study on information extraction of water body with the Env Sci. 2018. https://doi.org/10.1088/1755-1315/169/1/012098.
modified normalized difference water index (mndwi). J Remote 61. Lilay M, Taye G. Semantic segmentation model for land cover
Sen Beij. 2005;9(5):595. classification from satellite images in gambella national park,
43. Lu L, Guo H, Wang C, Pesaresi M, Ehrlich D. Monitoring bidec- ethiopia. SN Appl Sci. 2023;5:76.
adal development of urban agglomeration with remote sensing 62. Hu F, Xia G-S, Wang Z, Zhang L, Sun H. Unsupervised feature
images in the Jing-Jin-Tang area, China. J Appl Remote Sens. coding on local patch manifold for satellite image scene classifica-
2014;8(1):084592. tion. IEEE Geosci Remote Sens Sympos. 2014;2014:1273–6.
44. Kebede TA, Hailu BT, Suryabhagavan KV. Evaluation of spec-
tral built-up indices for impervious surface extraction using
SN Computer Science
1073 Page 20 of 20 SN Computer Science (2024) 5:1073
63. Cheriyadat AM. Unsupervised feature learning for aer- based lenet-5 architecture. Infrared Phys Tech. 2022. https://doi.
ial scene classification. IEEE Trans Geosci Remote Sens. org/10.1016/j.infrared.2022.104470.
2014;52(1):439–51.
64. Ma Z, Liu Z, Zhao Y, Zhang L, Liu D, Ren T, Zhang XF, Li S. Publisher's Note Springer Nature remains neutral with regard to
An unsupervised crop classification method based on principal jurisdictional claims in published maps and institutional affiliations.
components isometric binning. Intern J Geo Inform. 2020. https://
doi.org/10.3390/ijgi9110648. Springer Nature or its licensor (e.g. a society or other partner) holds
65. Yan S, Zhao JF, Zhao JL, Li QZ. A method for image classifica- exclusive rights to this article under a publishing agreement with the
tion based on kernel pca. Intern Confer Machine Learn Cyber. author(s) or other rightsholder(s); author self-archiving of the accepted
2008. https://doi.org/10.1109/icmlc.2008.4620498. manuscript version of this article is solely governed by the terms of
66. Fırat H, Asker ME, Bayindir M, Hanbay D. Spatial-spectral clas- such publishing agreement and applicable law.
sification of hyperspectral remote sensing images using 3d cnn
SN Computer Science