932
Views
0
CrossRef citations to date
0
Altmetric
Review Article

Optimizing multi-classifier fusion for seabed sediment classification using machine learning

, ORCID Icon, , , , & show all
Article: 2295988 | Received 12 Jun 2023, Accepted 12 Dec 2023, Published online: 21 Dec 2023

ABSTRACT

Seabed sediment mapping with acoustical data and ground-truth samples is a growing field in marine science. In recent years, multi-classifier ensemble models have gained prominence for classification problems by combining several base classifiers. However, traditional ensemble methods do not consider the confidence scores of base classifiers, leading to suboptimal fusion when there are conflicting predictions. The current study introduces a novel optimization strategy that enhances the ensemble’s accuracy by constructing an ideal ensemble predicted probability matrix based on the fusion of predicted probabilities of the base classifiers, to improve seabed sediment mapping. The proposed approach not only addresses the limitations of traditional ensemble methods but also significantly increases the ensemble’s performance. The proposed approach demonstrates significant accuracy improvements. On the under-sampled dataset, it achieves 73.5% improvement compared to individual classifiers (random forest, decision tree, support vector machine), surpassing their respective accuracies. On the standard dataset, the ensemble model attains an accuracy of 79.1%, surpassing individual classifiers. Employing over-sampling techniques further elevates accuracy to 94.9%, exceeding the individual classifier performances. The proposed method is evaluated on acoustical data obtained from the Irish Sea. The proposed method outperforms base classifiers in terms of accuracy, F1 score, and the Kappa coefficient.

1. Introduction

The ocean floor, covering 71% of Earth's surface, harbours useful resources, delivers essential ecosystem services, and offers diverse habitats for numerous species (Hosack et al. Citation2006; McGonigle et al. Citation2011). Seabed maps are crucial for ocean monitoring and management, primarily providing information on seafloor topography and habitats. Accurate habitat mapping aids in determining and measuring the human impact on the seafloor by offering data on the physical features, spatial distribution, and ecological roles of biological communities and habitats (Kostylev et al. Citation2001). Seabed sediments principally consist of the seabed's surface components, which include rocks and surface sediments (Damveld et al. Citation2019; Gary Greene, Cacchione, and Hampton Citation2017; Zajac Citation2008). Comprehensive and high-quality investigations of seabed sediments can enhance marine scientific research, resource development, engineering projects, environmental conservation, and military security across various fields (Barbagelata et al. Citation1991; Diesing et al. Citation2020; A. Z. Leon et al. Citation2020). For successful benthic environment evaluation, obtaining accurate information on seabed sediment distribution through accurate seabed maps is crucial. However, precise and appropriate maps and spatial models that accurately represent the seabed substrate types, habitats, and properties, which are useful for research, resource management, conservation, and spatial planning, are limited in the literature (Ierodiaconou et al. Citation2011). As a result, seabed sediment classification has progressively garnered increased interest from experts and researchers (McGonigle and Collier Citation2014).

Machine learning (ML) has been used for seabed sediment classification using multi-beam echosounder data has yielded remarkable results. ML is often adopted to perform both supervised and unsupervised classification of the seabed (Frederick, Villar, and Michalopoulou Citation2020; Viala, Lamouret, and Abadie Citation2021; Wan et al. Citation2022; Zhang et al. Citation2022). Recent studies that have used ML to study the seabed include Liu et al. (Citation2019), in which the sediment types of the Caminada dredge pit in the eastern portion of the submarine sandy Ship Shoal of the Louisiana inner shelf of the United States (USA) were identified using multiple machine learning classifiers and the performance of multiple supervised classification methods was analysed. Zhao et al. (Citation2020) suggested and verified a hybrid machine learning framework for predicting water levels in five representative high-sediment load reaches of China's Lower Yellow River (LYR). Wang et al. (Citation2021) used the XBoost algorithm to study offshore seabed sediment classification based on particle size parameters. In their work, they concluded that the developed model offered an efficient method for offshore sediment classification using Folk's scheme and grain-size parameters and could serve as an additional approach for studying sedimentary environments. To accomplish precise prediction and mapping of geographic seabed sediment information, Cui et al. (Citation2021) created a deep learning model based on feature optimization. By integrating ten major stream sediment geochemical datasets, principal component analysis (PCA) and self-organizing maps (SOMs) were used to delineate major lithologies in the Jining area, Inner Mongolia, China (G. Wu et al. Citation2021). ML algorithms that have been utilized in the field of acoustic sediment classification include random forest (RF), support vector machine (SVM), k-nearest neighbour (KNN) and artificial neural networks (ANNs) (Bressan et al. Citation2020; He et al. Citation2022; Lüdtke et al. Citation2012; McLaren, McIntyre, and Prospere Citation2019; Mudiyanselage et al. Citation2022; Prasad et al. Citation2020). Despite their strong performance and successes in classification tasks, ML algorithms require the support of other approaches to effectively and fully execute their tasks (Morgan and Jacobs Citation2020). An example is the normalization of data before classification (Singh and Singh Citation2020) when using ML algorithms such as SVM and KNN, since SVM and KNN algorithms are sensitive to large variations in data points. Other approaches that help solve the ML algorithm problem also occur after classification has been performed (Diesing and Stephens Citation2015; Gudiyangada Nachappa et al. Citation2020). In particular, post classification approaches are designed to help improve overall classification performance and usually eliminate errors that occur within individual classifiers by aggregating their decisions.

Multi-classifier ensemble learning is one such approach, in which the decision scores of many classifiers are combined to predict the final class label of an input sample. A multi-classifier model is designed to capture the key properties of all of its constituent models, outperforming individual base classifiers. Such models are resilient because ensembling reduces the dispersion or spread of the base models’ predictions. The multi-classifier model reduces the variation in the prediction errors of the basis classifiers by adding some bias to the competing base learners (Divyabarathi et al. Citation2022; Galvez et al. Citation2022; Kim et al. Citation2022).

Recent works have used rule-based majority voting aggregators to fuse multiple ML algorithms to classify seabed sediments. Yang et al. (Citation2022) performed marine oil spill detection using deep learning models. Wan et al. (Citation2022) also classified seabed sediment using a decision fusion method. The rule-based voting method has shown great results in improving classification. However, in some cases there are conflicting predictions among several classifiers, making it difficult to provide reliable results (F. Leon, Floria, and Badica Citation2017). Such cases occur when two or more classifiers predict the wrong class, and the aggregator simply uses that class, regardless of whether it is the wrong prediction. Furthermore, the classifiers utilize the predictions directly without the confidence score of the base classifiers. Finally, the quality of the fusion results is affected by the number of target classes (Tulyakov et al. Citation2008).

To address the aforementioned challenges in previous studies, the current study proposes a multi-classifier approach with ML algorithms to effectively classify seabed sediments. The proposed approach utilizes the confidence score of the base classifiers to make predictions. According to (Carpenter et al. Citation1999; Foody, Boyd, and Sanchez-Hernandez Citation2007), confidence scores are important in multi-classifiers due to the following: (1) agreement between classifiers and (2) correlations between the predicted accuracies. To implement this approach, the fusion process is formulated as an optimization problem where the accuracy of the fused classifier is maximized. The objective function is based on Bayesian theory (Bernardo and Smith Citation2008), where weights are assigned to each base model to yield an optimal posterior matrix with higher accuracy than the base classifiers. The question lies in which optimal weight can yield the best ensemble probability matrix with the highest accuracy for the fused model. Based on the formulated objective function, the differential evolution (DE) algorithm is used to optimize the whole process (Storn Citation1996). The algorithm explores the search space efficiently through mutation, crossover and selection operators to guide the population towards the global optimum. Notably, DE boasts several advantages, including its suitability for optimizing non-linear, non-convex, and non-differentiable objective functions, making it well-suited for the accuracy-based optimization (J. Parihar and Malik Citation2022; Priyavada and Kumar Citation2023; Zhao et al. Citation2022). Furthermore, DE’s population-based approach proves beneficial when dealing with problems involving multiple variables, such as optimizing the weights for three classifiers. Additionally, DE has relatively few control parameters, is straightforward to implement, and has stronger global optimization ability than other evolutionary algorithms. These characteristics make DE a good choice for maximizing the objective function representing multi-classifier accuracy by finding the optimal weight for fusing the base classifier confidence scores. The proposed approach is trained on acoustical data acquired from the Irish Sea. Additionally, multiple tests are performed to show the robustness and efficacy of the proposed approach.

The contribution of this paper lies in the development of an efficient and optimal multi-classifier ensemble method that fuses the base classifiers based on their confidence scores, as opposed to conventional methods that directly fuse base classifiers based on the predicted score values.

2. Materials and methods

2.1. Data

The Irish Sea, located in the middle of northwest Europe’s continental shelf, borders the North Atlantic to the north and south (see (a)). The depth of the water in this region varies, with coastal regions measuring less than 20 m deep and interior sections reaching more than 100 m deep. The western portion of the Irish Sea is deeper (>100 m), with weaker tidal energy and greater saline levels. The Irish Sea’s benthic habitat is influenced by a wide range of factors, resulting in a variety of complex, diverse, and dynamic physical conditions and biological communities, ranging from stony reefs to deep mud basins. The seafloor has a variety of landforms and underwater deposits. Acoustic sediment classification benefits greatly from studying materials such as silty clay, clayey silt, argillaceous sand, sandy mud, silt, gravel, mudstone, bedrock, and coarse sand, providing valuable insights into acoustic sediment categorization. The flow of water in the Irish Sea is mostly northwards, with a flow rate of approximately 8km3perday. However, there is some unpredictability in the northwards exchange of seawater. The Irish Sea's water is thoroughly mixed, resulting in a vertically homogeneous water column throughout the year.

Figure 1. The study area and the acquired acoustical data. (a) Location of the study area (Irish Sea). (b) Bathymetric data obtained from the acoustic data. (c) Backscatter data, representing the seabed texture, and the locations of grab samples colour-coded by Folk class (see ).

Figure 1. The study area and the acquired acoustical data. (a) Location of the study area (Irish Sea). (b) Bathymetric data obtained from the acoustic data. (c) Backscatter data, representing the seabed texture, and the locations of grab samples colour-coded by Folk class (see Table 1).

To better understand the geological and geographical characteristics of the Irish Sea, acoustical data were obtained between 9 February 2012 and 21 February 2012, and archived with the Marine Environmental Data and Information Network (MENIM), which is a part of the British Geological Survey (BGS) data archive centre. The acoustical data are presented in (a), with (b) showing the bathymetric data and (c) representing the backscatter data.

To support the classification of acoustical sediment, samples were collected using a grab method from various locations, as displayed in (c), and subsequently analysed based on Folk's classification scheme. provides a summary of the number and types of ground-truth samples obtained. Notably, a total of 34 grab samples were secured, with the most commonly occurring classes being muddy sand (nine samples), slightly gravelly muddy sand (five samples), and gravelly mud (five samples). The spatial distribution of these samples, as depicted in (c), ensures comprehensive coverage of the range of acoustic conditions and seabed sediment types prevalent in the study area. These ground-truth sediment samples are of utmost significance because they provide crucial training data for machine learning classifiers to establish associations between acoustics and sediment properties.

Table 1. Grabbed samples of the study area based on Folk’s classification.

2.2. Proposed approach

This paper presents a novel approach for tackling the challenges of classifying seafloor sediments discussed in the introduction. The proposed methodology employs a multi-classifier strategy that utilizes a combination of machine learning algorithms to accurately classify various seabed sediment types. The approach effectively blends the confidence scores of the base classifiers to enhance the accuracy of the multi-classifier model. represents the framework for the proposed approach. The proposed approach is divided into three stages: feature extraction and object-based segmentation of acoustical data, training of base classifiers, and classifier fusion optimization. Initially, significant acoustical features are retrieved from the multibeam data, and training examples are generated using object-based image analysis. The aforementioned examples are then used to train the three base classifiers: RF, SVM, and DT. Finally, by maximizing an objective function reflecting the multi-classifier’s accuracy, a differential evolution optimization technique is used to discover the ideal fusion of the base classifiers.

Figure 2. Schematic of the overall multi-classifier framework for this study. The three main parts include data preparation (feature extraction and object-based segmentation), classification, and optimization.

Figure 2. Schematic of the overall multi-classifier framework for this study. The three main parts include data preparation (feature extraction and object-based segmentation), classification, and optimization.

2.2.1. Feature extraction

To accurately identify the type of sediment present, feature extraction is crucial in establishing a stable mapping relationship with the raw acoustic data. The multibeam backscatter intensity feature is the most commonly used for feature extraction in studies on habitat mapping, as it provides the foundation for correct sediment identification. Furthermore, bathymetry and bathymetry-derived variables offer an intuitive representation of seafloor topography. The bathymetric features extracted for training in this paper encompass properties such as aspect, general curvature, geomorphons, landforms, morphometric features, the multiscale topographic position index, the multiresolution index of valley bottom flatness, plan curvature, profile curvature, slope, the terrain ruggedness index, the topographic position index, variance, and the vector ruggedness measure. Given that the topography of the seafloor is highly correlated with the distribution of different seabed sediment types, it has been demonstrated that the combination of backscatter intensity features and topographic features is more effective in the classification of seabed sediment (Marsh and Brown Citation2009). The study combines both bathymetry and backscattering to extract features for sediment classification, with extracted backscatter features including the angular second moment, contrast, energy, entropy, GLCM contrast, GLCM entropy, and inverse difference moment for sediment classification. The resulting extracted acoustic features for this study, after eliminating features with excessive noise and low separability between different classes, are summarized in .

Table 2. The extracted features from the acoustical data consist of 21 features, including both textures and topographic features.

2.2.2. Object-based image analysis.

The first stage in object-based image analysis (OBIA) is segmentation, a semi-automated procedure that separates image data into individual segments. Segmentation that successfully delimits the characteristics of interest is crucial to the success of object-based image analysis. OBIA operates on the basis of the idea that images should not be analysed based on individual pixels but rather on image segments, where an image segment is a collection of pixels. Furthermore, the image’s segments are hierarchically structured. The segments of an image are identical to the pixels at the pixel level. A finer level of visual objects is produced using segmentation, beginning at the pixel level. Image segments can be segmented further into intermediate and coarse levels to provide even more granular analysis. There is a plethora of data that can be derived from these image segments to form their characteristics. Afterwards, during the classification process, derived data are used. Statistics regarding the geometry, texture, or average colour of any of an object's layers constitute image segment characteristics.

The present study employs a hierarchical process for segmenting the study area into homogenous sections. The OBIA process is utilized to assist in the selection of training samples for the ML classifiers based on the location of the grabbed sample data. An assumption is made that generated training points located within the identical segment as a grabbed sample data point represent sediments of the same type. The segmentation process utilizes the multi-resolution segmentation (MRS) algorithm to group pixels into segments based on a predefined homogeneity value, known as the scale parameter. In addition, the shape parameter controls the influence of segment shape on the segmentation, while the compactness factor regulates the smoothness of the image object outline produced by the segmentation. Accurate segmentation provides an ideal outline of the targets of interest. However, often, many features of heterogeneous classes occur within the same segment, or one feature is demarcated by multiple segments. The spectral difference algorithm is employed to merge neighbouring segments where the difference between mean layer intensities is below a user-defined value. The pixel-based object resizing algorithm can also contract or expand image object boundaries until a specific geometry is achieved or a particular image layer value is met.

The following parameters are utilized for multi-resolution segmentation: The segmentation contains all 21 characteristics from , with a scale of 250, shape set to 0.5, and compactness set to 0.5. The resulting segmentation is shown in .

Figure 3. OBIA segmentation with sediment sample location. Based on this segmentation, training and validation samples are generated for the ML model for learning.

Figure 3. OBIA segmentation with sediment sample location. Based on this segmentation, training and validation samples are generated for the ML model for learning.

2.2.3. Base classifiers

The proposed seabed sediment classification using an adaptive differential evolution ensemble of machine learning classifiers is presented in this section. First, segments shown in , generated using the OBIA approach described above, are divided into training and testing samples, which are fed to the machine learning base classifiers. Three different machine learning algorithms with diverse backgrounds are used so that proper evaluations can be made. Confidence scores are extracted from these classifiers, and their fusion is performed by optimization to yield the best overall accuracy with respect to the individual accuracies of the base classifiers. The base classifiers used are as follows:

  1. Random Forest

Random Forest (Breiman Citation2001) is an ensemble learning technique that builds many decision trees during training and produces the mean prediction (regression) or mode of the classes (classification) of the individual trees. The underlying principle of RF is to average the outputs from many decision trees, each trained with a random subset of the data, to enhance the performance and decrease overfitting of a single decision tree. RF can be expressed mathematically as D={(x1,y1),(x2,y2),,(xN,yN),i=1,2,,N}, where N indicates the number of instances, D represents the dataset and (xi,yi) represent inputs and labels, respectively. A bootstrap sample of the dataset is used to train each of the T decision trees that make up the random forest classifier. The predictions from all decision trees that received the most votes are the output of the random forest classifier: (1) CRF(x)=argmaxkt=1T(Ct(x)=k)(1) where k represents the class label and Ct(x) reflects the prediction made by the t-th decision tree for the input x.
(2)

Decision Tree

Each internal node in a decision tree (Webb et al. Citation2011) corresponds to a feature, each branch to a decision rule, and each leaf node to an outcome or class label. Decision trees resemble flowcharts. Decision trees use recursive binary splitting to create the tree structure, aiming to create the purest partitions at each level. Gini impurity and entropy are the two most typical splitting criteria. The decision tree classifier CDT converts an input feature vector x to a class label by following the decision rules from the root node to a leaf node for a given dataset D={(x1,y1),(x2,y2),,(xN,yN)},i=1,2,,N, where N indicates the number of instances, D represents the dataset and (xi,yi) represent inputs and labels, respectively. The majority class of the instances in that partition is represented by the label given to the leaf node.

(3)

Support Vector Machine

The goal of SVM (Cristianini and Ricci Citation2008; Vapnik Citation1995), a supervised learning algorithm, is to identify the best hyperplane that maximizes the distance between two classes in the feature space. The best hyperplane, or support vector, is the one with the greatest distance to the closest data points from both classes in a linearly separable dataset. Let D={(x1,y1),(x2,y2),,(xN,yN)},i=1,2,,N, where N indicates the number of instances, D represents the dataset and (xi,yi) represent inputs and labels, respectively, be a dataset with binary class labels yi1,1. The SVM classifier can be formulated as the following optimization problem: (2) minw,b(12|w|2)subjecttoyi(wxi+b)1,fori=1,,N(2) where b stands for the bias term and w stands for the weight vector normal to the hyperplane. Using kernel functions, the SVM may be expanded to translate input data to a higher-dimensional space where a linear separator can be discovered in the case of non-linearly separable data. The SVM classifier's output is represented as follows: (3) CSVM(x)=sign(wϕ(x)+b)(3) where (x) refers to the input data point, w represents the weight vector, b is the bias term and ϕ(x) represents the feature mapping or transformation applied to the input data point.

2.2.4. Differential evolution optimization

Differential evolution (DE) (Storn Citation1996) is an optimization algorithm that uses a population to look for and find optimal solutions by utilizing mutation, crossover, and selection operators. DE excels at maximizing functions that are not linear, convex, or differentiable. The proposed method utilizes DE to fine-tune the multi-classifier's weights (p) for maximum accuracy on a validation set. The p for each classifier is used as an optimization variable and weight when fusing the classifiers. The main steps of DE can be summarized as follows:

First, create an initial population P(0)=x1(0),x2(0),,xN(0) of N size, where each element xi(0) is a possible solution within the search space. Next, create a new population with a mutated version of each target vector xi(t) by applying the mutation operator to obtain vi(t). The most popular mutation strategy, DE/rand/1, where ‘DE’ stands for differential evolution, is the overarching optimization algorithm. ‘rand’ indicates that this strategy involves random selection of individuals from the population, and ‘1’ signifies that it uses a single difference vector in the mutation process, which is defined as follows: (4) vi(t)=xr1(t)+F(xr2(t)xr3(t))(4) where r1, r2, and r3 are distinct random indices from 1,2,,N and F is the scaling factor. Crossover is the next step after mutation. Here, based on the target vector xi(t)and its corresponding mutant vi(t), a new trial vector ui(t) is created using the crossover operator. The most widely used crossover scheme is binomial crossover, which is expressed as (5) uij(t)={vij(t),ifrandjCRorj=jrandxij(t),otherwise(5) where randju(0,1), CR represents the crossover probability, and jrand is a randomly selected index of range 1,2,,N. Finally, a search criterion using greedy selection is performed. The search is performed for the objective function values corresponding to the target vector xi(0)and its trial vector ui(t): (6) xi(t+1)={ui(t),iff(ui(t))f(xi(t))xi(t),otherwise(6) In addition to finding the optimal classifier weights (p) that maximize validation accuracy, the weights can be inverted to give higher weight to poorer performing classifiers. The inversion provides a robustness check on the optimization by testing whether weighting the least accurate classifiers can still improve the multi-classifier performance.

To achieve a high objective value representing the accuracy of the multi-classifier model, the study employs an adaptive search space bound (bound) approach. By dynamically modifying the search space based on the best solutions found, the adaptive search approach enhances optimization algorithms and improves the balance between exploration and exploitation, accelerates convergence, and enhances the algorithm's robustness. The search space for classifier weights, p, ranges from 0 to 1. To facilitate convergence, these bounds adapt to the best solutions. The adaptive bounds approach concentrates the search around promising solutions while still allowing exploration, thus improving optimization performance. The adaptive bounds can be expressed mathematically as follows: (7) {n_loweri=min(xkiδ,0)n_upperi=max(xki+δ,1)(7) for i=1,2,,m, where m is the number of base classifiers, δ determines the maximum allowable change and xki is the current value in the i-th dimension. From (7), the new bounds for the search space are given as: (8) nbounds=[(new_lower1,new_upper1),,(new_lowerm,new_upperm)](8)

2.2.5. Bayesian-based objective function

The objective function of the optimization approach described in the previous section applies Bayesian theory to combine the classifiers (RF, DT, and SVM) in an optimal manner. The predicted probability matrixes of the base classifiers are fused together and then weighted by the weights (p) in the multi-classifier. The objective function is then formulated as follows: Given a set of classifier confidence scores, C=[C1,C2,C3,,Cm] and their corresponding probabilities P=[P1,P2,,Pm], for class k, the multi-classifier's posterior probability is calculated as: (9) P(y=kx,p)=j=1mpjP(y=kx,Cj)(9) In Equation (9), P(y=k|x,p) represents the probability that a given data point x belongs to class k based on the weighted combination of the base classifiers. m represents the number of base classifiers in the multi-classifier fusion, represents the jth base classifier, and represents the weights assigned to the jth base classifier.

The class with the greatest predicted probability is the class selected for an instance. To evaluate the robustness of the optimized p, an inversion test is performed where the lowest performing classifiers receive the highest weight. The inverted weights are obtained by ranking the classifiers by validation accuracy and assigning priorities proportional to the inverse of their rank. The objective function and DE optimization remain the same, using the inverted (p1) instead of p. (10) Censemble(x,p)=argmaxkP(y=kx,p)(10) The approach aims to determine an optimized p, i.e. p, that leads to the highest accuracy of the multi-classifier model on a validation set. Therefore, the optimization problem is posed as follows: (11) p=argmaxkAccuracy(Censemble(Xval,p))(11) where p is the vector of weights for each classifier, Xval is the dataset, and Censemble is the predicted output of the multi-classifier model based on the weighted combination of classifier probabilities according to p. Additionally, an optimization algorithm has been developed, and it is presented in . The optimal weighted vector (p) is sought using the differential evolution approach, considering class probability matrices from various classifiers, including RF, DT, and SVM.

Table 3. Algorithm for differential evolution optimization.

2.3. Implementation

In performing sediment classification using the ensemble method defined, each of the grabbed sample datasets in is identified and located in a separate, unique segment after the OBIA segmentation (see ). No two sample points are identified in the same OBIA segment. Each segment therefore defines the homogenous region for that sample type. Based on the location of each grabbed sample point, 500 data points are generated per segment, amounting to 17,000 over the study area. Features from are then extracted for each point. The dataset is split into portions of 70:30 for the training and validation sets, respectively. The three classifiers – RF, DT, and SVM – are trained to obtain the predicted probabilities. A few cases are evaluated to test the robustness of the proposed method: (1) Using the standard dataset, the classifiers are trained on the training set and validated on the validation set. (2) To account for data imbalance, split training using under-sampling and over-sampling of the dataset is performed using the imbalanced-learn library in Python. (3) Finally, training using 10-fold cross-validation is applied. The predicted probabilities per classifier are computed. The objective function computes the negative ensemble accuracy using the weights of the predicted class probabilities from the base classifiers. The weights are determined, which are the optimization variables for the DE algorithm.

The parameters used by the DE algorithm for optimization are as follows: the strategy parameter is ‘DE/rand/1’, the population size for optimization is 1000 data points for split training and over-sampled training and 500 for under-sampled training, the tolerance level is set to 108, mutation equals (0.5, 1), and the optimization is run for a configured number of 1000 iterations. Each weight value is limited to [0, 1]. The best weights are changed at each iteration by either adjusting the margin or keeping the boundaries inside the [0, 1] interval. The DE algorithm can be applied by coupling the objective function with the adaptive bounds (7) – (8). The best possible weights are acquired following the completion of optimization. The fused model is obtained by computing the predictions of the validation dataset using the acquired optimal weights. To perform the optimization, the ‘differential_evolution’ package of SciPy is employed (package: from scipy.optimize import differential_evolution). The package provides tools and functions necessary to effectively implement the DE algorithm for optimization.

The approach using DE proves to be successful. The adaptive bounds mechanism reduces the number of iterations while achieving convergence at the global maximum. The same DE approach and parameters are used for split training, imbalance training, and cross-validation training. The final prediction is evaluated on the basis of accuracy, F1 score, and the kappa coefficient. Additionally, two correlation tests are used to determine the performance of the multi-classifier model: Pearson’s correlation test and Spearman’s correlation test. The aim is to obtain a positive correlation between the confidence score and the predicted accuracy.

2.4. Evaluation metrics

The metrics of evaluation employed to ascertain the efficacy of the suggested technique are now defined.

Accuracy refers to the proportion of samples that have been classified correctly in relation to the size of the dataset. The formula utilized to compute accuracy is presented in (12), and it takes into account four fundamental variables: true positives (TPs), true negatives (TNs), false positives (FPs), and false negatives (FNs). True positives represent instances correctly classified as belonging to the target class, while true negatives are instances correctly classified as not belonging to the target class. False positives correspond to instances incorrectly classified as belonging to the target class, and false negatives are instances incorrectly classified as not belonging to the target class. (12) Accuracy=TP+TNTP+TN+FP+FN(12) The F1 score (F1) is computed as the harmonic average between the precision and recall scores. Its calculation is performed using the following formula: (13) F1=2×Precision×RecallPrecision×Recall(13) where F1 represents the F1 score, Precision denotes precision, and Recall stands for recall. Precision measures the proportion of true-positive predictions among all positive predictions made by the model. Recall, on the other hand, calculates the proportion of true-positive predictions among all actual positive instances in the dataset. By taking the harmonic average of precision and recall, the F1 score provides a single value that combines both measures, making it particularly useful when there is a need to balance precision and recall in a classification task. It serves as an important indicator of a model's ability to make accurate positive predictions while minimizing false positives and false negatives.

The kappa coefficient (κ) is a numerical indicator utilized to measure the level of agreement between the ground-truth data and the predicted outcomes. This measure can be computed using the following formula: (14) κ=PoPe1Po(14) where κ represents the kappa coefficient, P_o is the relative observed agreement between the ground-truth and predicted outcomes and P_e is the probability of agreement occurring by chance.

3. Results

3.1. Optimization results

The optimization approach adopted in this study aims to maximize the accuracy of the resulting multi-classifier model with respect to the weights. In other words, finding an optimal posterior probability matrix is based on the fusion of the base classifiers’ prediction probabilities. The results of the optimization based on the method described in Section 2 are presented in this section.

The optimization results are presented for three cases – under-sampled, standard, and over-sampled – to demonstrate the robustness of the proposed approach. presents the optimal weight values, which are assigned to each classifier as described in Section 2.2.5 and , along with their corresponding accuracies on the training dataset obtained after the three different cases: over-sampled, training on the standard, and under-sampled. To check for the effectiveness of the optimization, the weights are also inversed (see Section 2.2.4 and 2.2.5) to give more attention to the poorest-performing classifiers. The results for the inverted weights are also given in . presents changes in bounds observed during optimization compared to the initial bounds, where the bound adaptation is defined in Section 2.2.4. shows the objective value, convergence, and the steps taken during optimization.

Figure 4. Summary of the optimization results. The left column shows plots of the objective function over optimization iteration. The centre column shows the optimization steps for achieving the optimal weights, and the right column provides a 3D visualization of the convergence. The rows correspond to the different datasets: (a) under-sampled, (b) standard, and (c) over-sampled.

Figure 4. Summary of the optimization results. The left column shows plots of the objective function over optimization iteration. The centre column shows the optimization steps for achieving the optimal weights, and the right column provides a 3D visualization of the convergence. The rows correspond to the different datasets: (a) under-sampled, (b) standard, and (c) over-sampled.

Table 4. Optimization results on all training datasets (under-sampled, standard, and over- sampled).

Table 5. Search space bounds on the weights acquired for each optimization are presented in the table below.

3.2. Classification and fusion results

To evaluate the performance of the proposed multi-classifer approach, the accuracy, F1 score, and κ are compared to those of the individual base classifiers on the validation dataset. The validation dataset used for evaluation is described in Section 2.3. The results are presented across three cases – under-sampling, standard, and over-sampling – to demonstrate robustness. present a comparison of the accuracy, F1 score, and κ and AUC concerning the proposed approach and each of the base classifiers. All results in this section are evaluated based on the validation dataset described in Section 2.3. More specifically, gives the under-sampled training results. compares the proposed and base classifiers using the standard dataset. Finally, presents the over-sampled result. The resulting confusion matrix is presented in , and the AUC is also shown for the cases.

Figure 5. Confusion matrix. (a) Under-sampled training (b) Standard (c) Over-sampled training.

Figure 5. Confusion matrix. (a) Under-sampled training (b) Standard (c) Over-sampled training.

Table 6. Classification results for under-sampled training.

Table 7. Classification results for the standard dataset.

Table 8. Classification results for the over-sampled dataset.

Additional assessment was conducted utilizing 10-fold cross-validation on the standard data, whereby the accuracy, F1 score, and κ for each fold are documented in . The uniformity in performance across these folds demonstrates the resilience of the proposed methodology. Moreover, the favourable correlation values obtained from Pearson’s and Spearman’s tests, as presented in , evince a strong association between the anticipated accuracy and classifier confidence. The multi-classifier confidence in refers to the confidence scores (C) output by the multi-classifier for each validation sample, as defined in Section 2.2.5. The correlation is computed between these confidence scores (C) and the prediction accuracy for each validation sample. In other words, the confidence score vector is compared to a vector of 0s and 1s representing whether each prediction was correct over the validation set. This aligns with the motivation for the confidence-based fusion strategy. Finally, visually portrays the resulting seabed sediment classification maps generated by the proposed multi-classifier approach in contrast to the individual classifiers.

Figure 6. The generated seabed sediment maps after classification. (a) represents the sediment map for the proposed approach. (b) Sediment map for RF. (c) represents the SVM sediment map. (d) Sediment map for DT classification.

Figure 6. The generated seabed sediment maps after classification. (a) represents the sediment map for the proposed approach. (b) Sediment map for RF. (c) represents the SVM sediment map. (d) Sediment map for DT classification.

Table 9. Classification results for the 10-fold cross-validation training on the standard dataset.

Table 10. Pearson’s and Spearman’s correlation tests between the confidence (C) and predicted accuracy of the proposed multi-classifiers defined in Section 2.

4. Discussion

The optimization strategy developed and used in this paper aims to produce a multi-classifier model with the highest accuracy. The approach combines the confidence scores of the base classifiers into a single posterior probability matrix to accurately classify seabed sediments. This section discusses the results in Section 3 using the method outlined in Section 2.

As shown in , all optimization processes successfully converged at the global maximum. The optimization results in (left column) illustrate that training converged faster than the configured number of 1000 iterations defined in Section 2.3, while achieving a higher objective value (accuracy) than the base classifiers. The successful convergence observed in is a result of implementing the adaptive bounds approach during optimization, as detailed in Section 2.2.4 (7)–(8) (Rauf, Bangyal, and Lali Citation2021; K. Farda and Thammano Citation2022; Wu et al. Citation2022). The purpose of including this information in is to illustrate how the bounds on the weights were defined and how they evolved during the optimization process. These bounds are crucial in shaping the optimization problem and controlling the search space for the weights assigned to the classifiers, which, in turn, influence the optimization algorithm’s behaviour. Tighter bounds can restrict the search space, while wider bounds allow for more flexibility. The adaptive nature of these bounds is underscored by their relevance:

  • Initially, the algorithm explores a broader weight range. As it iterates and converges, it refines these bounds. This dynamic adjustment is akin to a self-regulating mechanism, where the algorithm adapts its search space based on observed improvements or deteriorations in accuracy.

  • The different datasets (under-sampled, standard, and over-sampled) present varying challenges due to their distinct data characteristics. As a result, the optimization algorithm may adapt its weight bounds differently for each dataset. This adaptability showcases the algorithm’s ability to tailor its strategy to the specific nuances of the dataset at hand.

  • Understanding how weight bounds evolve offers insight into the degree of autonomy granted to the optimization algorithm and can be perceived as a manifestation of the algorithm’s decision-making procedure, where it dynamically determines the constraints required to achieve optimal results (see ).

  • The evolution of bounds directly impacts the optimization’s efficiency and effectiveness. Tighter final bounds indicate more precise convergence, while broader bounds may imply a more exploratory optimization process.

The results of the objective values presented in indicate that the optimization effectively improves the classification performance of the multi-classifier fusion compared to individual classifiers. displays the optimized weight for each classifier on each training dataset. The higher objective accuracy achieved demonstrates that the proposed approach successfully learns an improved weighted combination of the classifier probabilities compared to equal weights. As shown in , the proposed optimization achieves higher training accuracy than the individual classifiers for each dataset. demonstrate the improvement in the multi-classifier generalization through higher validation metrics than those of the base classifiers across the different cases. Specifically, the results reveal the following: (1) For under-sampled training, the objective value of 73.656% is higher than the 71.81%, 60.81%, and 63.31% values for RF, DT, and SVM, respectively. (2) With standard dataset training, the objective value of 79.139% exceeds the values of 76.76%, 70.0%, and 70.96% for RF, DT, and SVM, respectively. (3) In the context of over-sampled training, the 95.01% objective value is greater than the respective values for RF (92.85%), DT (92.28%), and SVM (78.57%). As the quantity of data increases, evidenced by Batista, Prati, and Monard (Citation2004), the enhancement in performance is evident not only in accuracy but also in F1 score, κ, and AUC – critical performance indicators for classification tasks.

From an optimization perspective, the weights can be used to describe regions of high importance in the search space based on the confidence scores. In other words, based on the confidence score, the optimization is able to adjust the weights and select the predictions with the highest confidence score, thereby increasing the likelihood of finding the optimal global maximum. The adjustment is relevant when there are conflicting predictions among several base classifiers. The use of confidence scores in the optimization approach helps overcome the limitations of the traditional deterministic multi-classifier approach (Wan et al. Citation2022; Yang et al. Citation2022).

The positive values of the correlation test given in show that all three strategies demonstrate a positive association between predicted accuracy and the multi-classifier’s confidence, supporting the notion that more confidence in prediction is linked to better performance. According to (Carpenter et al. Citation1999; Foody, Boyd, and Sanchez-Hernandez Citation2007), confidence scores are important in multi-classifiers due to (1) agreement between classifiers and (2) correlations between the predicted accuracies. The argument is also supported by Kuncheva (Citation2004), who discussed the methods for achieving high multi-classifier performance by taking advantage of the positive correlation between the confidence score and the predicted accuracy.

The consistent accuracy of both weight and inverse weight findings across all trainings shown in provides strong evidence for the usefulness of inversing the weight in this setting. The inverse weights are used as a control to prioritize the least efficient classifiers, which boosts the efficiency of the algorithm even further. The approach excels when the multi-classifier model is made up of diverse base classifiers with overlapping capabilities. Even if they perform poorly, the weaker classifiers in such a multi-classifier model may hold information that improves the overall predictive accuracy (Bortolotti Citation2017; Dietterich Citation2000). The multi-classifier approach using inverse weights achieves high performance across a wide range of training situations by dynamically adjusting the weights given to individual base classifiers based on their individual performance.

The improved performance of the multi-classifier over the individual classifiers confirms the claim presented by Diesing and Stephens (Citation2015) that the mapping of maritime habitats is presently transitioning from expert analysis by interpretation to automatic seabed categorization techniques. While several methods and classifiers have been tried, there is no agreement on ‘what works best’, and there is unlikely to be a single solution that meets all requirements. Even though individual classifiers did perform well based on , their combination neutralizes their individual errors and enhances total performance because each model has the capacity to learn unique patterns from the underlying dataset (Chen et al. Citation2021).

The 10-fold cross-validation results further support the robustness of the proposed method. Based on the values in , the proposed approach outperforms the base classifiers in each fold. It should be noted that while cross-validation is performed on the standard, it achieves higher metric scores than the standard training approach in . The high average metric scores found in the cross-validation underline the significance of using a suitable quantity of training data to obtain improved results. This observation also explains the high performance in the over-sampled results in compared to that of the under-sampled and standard training results. The consistent success across various folds suggests that the technique generalizes well to new data and is not overly sensitive to the choice of training data.

The resulting sediment map for the four classifiers is shown in . It appears that the proposed approach shows more distinct sediment classes. Upon visual inspection, the proposed multi-classifier approach shown in produces sediment class distributions that appear more spatially consolidated and distinct than those in the individual classifier maps. For example, the mS class shows more continuity along the central region without fragmentation in the proposed map. Additionally, the (g)mS and (g)M areas are more confined to the eastern section without intermixing with other classes. Overall, the sediment class boundaries are better delineated in the proposed multi-classifier map, with less fragmentation and intermixing of classes due to misclassifications.

Figure 7. The figure illustrates discrepancies in sediment maps between the proposed method and the base classifiers. The figure is divided into four sub-figures, each representing a distinct area of interest. Sub-figure (a) corresponds to the area of interest in the sediment map generated by the proposed method, while sub-figures (b), (c), and (d) correspond to the areas of interest in the sediment maps generated by RT, SVM, and DT classifiers, respectively.

Figure 7. The figure illustrates discrepancies in sediment maps between the proposed method and the base classifiers. The figure is divided into four sub-figures, each representing a distinct area of interest. Sub-figure (a) corresponds to the area of interest in the sediment map generated by the proposed method, while sub-figures (b), (c), and (d) correspond to the areas of interest in the sediment maps generated by RT, SVM, and DT classifiers, respectively.

Additionally, the significant correlation coefficients derived from Pearson’s and Spearman’s tests in provide compelling evidence for a robust connection between the predicted accuracy and classifier confidence scores. The strong Pearson's correlations, with coefficients equal to or exceeding 0.675, clearly indicate a powerful positive linear association. In parallel, the elevated Spearman’s values also substantiate a pronounced monotonic relationship between accuracy and confidence. The correlation test aligns perfectly with the underlying rationale of employing confidence scores to enhance fusion, since heightened confidence values exhibit a strong correlation with more precise predictions.

Multiclassifier models enable the integration of the key attributes from all participating classifiers, resulting in superior performance compared to individual models. Over time, numerous well-known multi-classifier techniques have been introduced in prior research. provides a comprehensive comparative analysis of our proposed methodology in relation to previously established approaches found in the literature. The comparison in encompasses various scenarios, including under-sampled, standard, over-sampled, and cross-validation (cv) cases.

Table 11. Presents a comparative analysis of the proposed methodology with previously established approaches in the literature across various scenarios, encompassing under-sampled, standard, over-sampled, and cross-validation (cv) cases.

5. Conclusion

The prediction abilities of various basic classifiers (random forest, decision tree, and support vector machine) were combined in this study to produce a multi-classifer approach for precisely classifying seafloor sediments. The goal of the optimization strategy was to increase the multi-classifier's performance. The proposed multi-classifier approach outperformed each individual basic classifier in terms of accuracy, F1 score, the κ, and the AUC. The 10-fold cross-validation results demonstrated the generalizability of the method by showing that the classification is not overly dependent on training data selection and may generate consistent results across different data divisions. In the future, different optimization methods can be tested to further improve the proposed approach.

Acknowledgments

The authors want to thank the British Geological Survey (BGS) provides the multi-beam dataset and ground-truth samples.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Data availability statement

The data that support the findings of this study are openly available in ‘public geographic science database belongs to the British Geological Survey (BGS)’ at http://mapapps2.bgs.ac.uk/geoindex_offshore/home.html.

Additional information

Funding

This research was supported by the National Natural Science Foundation of China under grant [number 52201400], and supported by Shandong Provincial Natural Science Foundation under grant [number ZR202111260306], and supported by Special Projects for Promoting High Quality Economic Development (Marine Economic Development) in Guangdong Province under grant (GDNRC[2023]42), and Funding of Key Laboratory of Submarine Geoscience, Ministry of Natural Resources, under grant number KLSG2203.

References

  • Barbagelata, A., M. Richardson, B. Miaschi, E. Muzi, P. Guerrini, L. Troiano, and T. Akal. 1991. “ISSAMS: An In Situ Sediment Acoustic Measurement System.” Shear Waves in Marine Sediments, 305–312. https://doi.org/10.1007/978-94-011-3568-9_34.
  • Batista, Gustavo E. A. P. A., Ronaldo C. Prati, and Maria Carolina Monard. 2004. “A Study of the Behavior of Several Methods for Balancing Machine Learning Training Data.” ACM SIGKDD Explorations Newsletter 6 (1): 20–29. https://doi.org/10.1145/1007730.1007735.
  • Bernardo, José M., and Adrian F. M. Smith. 2008. “Bayesian Theory.” Bayesian Theory 5: 1–595. https://doi.org/10.1002/9780470316870.
  • Bortolotti, Roberta. 2017. “Data Prep 2-1: Data Cleaning and Recoding.” In Handbook of Statistical Analysis and Data Mining Applications. Elsevier. https://doi.org/10.1016/B978-0-12-416632-5.00032-3.
  • Breiman, Leo. 2001. “Random Forests.” Machine Learning 45 (1): 5–32. https://doi.org/10.1023/A:1010933404324.
  • Bressan, Thiago Santi, Marcelo Kehl de Souza, Tiago J. Girelli, and Farid Chemale Junior. 2020. “Evaluation of Machine Learning Methods for Lithology Classification Using Geophysical Data.” Computers & Geosciences 139 (6): 104475. https://doi.org/10.1016/J.CAGEO.2020.104475.
  • Carpenter, Gail A., Sucharita Gopal, Scott Macober, Siegfried Martens, Curtis E. Woodcock, and Janet Franklin. 1999. “A Neural Network Method for Efficient Vegetation Mapping.” Remote Sensing of Environment 70 (3): 326–338. https://doi.org/10.1016/S0034-4257(99)00051-6.
  • Chen, Jiefeng, Frederick Liu, Besim Avci, Xi Wu, Yingyu Liang, and Somesh Jha. 2021. “Detecting Errors and Estimating Accuracy on Unlabeled Data with Self-training Ensembles.” Advances in Neural Information Processing Systems 34:14980–14992. https://arxiv.org/abs/2106.15728v3.
  • Cristianini, Nello, and Elisa Ricci. 2008. “Support Vector Machines.” Encyclopedia of Algorithms, 928–932. https://doi.org/10.1007/978-0-387-30162-4_415.
  • Cui, Xiaodong, Fanlin Yang, Xin Wang, Bo Ai, Yu Luo, and Dan Ma. 2021. “Deep Learning Model for Seabed Sediment Classification Based on Fuzzy Ranking Feature Optimization.” Marine Geology 432:106390. https://doi.org/10.1016/j.margeo.2020.106390.
  • Damveld, Johan H., Pieter C. Roos, Bas W. Borsje, and Suzanne J. M. H. Hulscher. 2019. “Modelling the Two-way Coupling of Tidal Sand Waves and Benthic Organisms: A Linear Stability Approach.” Environmental Fluid Mechanics 19 (5): 1073–1103. https://doi.org/10.1007/s10652-019-09673-1.
  • Diesing, Markus, Peter J. Mitchell, Eimear O’keeffe, Giacomo O. A. Montereale Gavazzi, and Tim Le Bas. 2020. “Limitations of Predicting Substrate Classes on a Sedimentary Complex But Morphologically Simple Seabed.” Remote Sensing 12 (20): 3398. https://doi.org/10.3390/rs12203398.
  • Diesing, Markus, and David Stephens. 2015. “A Multi-model Ensemble Approach to Seabed Mapping.” Journal of Sea Research 100:62–69. https://doi.org/10.1016/j.seares.2014.10.013.
  • Dietterich, Thomas G. 2000. “Ensemble Methods in Machine Learning.” Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) 1857:1–15. https://doi.org/10.1007/3-540-45014-9_1/COVER.
  • Divyabarathi, G., S. Shailesh, M. V. Judy, and R. Krishnakumar. 2022. “A Novel Ensemble Method for Underwater Mines Classification.” Lecture Notes on Data Engineering and Communications Technologies 96:555–565. https://doi.org/10.1007/978-981-16-7167-8_40.
  • Farda, Irfan, and Arit Thammano. 2022. “A Self-adaptive Differential Evolution Algorithm for Solving Optimization Problems.” Lecture Notes in Networks and Systems 453:68–76. https://doi.org/10.1007/978-3-030-99948-3_7.
  • Foody, G. M., D. S. Boyd, and C. Sanchez-Hernandez. 2007. “Mapping a Specific Class with an Ensemble of Classifiers.” International Journal of Remote Sensing 28 (8): 1733–1746. https://doi.org/10.1080/01431160600962566.
  • Frederick, Christina, Soledad Villar, and Zoi-Heleni Michalopoulou. 2020. “Seabed Classification Using Physics-based Modeling and Machine Learning.” The Journal of the Acoustical Society of America 148 (2): 859–872. https://doi.org/10.1121/10.0001728.
  • Gallant, J. C., and T. I. Dowling – Water resources research, and undefined. 2003. “A Multiresolution Index of Valley Bottom Flatness for Mapping Depositional Areas.” Wiley Online Library 39 (12): 1347. https://doi.org/10.1029/2002WR001426.
  • Galvez, Daphnie S., Svenja Papenmeier, Lasse Sander, Alexander Bartholomä, and Karen Helen Wiltshire. 2022. “Ensemble Mapping as an Alternative to Baseline Seafloor Sediment Mapping and Monitoring.” Geo-Marine Letters 42 (3): 1–13. https://doi.org/10.1007/S00367-022-00734-X/FIGURES/5.
  • Gary Greene, H., David A. Cacchione, and Monty A. Hampton. 2017. “Characteristics and Dynamics of a Large Sub-Tidal Sand Wave Field—Habitat for Pacific Sand Lance (Ammodytes Personatus), Salish Sea, Washington, USA.” Geosciences 7 (4): 107. https://doi.org/10.3390/geosciences7040107.
  • Goulden, T., C. Hopkinson, R. Jamieson, and S. Sterling. 2016. “Sensitivity of DEM, Slope, Aspect and Watershed Attributes to LiDAR Measurement Uncertainty.” Remote Sensing of Environment 179:23–35. https://doi.org/10.1016/j.rse.2016.03.005.
  • Gudiyangada Nachappa, Thimmaiah, Sepideh Tavakkoli Piralilou, Khalil Gholamnia, Omid Ghorbanzadeh, Omid Rahmati, and Thomas Blaschke. 2020. “Flood Susceptibility Mapping with Machine Learning, Multi-criteria Decision Analysis and Ensemble Using Dempster Shafer Theory.” Journal of Hydrology 590:125275. https://doi.org/10.1016/J.JHYDROL.2020.125275.
  • Guisan, Antoine, Stuart B. Weiss, and Andrew D. Weiss. 1999. “GLM versus CCA Spatial Modeling of Plant Species Distribution.” Plant Ecology 143 (1): 107–122. https://doi.org/10.1023/A:1009841519580.
  • Haralick, Robert M., Its’hak Dinstein, and K. Shanmugam. 1973. “Textural Features for Image Classification.” IEEE Transactions on Systems, Man and Cybernetics SMC-3 (6): 610–621. https://doi.org/10.1109/TSMC.1973.4309314.
  • He, Linbang, Jianhu Zhao, Zhenge Qiu, and Jie Feng. 2022. “High-accuracy Seabed Sediment Classification Using Multi-beam Acoustic Backscatter Data.” Oceans Conference Record (IEEE). https://doi.org/10.1109/OCEANSCHENNAI45887.2022.9775530.
  • Hosack, Geoffrey R., Brett R. Dumbauld, Jennifer L. Ruesink, and David A. Armstrong. 2006. “Habitat Associations of Estuarine Species: Comparisons of Intertidal Mudflat, Seagrass (Zostera Marina), and Oyster (Crassostrea Gigas) Habitats.” Estuaries and Coasts 29 (6): 1150–1160. https://doi.org/10.1007/BF02781816.
  • Ierodiaconou, D., J. Monk, A. Rattray, L. Laurenson, and V. L. Versace. 2011. “Comparison of Automated Classification Techniques for Predicting Benthic Biological Communities Using Hydroacoustics and Video Observations.” Continental Shelf Research 31 (2): S28–S38. https://doi.org/10.1016/j.csr.2010.01.012.
  • Jasiewicz, Jarosław, and Tomasz F. Stepinski. 2013. “Geomorphons-a Pattern Recognition Approach to Classification and Mapping of Landforms.” Geomorphology 182: 147–156. https://doi.org/10.1016/j.geomorph.2012.11.005.
  • Kim, Byung Chul, Hoe Chang Kim, Sungho Han, and Dong Kyou Park. 2022. “Inspection of Underwater Hull Surface Condition Using the Soft Voting Ensemble of the Transfer-learned Models.” Sensors 22 (12): 4392. https://doi.org/10.3390/s22124392.
  • Kostylev, V. E., B. J. Todd, G. B. J. Fader, R. C. Courtney, G. D. M. Cameron, and R. A. Pickrill. 2001. “Benthic Habitat Mapping on the Scotian Shelf Based on Multibeam Bathymetry, Surficial Geology and Sea Floor Photographs.” Marine Ecology Progress Series 219:121–137. https://doi.org/10.3354/meps219121.
  • Kuncheva, Ludmila I. 2004. “Combining Pattern Classifiers.” Combining Pattern Classifiers, July. https://doi.org/10.1002/0471660264.
  • Leon, Florin, Sabina Adriana Floria, and Costin Badica. 2017. “Evaluating the Effect of Voting Methods on Ensemble-Based Classification.” Proceedings – 2017 IEEE International Conference on Innovations in Intelligent Systems and Applications, 1–6. https://doi.org/10.1109/INISTA.2017.8001122.
  • Leon, America Zelada, Veerle A. I. Huvenne, Noëlie M. A. Benoist, Matthew Ferguson, Brian J. Bett, and Russell B. Wynn. 2020. “Assessing the Repeatability of Automated Seafloor Classification Algorithms, with Application in Marine Protected Area Monitoring.” Remote Sensing 12 (10): 1572. https://doi.org/10.3390/rs12101572.
  • Liu, Haoran, Kehui Xu, Bin Li, Ya Han, and Guandong Li. 2019. “Sediment Identification Using Machine Learning Classifiers in a Mixed-texture Dredge Pit of Louisiana Shelf for Coastal Restoration.” Water 11 (6): 1257. https://doi.org/10.3390/w11061257.
  • Lüdtke, A., K. Jerosch, O. Herzog, and M. Schlüter. 2012. “Development of a Machine Learning Technique for Automatic Analysis of Seafloor Image Data: Case Example, Pogonophora Coverage at Mud Volcanoes.” Computers & Geosciences 39:120–128. https://doi.org/10.1016/j.cageo.2011.06.020.
  • Marsh, Ivor, and Colin Brown. 2009. “Neural Network Classification of Multibeam Backscatter and Bathymetry Data from Stanton Bank (Area IV).” Applied Acoustics 70 (10): 1269–1276. https://doi.org/10.1016/j.apacoust.2008.07.012.
  • McGonigle, Chris, and Jenny S. Collier. 2014. “Interlinking Backscatter, Grain Size and Benthic Community Structure.” Estuarine, Coastal and Shelf Science 147:123–136. https://doi.org/10.1016/j.ecss.2014.05.025.
  • McGonigle, Chris, Jonathan H. Grabowski, Craig J. Brown, Thomas C. Weber, and Rory Quinn. 2011. “Detection of Deep Water Benthic Macroalgae Using Image-based Classification Techniques on Multibeam Backscatter at Cashes Ledge, Gulf of Maine, USA.” Estuarine, Coastal and Shelf Science 91 (1): 87–101. https://doi.org/10.1016/j.ecss.2010.10.016.
  • McLaren, Kurt, Karen McIntyre, and Kurt Prospere. 2019. “Using the Random Forest Algorithm to Integrate Hydroacoustic Data with Satellite Images to Improve the Mapping of Shallow Nearshore Benthic Features in a Marine Protected Area in Jamaica.” GIScience & Remote Sensing 56 (7): 1065–1092. https://doi.org/10.1080/15481603.2019.1613803.
  • Morgan, Dane, and Ryan Jacobs. 2020. “Opportunities and Challenges for Machine Learning in Materials Science.” Annual Review of Materials Research 50 (1): 71–103. https://doi.org/10.1146/annurev-matsci-070218-010015.
  • Mudiyanselage, S. S. J. D., A. Abd-Elrahman, B. Wilkinson, and V. Lecours. 2022. “Satellite-derived Bathymetry Using Machine Learning and Optimal Sentinel-2 Imagery in South-West Florida Coastal Waters.” GIScience & Remote Sensing 59 (1): 1143–1158. https://doi.org/10.1080/15481603.2022.2100597.
  • Parihar, Shradha Singh, and Nitin Malik. 2022. “Fuzzy-based Real-coded Genetic Algorithm for Optimizing Non-convex Environmental Economic Loss Dispatch.” Facta Universitatis – Series: Electronics and Energetics 35 (4): 495–512. https://doi.org/10.2298/FUEE2204495P.
  • Prasad, Pankaj, Victor Joseph Loveson, Mahender Kotha, and Ramanand Yadav. 2020. “Application of Machine Learning Techniques in Groundwater Potential Mapping along the West Coast of India.” GIScience & Remote Sensing 57:735–752. https://doi.org/10.1080/15481603.2020.1794104.
  • Priyavada, Binay Kumar. 2023. “Solution of Unconstrained Non-linear Programming Problems Using Differential Evolution, Genetic Algorithm and Artificial Bee Colony Evolutionary Algorithms: A Comparative Study.” Mathematical Statistician and Engineering Applications 72 (1): 1692–1712. https://doi.org/10.17762/MSEA.V72I1.2399.
  • Rauf, Hafiz Tayyab, Waqas Haider Khan Bangyal, and M. Ikramullah Lali. 2021. “An Adaptive Hybrid Differential Evolution Algorithm for Continuous Optimization and Classification Problems.” Neural Computing and Applications 33 (17): 10841–10867. https://doi.org/10.1007/s00521-021-06216-y.
  • Riley, Shawn J., Stephen D. DeGloria, and Robert Elliot. 1997. “Index that Quantifies Topographic Heterogeneity.” intermountain Journal of sciences 5 (1–4): 23–27.
  • Sappington, J. Mark, Kathleen M. Longshore, and Daniel B. Thompson. 2007. “Quantifying Landscape Ruggedness for Animal Habitat Analysis: A Case Study Using Bighorn Sheep in the Mojave Desert.” The Journal of Wildlife Management 71 (5): 1419–1426. https://doi.org/10.2193/2005-723.
  • Singh, Dalwinder, and Birmohan Singh. 2020. “Investigating the Impact of Data Normalization on Classification Performance.” Applied Soft Computing 97:105524. https://doi.org/10.1016/J.ASOC.2019.105524.
  • Storn, Rainer. 1996. “On the Usage of Differential Evolution for Function Optimization.” Biennial Conference of the North American Fuzzy Information Processing Society – NAFIPS, 519–523. https://doi.org/10.1109/NAFIPS.1996.534789.
  • Tulyakov, Sergey, Stefan Jaeger, Venu Govindaraju, and David Doermann. 2008. “Review of Classifier Combination Methods.” Studies in Computational Intelligence 90:361–386. https://doi.org/10.1007/978-3-540-76280-5_14/COVER.
  • Vapnik, Vladimir N. 1995. “The Nature of Statistical Learning Theory.” The Nature of Statistical Learning Theory. https://doi.org/10.1007/978-1-4757-2440-0.
  • Viala, Christophe, Marie Lamouret, and Arnaud Abadie. 2021. “Seafloor Classification Using a Multibeam Echo Sounder: A New Rugosity Index Coupled with a Pixel-Based Process to Map Mediterranean Marine Habitats.” Applied Acoustics 179 (August): 108067. https://doi.org/10.1016/j.apacoust.2021.108067.
  • Wan, Jiaxin, Zhiliang Qin, Xiaodong Cui, Fanlin Yang, Muhammad Yasir, Benjun Ma, and Xueqin Liu. 2022. “MBES Seabed Sediment Classification Based on a Decision Fusion Method Using Deep Learning Model.” Remote Sensing 14 (15): 3708. https://doi.org/10.3390/rs14153708.
  • Wang, Fengfan, Jia Yu, Zhijie Liu, Min Kong, and Yunfan Wu. 2021. “Study on Offshore Seabed Sediment Classification Based on Particle Size Parameters Using XGBoost Algorithm.” Computers & Geosciences 149:104713. https://doi.org/10.1016/j.cageo.2021.104713.
  • Webb, Geoffrey I., Johannes Fürnkranz, Geoffrey Hinton, Claude Sammut, and Joerg Sander. 2011. “Decision Tree.” Encyclopedia of Machine Learning, 263–267. https://doi.org/10.1007/978-0-387-30164-8_204.
  • Weiss, A. 2001. “Topographic Position and Landforms Analysis.” Jennessent.Com. http://www.jennessent.com/downloads/TPI-poster-TNC_18x22.pdf.
  • Wu, Guopeng, Guoxiong Chen, Qiuming Cheng, Zhenjie Zhang, and Jie Yang. 2021. “Unsupervised Machine Learning for Lithological Mapping Using Geochemical Data in Covered Areas of Jining, China.” Natural Resources Research 30 (2): 1053–1068. https://doi.org/10.1007/s11053-020-09788-z.
  • Wu, Kaijun, Zhengnan Liu, Ning Ma, and Dicong Wang. 2022. “A Dynamic Adaptive Weighted Differential Evolutionary Algorithm.” Computational Intelligence and Neuroscience 2022:1–10. https://doi.org/10.1155/2022/1318044.
  • Yang, Junfang, Yi Ma, Yabin Hu, Zongchen Jiang, Jie Zhang, Jianhua Wan, and Zhongwei Li. 2022. “Decision Fusion of Deep Learning and Shallow Learning for Marine Oil Spill Detection.” Remote Sensing 14 (3): 666. https://doi.org/10.3390/rs14030666.
  • Zajac, Roman N. 2008. “Challenges in Marine, Soft-sediment Benthoscape Ecology.” Landscape Ecology 23 (S1): 7–18. https://doi.org/10.1007/s10980-007-9140-4.
  • Zevenbergen, Lyle W., and Colin R. Thorne. 1987. “Quantitative Analysis of Land Surface Topography.” Earth Surface Processes and Landforms 12 (1): 47–56. https://doi.org/10.1002/esp.3290120107.
  • Zhang, Quanyin, Jianhu Zhao, Shaobo Li, and Hongmei Zhang. 2022. “Seabed Sediment Classification Using Spatial Statistical Characteristics.” Journal of Marine Science and Engineering 10 (5): 691. https://doi.org/10.3390/jmse10050691.
  • Zhao, Gang, Bo Pang, Zongxue Xu, and Liyang Xu. 2020. “A Hybrid Machine Learning Framework for Real-time Water Level Prediction in High Sediment Load Reaches.” Journal of Hydrology 581 (February): 124422. https://doi.org/10.1016/j.jhydrol.2019.124422.
  • Zhao, Jian, Bochen Zhang, Xiwang Guo, Liang Qi, and Zhiwu Li. 2022. “Self-adapting Spherical Search Algorithm with Differential Evolution for Global Optimization.” Mathematics 10 (23): 4519. https://doi.org/10.3390/math10234519.