713
Views
0
CrossRef citations to date
0
Altmetric
Original Articles

Feature subset selection in structural health monitoring data using an advanced binary slime mould algorithm

&

ABSTRACT

Feature Selection (FS) is an important step in data-driven structural health monitoring approaches. In this paper, an Advanced version of the Binary Slime Mould Algorithm (ABSMA) is introduced for feature subset selection to improve the performance of structural damage classification techniques. Two operators of mutation and crossover are embedded to the algorithm, to overcome the stagnation situation involved in the Binary Slime Mould Algorithm (BSMA). The proposed ABSMA is then embedded in a new data-driven SHM framework which consists of three main steps. In the first step, structural time domain responses are collected and pre-processed to extract the statistical features. In the second step, the order of the extracted features is reduced using an optimization algorithm to find a minimal subset of salient features by removing irrelevant, and redundant data. Finally, the optimized feature vectors are used as inputs to Neural Network (NN) based classification models. Benchmark datasets of a timber bridge model and a three-story frame structure are employed to validate the proposed algorithm. The results show that the proposed ABSMA provides a better performance and convergence rate compared to other commonly used binary optimization algorithms.

Introduction

Vibration-based structural health monitoring (SHM) has been widely explored over the past decades. Avci et al. (Citation2021) and Das et al. (Citation2016) presented comprehensive reviews of vibration-based damage detection methods and their applications to civil infrastructure. Recently, with the fast development of sensing technologies (Corbally & Malekjafarian, Citation2022; Malekjafarian et al., Citation2021), signal processing techniques (Silik et al., Citation2021, Citation2022), and machine learning approaches (Ghiasi et al., Citation2016; Malekjafarian et al., Citation2019), data-driven SHM approaches have attracted a high attention for damage detection of civil infrastructure (Gharehbaghi et al., Citation2021; Gomes et al., Citation2018). Vibration-based SHM methods can be mainly classified into two categories: (a) modal-based approaches which are based on vibratory characteristics of structural systems, such as natural frequencies, mode shapes and curvatures (Avci et al., Citation2021) and (b) data-driven approaches which extract sensitive features from time domain responses to assess the structural conditions (Dadras Eslamlou & Huang, Citation2022). Data-driven damage detection approaches can be performed in the time domain from the raw sensor data or in the feature domain, in which damage-sensitive features are first extracted from the time series. This process is referred to as feature extraction (FE) (Soleimani-Babakamali et al., Citation2022). Due to the dimensionality of high-frequency acceleration data, the selection of features to be extracted from raw time domain signals is central to success of data-driven SHM methods. For this purpose, feature extraction methods are normally employed to find useful lower dimension metrics from the raw time domain signals. The features extracted from acceleration signals are normally calculated over a set time window and provide a summary of the dynamic characteristics of the structure over that window. The change in these features over time is indicative of the behaviour of the dynamical system under measurement. However, most datasets contain irrelevant, highly correlated or noisy features that can be removed without a significant loss of information. This process is referred to as feature selection (FS) (Paniri et al., Citation2021).

FS is normally used in machine learning-based algorithms, especially when the learning task involves high-dimensional datasets. The primary purpose of FS is to choose a subset of available features, by eliminating features with little or no predictive information and also redundant features that are strongly correlated (Buckley et al., Citation2022; Paniri et al., Citation2021). In addition, the large volume of data represents a challenge to classification algorithms. This means that each feature used in the classification process should ideally provide an independent set of information. However, features are often highly correlated, and this can suggest a degree of redundancy in the available information which may have a negative impact on the classification accuracy (CA) (Pashaei & Pashaei, Citation2022a). Thus, FS approaches are needed to tackle these shortcomings.

The current methods for FS are generally divided into three categories: filter-based methods, wrapper-based methods and hybrid methods (Pashaei & Pashaei, Citation2022a). The filter-based methods use the statistical information of data to select features before the actual learning algorithm. These methods calculate the relevance of each feature with respect to the target classes. The features can then be sorted based on their individual relevance to the classes and then the top ranked features can be selected for modelling (Buckley et al., Citation2022). Examples of filter methods include analysis of variance (ANOVA) (Buckley et al., Citation2022), maximally relevant and minimally redundant (MRMR) (Zhao et al., Citation2019) and minimisation of joint mutual information (JMI) (Bennasar et al., Citation2015). The wrapper-based methods use the CA of the predetermined learning model, as a fitness function for the subset evaluation. The best feature set is the one which maximises the classification prediction accuracy (Pashaei & Pashaei, Citation2022b). The hybrid methods use independent measures to decide the best subsets for a given features and use a mining algorithm to select the final best subset between the available subsets (Cai et al., Citation2018). The FS algorithms have been reviewed in (Colaco et al., Citation2019).

When there is a large number of features, evaluating all states is computationally challenging and therefore metaheuristic search methods are required. Due to the inefficiency of the traditional search approaches in solving complex combinatorial optimization problems, several researchers have adopted metaheuristics algorithms (Xue et al., Citation2019). For instance, the Binary Coot bird optimization algorithm (BCOOT) (Pashaei & Pashaei, Citation2023) was developed as a wrapper feature selection method. Additionally, an enhanced version of the Black Hole Algorithm (BHA), namely hybrid dragonfly black hole algorithm was designed for real-world applications (Pashaei & Pashaei, Citation2021). Pashaei and Pashaei (Citation2022a) introduced an efficient Binary Chimp Optimization Algorithm (BChoA) that integrated the crossover operator to improve the ChOA’s exploratory behaviour. In another study, Pashaei and Pashaei (Citation2022b) developed the modified Binary Arithmetic Optimization Algorithm (BAOA) to evolve the usage of RSO for gene selection in high‑dimensional biomedical data. Moreover, the binary versions of the Rat Swarm Optimizer (BRSO) (Awadallah, Al-Betar, et al., Citation2022) and the Horse herd optimization (BHHO) (Awadallah, Hammouri, et al., Citation2022) were proposed to solve the feature selection problems as the wrapper methods. Recently, the non-dominated sorting genetic algorithm-III (NSGA-III) was developed for feature selection in databases with missing data (Xue et al., Citation2021).

The slime mould algorithm (SMA) (Li et al., Citation2020) is a novel and robust metaheuristic algorithm proposed to solve continuous problem and it is inspired by the propagation and foraging of the slime mould which includes a unique mathematical model. Feature selection is inherently a binary optimization problem (Ghiasi et al., Citation2021). The dimension of the problem is equal to the number of features and each solution vector represents the selection (1) or non-selection (0) of each feature. The binary version of the SMA (BSMA) proposed in (Abdollahzadeh et al., Citation2021) is used as the main optimization algorithm in this article. A comprehensive survey about SMA applications and its variants presented in (Soleimanian et al., Citation2023). Moreover, the robustness of using four variants of BSMA as FS algorithm is shown in (Abdel-Basset et al., Citation2021). Ghiasi and Malekjafarian (Citation2022) discussed that BSMA contains stagnation situation and population diversity in SHM applications which might reduce the efficiency of classification.

In this paper, an Advanced version of the Binary Slime Mould Algorithm (ABSMA) is introduced by incorporating two new operators of mutation and crossover into the BSMA. Mutation and crossover are mainly used as the key operators in genetic algorithm to make changes in the genes of the chromosomes (Sivanandam & Deepa, Citation2008). These operators have also been used in several optimization algorithms such as Whale Optimization Algorithm (WOA) (Qi et al., Citation2022) and Salp Swarm Algorithm (SSA) (Faris et al., Citation2018) to increase their efficiency. In this paper, for the first time, they are used in a combination in the BSMA. The main focus of this work is to identify the minimal set of features which maximises the ability of a learning model to detect and distinguish between damages in structural systems. For this purpose, a three-step framework is presented in this paper based on the proposed ABSMA. Firstly, statistical characteristics of structural response signals under ambient vibration are extracted, and feature vectors are obtained. Subsequently, the best feature subset is selected by the ABSMA algorithm based on desirability index using F-score (Kashef & Nezamabadi-Pour, Citation2015). In the final step, the selected feature subset is employed for training the classification model based on radial basis function Neural Network (NN). The performance of the proposed framework is evaluated statistically using a benchmark datasets of a timber bridge model (Kullaa, Citation2011) and a three-story frame structure (Figueiredo & Flynn, Citation2009; Ghiasi & Ghasemi, Citation2018). Furthermore, the efficiency of using ABSMA as the main algorithm for FS is compared to several state-of-the-art MOAs such as Binary Particle Swarm Optimization (BPSO) (Chuang et al., Citation2011), Binary Harris Hawks Optimization (BHHO) (Thaher et al., Citation2020), Binary Whale Optimization Algorithm (BWOA) (Qi et al., Citation2022) and Binary Farmland Fertility Optimization Algorithm (BFFA) (Naseri & Gharehchopogh, Citation2022). Moreover, the main part of the binary version of the SMA is a transfer function that is responsible to map a continuous search space to a discrete search space (Abdollahzadeh et al., Citation2021; Too et al., Citation2019). Therefore, in this paper, the impact of various transfer functions (such as S-shaped and V-shaped) on the accuracy of the proposed ABSMA is also assessed. The primary contributions of this study can be summarized as below:

  1. A novel algorithm called ABSMA is proposed by adding two operators of mutation and crossover to the BSMA algorithm, to overcome the stagnation situation involved in the original version of BSMA.

  2. The proposed ABSMA is embedded in a data-driven SHM framework to show its performance against other algorithms in the literature.

The rest of this article is organized as follows: SMA algorithm and its binary version are explained in Section 2. The proposed ABSMA is presented in Section 3. Details of the proposed data driven SHM framework using ABSMA are provided in Section 4. In Section 5, the proposed framework’s performance is examined using two real-data sets from the SHM community.

Theoretical background of SMA

Traditional SMA

The SMA is proposed by Li et al. (Citation2020) based on the oscillation mode of slime moulds in nature. The SMA has a unique mathematical model that uses adaptive weights to simulate the process of producing positive and negative feedback of the propagation wave of slime moulds based on bio-oscillator. It uses these features to form the optimal path for connecting food with excellent exploratory ability and exploitation propensity (Ghiasi et al., Citation2022). This is normally carried out within three phases: (1) Approach food, (2) Wrap food and (3) Grabble food. The logic of the SMA and each of these phases are shown in and each of these phases are explained here. More details of the SMA is provided by (Li et al., Citation2020).

Figure 1. The SMA framework (Li et al., Citation2020).

Figure 1. The SMA framework (Li et al., Citation2020).

Approach food

EquationEquation (1) represents the approaching behaviour of the slime mould to replicate the contraction mode (Li et al., Citation2020):

(1) Xt+1=Xbt+vbWXAtXBt,r<pvcXt,rp(1)

where W is the weight of the slime mould,vbis a parameter with a range of a,a, vc decreases linearly from 1 to 0, trepresents the current iteration, Xbrepresents the individual location with the highest odor concentration currently found, Xrepresents the location of slime mould, XA and XB represent two individuals randomly selected from the swarm, W represents the weight of the slime mould. The formula for pcan be given as:

(2) p=tanhSiDF(2)

where i1,2,,n (n=number of moulds), and Si represents the fitness of X. The best fitness acquired in all iterations is denoted by the DF. The vbformula is as follows:

(3) vb=a,a,(3)
(4) a=arctanhtmax_t+1,(4)

The formula of W is organized as follows:

(5) WSmellIndexi=1+rlogbFSibFwF+1,condition1rlogbFSibFwF+1,others(5)
(6) SmellIndex=sortS,(6)

where conditionrepresents that Si ranks first half of the population,r denotes the random value in the interval of 0,1bF and wF denote the optimal and worst fitness obtained in the current iterative process, respectively. SmellIndex represents the sequence of fitness values sorted (ascends in the minimum value problem).

Wrap food

This phase replicates the contraction mode of venous tissues of slime mode. EquationEquation (7) describes the updating position of slime mould:

(7) X=randUBLB+LB,rand<zXbt+vbWXAtXBt,r<pvcXt,rp(7)

where LB and UB represents the lower and upper boundaries of the searching range, randand rdenote the random value in [0,1].

Grabble food

As the number of iterations increases, the value of vboscillates randomly between a,a and gradually approaches zero. The value of vc oscillates between 1,1 and eventually tends to zero. The pseudo-code of the SMA is presented in Algorithm 1

Algorithm 1 Pseudo-code of SMA

Initialize the parameters popsize, Max_iteraition;

Initialize the positions of slime mould Xii=1,2,,n;

While (tMax_iteraition)

Calculate the fitness of all slime mould;

update bestFitness, Xb

Calculate the W by EquationEquation (5);

For each search portion

Update p,vb,vc;

Updated positions by EquationEquation (7);

End For

t=t+1;

End While

Return bestFitness, Xb;

Binary Slime Mould Algorithm (BSMA)

FS is generally an NP-hard combinatorial binary optimization problem, in which the number of possible solutions increases exponentially with the number of features. For example, if D is the total number of features, the number of possible solutions is 2D1 (Too et al., Citation2019). The BSMA was first proposed by Abdollahzadeh et al. (Citation2021) for solving binary optimization problems. They compared the effectiveness of BSMA with several binary metaheuristics such as Binary Harris Hawks Optimization (BHHO), Branch and Bound algorithm (BB), Binary Tunicate Awarm Algorithm (BTSA), Binary Farmland Fertility optimization Algorithm (BFFA), Binary Particle Swarm Optimization (BPSO), Binary Teaching – Learning-based Optimization (BTLBO), Binary Archimedes Optimization Algorithm (BAOA) and etc., and they concluded that BSMA is the most robust method compared to other algorithms. Therefore, BSMA was chosen as the main algorithm in this study. The meta-heuristic optimization algorithms (MOAs) normally start with the initialization step to spread the solutions within the search space of the problem. Accordingly, the BSMA is initialized by creating a population of n moulds. Each mould which represents a solution to the optimization process that has D dimensions equal to the number of features in the used dataset. The FS problem is considered as a discrete problem as it is based on choosing a number of features that leads to better accuracy in the classification method. Therefore, for each dimension, BSMA is randomly initialized with a value of 1 for the accepted feature or 0 for the rejected one as shown in . This provides the representation of an initial solution for the FS. At the end of each iteration, each mould has a solution in the form of a binary vector with the same length as the number of the features, where 1 means selecting and 0 means deselecting the corresponding feature. This process continues for all iterations and at last, the best feature subset with the least classification error of the classifier is suggested as the best result.

Figure 2. An initial solution to the FS.

Figure 2. An initial solution to the FS.

It should be noted that the values generated by the standard SMA are continuous, but the features in FS problems are binary: e.g. using 0 (selected feature) and 1 (not selected) values. Therefore, a transfer function is needed to convey solutions from the continuous space to the binary space. According to literature (Saremi et al., Citation2015), the using a transfer function is one of the effective ways to convert continuous optimizer into a binary one. In comparison with other operators, the transfer function is user-friendly and less computationally expensive (Saremi et al., Citation2015). A wide range of transfer functions belonging to the family of the V-Shaped and S-Shaped functions (Mirjalili & Lewis, Citation2013) can convert continuous values into binary ones. In this study, The V-Shaped and S-shaped transfer functions are used in this study are listed in . A transfer function receives a real value from the standard SMA as an input and then normalizes this value between 0 and 1 using one of the formulas in . The normalized value is then converted into a binary value using EquationEquation (8) (Abdollahzadeh et al., Citation2021).

(8) Sbinary=fx1,\ifSa>0.50,otherwise(8)

Table 1. V-shaped and S-shaped transfer function.

In EquationEquation (8), S(a) is the S-shaped transfer function.

The proposed advanced binary slime mould algorithm

To overcome the inefficiency of the BSMA in solving the feature selection problems in SHM domain, an advanced version of BSMA (Ghiasi & Malekjafarian, Citation2022) is proposed in this section. In the proposed version of BSMA, two ideas from GA (Sivanandam & Deepa, Citation2008) are implemented on the BSMA to enhance its capability for the FS and solve low population diversity and stagnation situation. The new solutions in GA are mainly created by the two operators: crossover and mutation. In the crossover operator, two solution sets are randomly selected, and some portions are exchanged which result in two new solutions. In the mutation operator, a randomly selected bit of a particular solution is mutated; means that 1 is changed to 0 and 0 is changed to 1. To implement these two operations on BSMA, a three-step procedure is developed as shown in . A random solution is generated in the first step, and then a crossover operation is applied to the randomly generated solution and the best available solution. In the second step, the solution obtained from the crossover operation is given as inputs to the mutation operation. Finally, if the new solution is better than the current one, the new solution replaces the current solution. The main purpose of these operations is to increase the population diversity and escaping from the local optimal points to improve the quality of the solutions. In other words, the integration of BSMA with both crossover and mutation operators simultaneously, improve both the exploration and exploitation capabilities of BSMA. Mutation operators improve the exploitation capability of the algorithm by searching around the best solution and crossover improves the exploration capability by searching around a slime mould created randomly. The pseudo-code of the ABSMA is presented in Algorithm 2.

Figure 3. The process of implementing the crossover and mutation on the solution vector of the ABSMA.

Figure 3. The process of implementing the crossover and mutation on the solution vector of the ABSMA.

Algorithm 2 Pseudo-code of ABSMA

Initialize the parameters popsize, Max_iteraition;

Initialize the positions of slime mould Xii=1,2,,n;

While (tMax_iteraition)

Calculate the Fitness of all slime mould;

UpdatebestFitness,Xb

Calculate the W by EquationEquation (5);

For each search portion

Updatep,vb,vc;

UpdatepositionsbyEquation7;

Apply one of the eight transfer-functions on slime mould;

Calculate the Fitness of slime mould;

UpdatebestFitness,Xb

Apply crossover and mutation operations on the current best slime mould (Xb)

If (Fitness of new slime mould> Fitness of current slime mould)

Replace new slime mould with current slime mould

End if

End For

t=t+1;

End While

Return bestFitness,Xb;

Fitness function

The fitness function (FF) plays an important role in the efficiency of the ABSMA algorithm as it is shown in Algorithm 2. As the developed framework in this study is based on a wrapper feature selection method (Pashaei & Pashaei, Citation2022a), the fitness function is developed based on the classification model accuracy and the efficiency of selected subset of features. The classification model accuracy is obtained by the evaluation of the test data classification using the trained model. In addition, the efficiency of the selected subset of features is evaluated using Fscore which measures the desirability of the features and will be defined in the next subsection. The ABSMA selects a vector with the smallest fitness value when the completion conditions are satisfied. The fitness function of the ABSMA is formed as follows:

(9) FF=1W×CA+1W×1ni=1nFscorei(9)

Where W is weighting factor between 0 and 1 and n is the total number of features. Fscorei is defined below. The CA is used to define the quality function of a solution, which is the percentage of samples correctly classified and evaluated as EquationEquation (10):

(10) CA=NumberofsamplescorrectlyclassifiedTotalnumberofsamplestakenforexprimentation(10)

F-score

A desirability value, for each feature generally represents the attractiveness of the features, and can be any subset evaluation function like an entropy-based measure or rough set dependency measure (Kashef & Nezamabadi-Pour, Citation2015). In this paper, Fscore is used as an index for measuring the desirability of the features. Fscore is a measurement to evaluate the discrimination ability of the feature i. EquationEquation (11) defines the F-score of theith feature. The numerator specifies the discrimination among the categories of the target variable, and the denominator indicates the discrimination within each category. A larger Fscore implies to a greater likelihood that this feature is discriminative (Kashef & Nezamabadi-Pour, Citation2015).

(11) Fscorei=k=1cxˉikxˉi2k=1c1Nik1j=1Nikxijkxˉik2(11)

where c is the number of classes and n is the number of features; Nik is the number of samples of the feature i in class k, (k=1,2,,c;i=1,2,,n), xijk is the j-th training sample for the feature i in class k, (j=1,2,,Nik), xi is the mean value of feature i of all classes and xik is the mean value of feature i of the samples in class k (Kashef & Nezamabadi-Pour, Citation2015).

It should be mentioned that the performance of the proposed algorithm is evaluated with the state-of-the-art metrics such as precision, recall, accuracy, F1-score and Feature-Reduction index (Fr). F1-score is a weighted average of the precision and recall. This score can be additionally weighted to account for class imbalance. The F1-score is calculated independently for each class and the weighted based on the number of true instances for each class. Fr which is used to compare the feature reduction rate in different algorithms, is defined as below:

(12) Fr=npn(12)

where n is the total number of features and p is the number of selected features by the FS algorithm. Fr is the average feature reduction. The more it is close to 1, the more features are reduced, and the classifier complexity is less.

The novel data-driven SHM using the proposed ABSMA

In this section, an SHM framework is presented using the optimal feature subset selection proposed in Section 3. The method consists of three main steps: (A) The Feature Extraction, (B) The FS using ABSMA and (C) The Feature Classification. The detail of FS using ABSMA is described in the previous section. The following subsections will show the detail of steps A and C. The detailed flowchart of the proposed three-stage framework is depicted in .

Figure 4. The detailed flowchart of proposed framework for feature selection and classification.

Figure 4. The detailed flowchart of proposed framework for feature selection and classification.

Feature extraction

In this paper, the functions given in , are used to form the feature vectors from the time domain sensors collected by sensors. These features are computed in time domain and provide a summary of the statistical characteristics of the signal over the feature extraction window. It should be mentioned that these features are selected based on the recommendations from the previous works in this field (Buckley et al., Citation2022; Ghiasi et al., Citation2021). These features represent the energy, the time series distribution and the vibration amplitude of the signals in time-domain (Buckley et al., Citation2022).

Table 2. Time-domain features.

Feature classification

Wrapper-based feature selection methods require a supervised learning approach where knowledge of the varying damage states or classes is available in order to identify the subset of informative features that best discriminate different classes (Pashaei & Pashaei, Citation2022a). Therefore, in this step, a well-trained classification model is applied to classify various conditions of the structure. In this model, the input matrix includes the selected features, and the outputs are the corresponding damage conditions. In recent years, many neural network models have been proposed or employed for various components of SHM in order to perform pattern classification, function approximation, and regression (Altabey et al., Citation2021). Among them, the RBF network is a type of feed forward neural networks that learns using a supervised training technique. Lowe and Broomhead (Citation1988) first exploited the use of the RBF for designing neural networks. Radial functions are a type of function in which the response reduces or grows monotonically with the distance from the centre point. The RBF network is a popular alternative to the well-known multilayer perceptron (MLP) since it has a simpler structure and a much faster training process (Wu et al., Citation2012). Therefore, RBF neural network is used as feature classifier in this paper.

Experimental results

In this section, two benchmark data set from the SHM community are selected to evaluate the effectiveness of the proposed FS algorithm. The first dataset includes acceleration responses of a timber bridge model recoded in the laboratory of Helsinki Polytechnic Stadia (Kullaa, Citation2011) and the second dataset contains the responses measured from a three-story frame scale structure published by Los Alamos National Laboratory (Figueiredo et al., Citation2011). outlines the datasets used in this work. All the analysis in this paper are done in MATLAB 9.13 using a computer with processing Intel Core i7–3340 3.1 GHz with 16 GB random access memory (RAM).

Table 3. The list of used datasets.

The timber bridge dataset

This data set was collected in the laboratory of Helsinki Polytechnic Stadia (Kullaa, Citation2011) and is available open access (Kullaa, Citation2018). The data was collected from a timber bridge model as shown in . In this experimental campaign, Kullaa (Citation2013) used a random excitation generated by an electrodynamic shaker to activate the vertical, transverse, and torsional modes. They measured the responses at three different longitudinal positions using 15 accelerometers. The sampling frequency was 256 Hz and the length of the signals was 32 s. The data were filtered using a low pass filter at 128 Hz and re-sampled for sufficient redundancy. The measurements were repeated several times and it was noticed that the dynamic properties of the structure vary due to the environmental changes. The main influencing factors were assumed to be the changes in the temperature and humidity (Kullaa, Citation2013).

Figure 5. The experimental case study; (a) the timber bridge model and (b) the locations of 15 sensors and the damage (D) are indicated (Kullaa, Citation2013).

Figure 5. The experimental case study; (a) the timber bridge model and (b) the locations of 15 sensors and the damage (D) are indicated (Kullaa, Citation2013).

Kullaa (Citation2013) modelled the damage by adding masses to the original structure. As described in the original paper (Kullaa, Citation2013), five artificial damage scenarios were introduced by adding small point masses of different sizes on the structure. The mass sizes were 23.5, 47.0, 70.5, 123.2 and 193.7 gr. The point masses were attached on the top flange, 600 mm left from the midspan (). The added masses were relatively small compared to the total mass of the bridge (36 kg), where the highest mass increase was only 0.5 %. The total number of experiments were carried out on the structure was 273. One hundred and ninety of the measurements were selected as the training data. The test data consisted of both the healthy and abnormal measurements.

Feature extraction

The first step in the framework proposed in this research is to extract statistical features from the recorded acceleration response of the structure. For this purpose, the statistical features shown in are extracted from the responses of 15 sensors on the timber bridge. The total number of extracted features for each experiment based on is 105 features (15 sensors × 7 features). To show the changes of features for different damage classes, eight samples were randomly selected from each damage class and the results are shown in . It can be seen in the figures that the RMS of sensor 7 has a specific threshold and boundary for different damage classes, but the Skewness of sensor 10 does not have such a specific threshold.

Figure 6. Features trend for various classes: (a) RMS of sensor 7 (b) Skewness of sensor 10.

Figure 6. Features trend for various classes: (a) RMS of sensor 7 (b) Skewness of sensor 10.

Feature selection

The automatic feature selection approach introduced in this research is then used to select the best subset of features. As mentioned in the objective function of ABSMA, the selected features should meet the following two conditions: the maximum distance between classes and the minimum distance within classes. Considering the high impact of the transfer function on the performance of the ABSMA, this function must be selected first, which will be discussed in the next subsection. For providing the stochastic behaviour of MOAs, the performance of the algorithms is compared using the best, worst, average and standard deviation (SD) of the obtained fitness values over 20 independent runs in . Columns ABSMA-V1, ABSMA-V2, ABSMA-V3, ABSMA-V4, ABSMA-S1, ABSMA-S2, ABSMA-S3, and ABSMA-S4 give the results of the transfer functions V1, V2, V3, V4, S1, S2, S3, and S4, respectively. As stated above, MOAs have stochastic nature and in each independent run, they may have slightly different results. Therefore, for comparing their performance, the approach used by other researchers (Varaee & Ghasemi, Citation2017) (considering the best, worst, average, and standard deviation of fitness values) is employed here. The results of this analysis are given in , where ABSMA-V2 shows the best performance in most indexes (best, average, and worst) in comparison with other transfer functions. Therefore, V2 is selected as the transfer function in this study. For simplicity, ABSMA-V2 will be denoted as ABSMA in the rest of the article.

Table 4. The best fitness values under eight different transfer functions.

Feature classification

In this section, the accuracy and effectiveness of the proposed framework for feature extraction/selection in SHM domain is evaluated. Furthermore, the results obtained by the proposed ABSMA algorithm are compared to BSMA, BPSO, BHHO, BWOA, and BFFA which are reported to be effective algorithms in FS (Abdollahzadeh et al., Citation2021). The parameters that need to be set in these algorithms are set to the best values are reported in the original studies and are shown in . In order to maintain a fair comparison, the population size for all the algorithms is set to be 50 and the maximum iterations is set to be 200. The weighting factor W in the fitness function is varied from 0.6 to 0.9 to get the different sets of features. The results are averaged over 20 independent runs in every algorithm. The dimension of the search space is equal to the total number of features of each dataset.

Table 5. Parameter settings for the comparative algorithms.

gives the mean of the CA, best, worst, average and SD of the results for each algorithm. The number in the brackets in each table slot shows the ranking of each algorithm and the best result is highlighted with the bold text. It can be seen that the ABSMA scored the best fitness value, followed by BSMA. From , the best algorithm that contributed to the lowest average fitness value was found to be ABSMA, followed by BSMA and BHHO. On the one hand, BSMA perceived the most consistent results due to the lowest STD values in this dataset.

Table 6. CA of each algorithm for the tested datasets of timber bridge.

A comparison of the average precision, recall, F1 score and the amount of Fr for other algorithms are given in . It can be concluded that the proposed ABSMA algorithm can obtain, in most of cases, better CA using a smaller feature set, compared to other algorithms. On the one hand, even though BFFA and BPSO algorithms can reduce the number of features; however, the relevant features are eliminated and thus resulting in unsatisfactory performance.

Table 7. Comparison of the performance (precision, recall, F1-score and Fr) of the algorithms on timber bridge.

The number of selected features and the average of Fr for each optimization algorithm are shown in . It can be seen that the ABSMA not only finds smaller feature subsets than the other algorithms but also the number of selected features also decreases much faster. It can be concluded that the ABSMA provides a higher degree of exploration than the other algorithms, which enables it to explore the search space to find a solution that selects a smaller number of features and better performance.

Figure 7. Number of selected features of each optimization algorithms.

Figure 7. Number of selected features of each optimization algorithms.

Figure 8. Average of Fr for each optimization algorithms with respect to number of iterations.

Figure 8. Average of Fr for each optimization algorithms with respect to number of iterations.

Among the 20 independent runs in Section 5.1.3, the highest overall performance of the ABSMA is achieved in 19th and the worst in 13th run (From 20 independent runs). shows a confusion matrix for the 13th run of the ABSMA, the worst performing of it. The majority of misclassifications are occurring between successive damage classes with values being misclassified as the previous damage state. The separation between the healthy and damaged states and the reduction of false alarms is critical for SHM applications (Buckley et al., Citation2022). For the healthy state, 83.3% of the unseen healthy data is being correctly predicted and the False Negative Rate (FNR) for the Healthy state is 16%. Therefore, despite the poor overall classification prediction for this run, the coupled system of the ABSMA and NN have reasonable accuracy in distinguishing between the healthy state and the damaged states.

Figure 9. Confusion matrix for worst result of ABSMA (13th run).

Figure 9. Confusion matrix for worst result of ABSMA (13th run).

shows a confusion matrix for 19th run, which has the best classification performance of ABSMA. In this run, only 2 samples are misclassified. One of the concerns with an imbalanced dataset is that a classifier may learn to improve prediction performance by randomly assigning datapoints to majority classes (Krawczyk, Citation2016). The confusion matrices show that this is not the case as the majority of miss classifications across the runs are when the unseen test data is at boundaries between classes particularly between the damage class 2 and 3.

Figure 10. Confusion matrix for best result of ABSMA (19th run).

Figure 10. Confusion matrix for best result of ABSMA (19th run).

In order to confirm the efficiency of the proposed feature selection framework, the CA of the NN for the selected feature subsets is compared with the one for all features in . The results show a higher CA value for the reduced number of features. In RBF case, the accuracy is increased from 87% to 94% with 81% data reduction. This result is reasonable because the main benefit of FS is to improve prediction performance and provide faster and more cost-effective predictors. Using too many features degrades prediction performance even when all features are relevant and contain information about the response variable.

Table 8. Comparing the performance of RBF neural network for selected features and all features.

The three-story frame structure dataset

The experimental dataset of a three-story frame structure published by Los Alamos National Laboratory is used here as the second case study (Figueiredo et al., Citation2011). shows the three-story frame structure where an electrodynamic shaker was used to excite the frame structure with various damage conditions with Gaussian white noise laterally on the base floor along the structural centreline. The excitation force applied from the shaker to the structure was recorded with a load cell mounted on the stringer and the structural responses were measured using four accelerometers attached at the centre line of each floor as shown in . The data were collected and processed at a sampling frequency of 320 Hz with a data acquisition system. For each structural damage state, 10 shaking tests were conducted considering the variability of excitations and structural properties (Figueiredo et al., Citation2011).

Figure 11. The three-story frame structure. (a) Experimental setup (b) acceleration sensor positions (Figueiredo & Flynn, Citation2009).

Figure 11. The three-story frame structure. (a) Experimental setup (b) acceleration sensor positions (Figueiredo & Flynn, Citation2009).

The main goal of this benchmark study is to detect damage when the structure has undergone structural changes caused by operational and environmental effects (Figueiredo et al., Citation2011). For this purpose, the present study selects four structural conditions of the three-story building from the database available for open access (Figueiredo, Citation2007) to examine the effectiveness of proposed FS in damage localization. The damage cases were simulated through the introduction of nonlinearities into the structure. A bumper and a suspended column were used with different gaps in between them as shown in . The gap between the bumper and the suspended column was varied (0.1 and 0.20 mm) to introduce different degrees of nonlinearity.

Figure 12. A adjustable bumper and the suspended column (Figueiredo & Flynn, Citation2009).

Figure 12. A adjustable bumper and the suspended column (Figueiredo & Flynn, Citation2009).

The selected conditions include the baseline condition without structural damage (termed as D0), structural condition with Gap equal to 0.20 mm and mass on the 1st floor as representer of operational and environmental condition changes (D1), structural condition Gap equal to 0.10 mm and mass on the 1st floor (D2), and structural condition with Gap equal to 0.20 mm and mass on the base floor (D3). summarizes all structural conditions investigated in this paper.

Table 9. The structural conditions of the three-story frame structure dataset.

Feature selection

The ABSMA proposed in this study is used for FS using this data set. gives the CA of ABSMA compared to other methods. Like the first dataset, each feature selection algorithm is executed for 20 runs, with different random seed. The averaged results of 20 runs are used for performance comparison.

Table 10. CA of each algorithm for the tested datasets of three-story frame.

shows that the best mean CA is obtained by ABSMA (90%), followed by BFFA (85%). In comparison with original versions of BSMA, ABSMA has a higher chance to prevent itself from being trapped in the local optimum.

In addition, illustrates the mean, best and STD computational time of the proposed methods over 20 independent runs. In addition, ABSMA shows the fastest processing speed in this work. This indicates that ABSMA can achieve the optimal feature subset in a very short period. The reason ABSMA has a very short computational time is because it utilizes a mutation and crossover strategies together, which performs a position update for best slime mould. It can be concluded that ABSMA not only provides a great performance in feature selection but also provides the lowest computational cost.

Table 11. Computational time of each algorithm for the tested datasets of three-story frame.

Moreover, shows the number of selected features for proposed method in comparison with other MOAs. It is observed that not all the features are required in the classification process. A proper selection of features could lead to a higher classification performance with lower complexity. As presented in , ABSMA contributes to the smallest number of the features in comparison with other FS wrapper-based algorithm. This means that ABSMA can achieve a promising CA while keeping a smaller number of features. On one side, BFFA has a higher mean number of selected features, 18. It can be inferred that BFFA does not evaluate the relevant features very well, thus leading to a poor classification performance in this work. Finally, according to the results shown, adding desirability index, mutation and crossover operators to the BSMA, increases the exploration of the search and guide the algorithm to more salient features.

Table 12. Number of selected features in each algorithm for the tested datasets of three-story frame.

demonstrates the convergence curve of the proposed methods on Los Alamos dataset which shows the fitness value in each iteration for different algorithms. It can be seen that ABSMA provides the lowest fitness value compared to others which means it has a good diversity which gives the ability to escape from the local optimum. Unlike BPSO and BHHO, ABSMA keeps tracking for the global optimum, thus leading to a very good performance. On the other side, BPSO and BHHO converged faster, but without acceleration. This shows that BPSO and BHHO are easily getting trapped in the local optimum. It can be concluded that ABSMA is effective and reliable in evaluating the optimal feature subset.

Figure 13. The convergence curve of six different feature selection methods for Los Alamos dataset.

Figure 13. The convergence curve of six different feature selection methods for Los Alamos dataset.

In order to investigate the effect of using the two operators on the efficiency of ABSMA, the convergence history of the algorithm in cases where only one operator is implemented on the ABMSA is compared with the case where both of them are employed in . It can be seen that the separate implementation of mutation and combination operators does not significantly increase the efficiency of ABSMA. But, if they are used simultaneously, the algorithm could have the ability to escape from local optima and reach the best subset of features.

Figure 14. The convergence curve of implementation of mutation and cross over operator on ABSMA based on Los Alamos dataset.

Figure 14. The convergence curve of implementation of mutation and cross over operator on ABSMA based on Los Alamos dataset.

Furthermore, the Wilcoxon signed-rank test and t-test are employed to compare the performance of the proposed algorithm with other optimization algorithms in terms of CA and the number of selected features. In the Wilcoxon signed-rank test, if the p-value achieved is less than 0.05, then the performances of the two algorithms are significantly different; otherwise, the performances of the two algorithms are similar (Too et al., Citation2019). exhibit the result of Wilcoxon signed-rank test and t-test with p-values. By conducting these tests, it is seen that there is a significant difference in classification performance and number of features between ABSMA versus BPSO (p = 6.4×103, 1.4×103), ABSMA versus BWOA (p = 8.5×103, 8.82×105), ABMSA versus BHHO (p = 4.8×104, 7.72×104), ABMSA versus BFFA (p = 7.8×103, 9.97×105), and ABSMA versus BSMA (p = 2.64×103, 2.1×104). The statistical results show the superiority of ABMSA over other algorithms in feature selection.

Table 13. P-values of Wilcoxon signed-rank test.

Table 14. P-values of Wilcoxon signed-rank test.

In order to present the underlying relationships between the features and structural damages, the features are plotted using data scatter diagrams at different damage classes in . shows two features randomly selected from the set of features, and shows two features selected by ABMSA. It can be concluded that by minimizing the objective function, ABMSA selects the features that have the highest ability to differentiate between the different damage classes.

Figure 15. Scatter diagram of features for various damage classes. (a) Shape factor of sensor 1 vs mean of sensor 4 (b) Std of sensor 1 vs Skewness of sensor 4.

Figure 15. Scatter diagram of features for various damage classes. (a) Shape factor of sensor 1 vs mean of sensor 4 (b) Std of sensor 1 vs Skewness of sensor 4.

Feature classification

In this subsection, the performance of the NN as a classification algorithm is compared with 5 other commonly used machine learning (ML) classifiers: Random Forest (RF) (Ghiasi et al., Citation2018), k-nearest neighbor (KNN) algorithm (with Euclidean distance and k = 5) (Too et al., Citation2019), Support Vector Machine (SVM) (with radial basis kernel function) (Santos et al., Citation2016), Decision Tree (DT) (Charbuty & Abdulazeez, Citation2021), and Cascade Forward Neural Network (CFNN) (with 2 hidden layer) (Fathnejat et al., Citation2014). shows the boxplots of CA for different ML algorithms on Los Alamos datasets. In these figures, the red line in the box represents the median value, and the symbol “+” denotes the outlier. As can be seen, ABSMA showed competitive median value in most cases. Furthermore, in comparison between the ML algorithm, NN and CFNN provides better classification performance than KNN, SVM, RF and DT algorithms.

Figure 16. Boxplot for CA of six different classifiers algorithms on Los Alamos dataset. (a) KNN (b) SVM (c) RF (d) DT (e) NN (f) CFNN.

Figure 16. Boxplot for CA of six different classifiers algorithms on Los Alamos dataset. (a) KNN (b) SVM (c) RF (d) DT (e) NN (f) CFNN.

In addition, to compare the performance of proposed wrapper-based approach, 5 well-known filter-based methods are selected from the literature. These methods are as follows: Principal Component Analysis (PCA) (Santos et al., Citation2016), Neighborhood Component Analysis (NBC) (Malan & Sharma, Citation2019), Term Variance (TV) (Malan & Sharma, Citation2019), Pearson Correlation Coefficient (PCC) (Saidi et al., Citation2019) and Relief-F (Urbanowicz et al., Citation2018). shows a comparison between performance of selected filter-based approaches with the proposed framework.

Table 15. CA of five different filter-based feature selection method and proposed approach.

Generally, in comparison with the filter based models, the wrapper model achieve a higher CA and tend to have a smaller subset size; however, it has high time complexity (Kashef & Nezamabadi-Pour, Citation2015).

Finally, in order to compare the performance of the proposed framework on damage detection of the Los Almos dataset, the CA of similar ML algorithms from other paper (He et al., Citation2022) has been chosen and is shown in . As can be seen in , by using ABSMA, the SHM framework select more salient feature that enhances the capability of classifier in detecting damage class.

Table 16. CA of proposed framework for classification of damage state in comparison with other framework.

In practice, users might have difficulty in selecting the best features for each SHM problem. Unlike other traditional feature selection methods, users can apply the ABSMA to select the potential features without prior knowledge. Successively, ABSMA will automatically select the optimal features for specific subjects, and that feature subset will be used in real world application. This, in turn, will reduce the complexity and improve the performance of the damage detection system. In sum, the proposed ABSMA is useful in feature selection.

Conclusion

In this paper, a new algorithm called ABSMA is proposed for FS in SHM problems. ABSMA is proposed for enhancing the capability of the SMA in this domain. The mutation and crossover operators are employed in the proposed ABSMA which could increase diversity, prevent excessive convergence during the optimization process, and local optimal trap escape. Two benchmark data sets selected from the SHM community are employed in this paper. The ABSMA is initially evaluated using eight transfer functions that convert continuous solutions to binary ones, in which the best transfer function (transfer function V2) is selected. The results obtained from the proposed algorithm are compared with 4 state-of-the-art metaheuristic-based algorithms including BHHO, BPSO, BWOA and BFFA. The results of the experiments indicate that a significant improvement in the proposed algorithm compared to other ones. Moreover, the proposed framework can remove the irrelevant and redundant information by choosing useful features as the input of the classification model. It is also shown that the proposed FS approach based on the ABSMA optimization algorithm reaches a better feature set in terms of CA in comparison with the full feature set. In addition, it can be concluded that ABMSA not only yields the optimal classification performance but also provides the minimal feature size and consumes a very low computational cost. Finally, the experimental results show that NN and CFNN can usually achieve the highest CA in comparison with KNN, SVM, RF and DT.

The features extracted in time domain are used in this paper to identify the state of the structure. However, using the extracted features in frequency domain and comparing their performance in detecting the state of the structure can be considered as a future extension of current work. Furthermore, the supervised scheme is used for training/testing of ML algorithms in the proposed framework, while in some real-life cases, there is a limited access to labelled data. In such cases, unsupervised learning schemes should be used. Moreover, it is suggested to use a chaotic map to fine-tune the parameters of ABSMA in future works. The base code and extracted feature data have been made available at https://github.com/raminqs/ABSMA.git.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

This publication has emanated from research conducted with the financial support of Science Foundation Ireland under Grant number [20/FFP-P/8706].

References

  • Abdel-Basset, M., Mohamed, R., Chakrabortty, R. K., Ryan, M. J., & Mirjalili, S. (2021). An efficient binary slime mould algorithm integrated with a novel attacking-feeding strategy for feature selection. Computers & Industrial Engineering, 153(January), 107078. https://doi.org/10.1016/j.cie.2020.107078
  • Abdollahzadeh, B., Barshandeh, S., Javadi, H., & Epicoco, N. (2021). An enhanced binary slime mould algorithm for solving the 0–1 knapsack problem. Engineering with Computers, 38(s4), 3423–3444. https://doi.org/10.1007/s00366-021-01470-z
  • Altabey, W. A., Noori, M., Wang, T., Ghiasi, R., Kuok, S.-C., & Wu, Z. (2021). Deep learning-based crack identification for steel pipelines by extracting features from 3d shadow modeling. Applied Sciences (Switzerland), 11(13), 6063. https://doi.org/10.3390/app11136063
  • Avci, O., Abdeljaber, O., Kiranyaz, S., Hussein, M., Gabbouj, M., & Inman, D. J. (2021). A review of vibration-based damage detection in civil structures: From traditional methods to machine learning and deep learning applications. Mechanical Systems and Signal Processing, 147, 107077. https://doi.org/10.1016/j.ymssp.2020.107077
  • Awadallah, M. A., Al-Betar, M. A., Braik, M. S., Hammouri, A. I., Doush, I. A., & Zitar, R. A. (2022). An enhanced binary rat swarm optimizer based on local-best concepts of PSO and collaborative crossover operators for feature selection. Computers in Biology and Medicine, 147, 105675. https://doi.org/10.1016/j.compbiomed.2022.105675
  • Awadallah, M. A., Hammouri, A. I., Al-Betar, M. A., Braik, M. S., & Abd Elaziz, M. (2022). Binary Horse herd optimization algorithm with crossover operators for feature selection. Computers in Biology and Medicine, 141, 105152. https://doi.org/10.1016/j.compbiomed.2021.105152
  • Bennasar, M., Hicks, Y., & Setchi, R. (2015). Feature selection using joint mutual information maximisation. Expert Systems with Applications, 42(22), 8520–8532. https://doi.org/10.1016/j.eswa.2015.07.007
  • Buckley, T., Ghosh, B., & Pakrashi, V. (2022). A feature extraction & selection benchmark for structural health monitoring. Structural Health Monitoring, 14. https://doi.org/10.1177/14759217221111141
  • Cai, J., Luo, J., Wang, S., & Yang, S. (2018). Feature selection in machine learning: A new perspective. Neurocomputing, 300, 70–79. https://doi.org/10.1016/j.neucom.2017.11.077
  • Charbuty, B., & Abdulazeez, A. (2021). Classification based on decision tree algorithm for machine learning. Journal of Applied Science and Technology Trends, 2(1), 20–28.
  • Chuang, L.-Y., Yang, C.-H., & Li, J.-C. (2011). Chaotic maps based on binary particle swarm optimization for feature selection. Applied Soft Computing, 11(1), 239–248. https://doi.org/10.1016/j.asoc.2009.11.014
  • Colaco, S., Kumar, S., Tamang, A., & Biju, V. G. (2019). A review on feature selection algorithms. In N. Shetty, L. Patnaik, H. Nagaraj, P. Hamsavath, & N. Nalini (Eds.), Emerging Research in Computing, Information, Communication and Applications. Advances in Intelligent Systems and Computing (Vol. 906). Springer. https://doi.org/10.1007/978-981-13-6001-5_11.
  • Corbally, R., & Malekjafarian, A. (2022). A data-driven approach for drive-by damage detection in bridges considering the influence of temperature change. Engineering Structures, 253(December 2021), 113783. https://doi.org/10.1016/j.engstruct.2021.113783
  • Dadras Eslamlou, A., & Huang, S. (2022). Artificial-neural-network-based surrogate models for structural health monitoring of civil structures: A literature review. Buildings, 12(12), 2067. https://doi.org/10.3390/buildings12122067
  • Das, S., Saha, P., & Patro, S. K. (2016). Vibration-based damage detection techniques used for health monitoring of structures: A review. Journal of Civil Structural Health Monitoring, 6(3), 477–507. https://doi.org/10.1007/s13349-016-0168-5
  • Faris, H., Mafarja, M. M., Heidari, A. A., Aljarah, I., Ala’M, A.-Z., Mirjalili, S., & Fujita, H. (2018). An efficient binary salp swarm algorithm with crossover scheme for feature selection problems. Knowledge-Based Systems, 154, 43–67. https://doi.org/10.1016/j.knosys.2018.05.009
  • Fathnejat, H., Torkzadeh, P., Salajegheh, E., & Ghiasi, R. (2014). Structural damage detection by model updating method based on cascade feed-forward neural network as an efficient approximation mechanism. International Journal of Optimization in Civil Engineering, 4(4), 451–472. http://ijoce.iust.ac.ir/browse.php?a_code=A-10-66-43&slc_lang=en&sid=1
  • Figueiredo, E. (2007). Los Alamos National Laboratory dataset. https://www.lanl.gov/projects/national-security-education-center/engineering/software/shm-data-sets-and-software.php
  • Figueiredo, E., & Flynn, E. (2009). Three-story building structure to detect nonlinear effects. Report SHMTools Data Description.
  • Figueiredo, E., Park, G., Farrar, C. R., Worden, K., & Figueiras, J. (2011). Machine learning algorithms for damage detection under operational and environmental variability. Structural Health Monitoring, 10(6), 559–572. https://doi.org/10.1177/1475921710388971
  • Gharehbaghi, V. R., Noroozinejad Farsangi, E., Noori, M., Yang, T. Y., Li, S., Nguyen, A., Málaga-Chuquitaype, C., Gardoni, P., & Mirjalili, S. (2021). A critical review on structural health monitoring: Definitions, methods, and perspectives. Archives of Computational Methods in Engineering, 1–27.
  • Ghiasi, R., & Ghasemi, M. R. (2018). An intelligent health monitoring method for processing data collected from the sensor network of structure. Steel and Composite Structures, 29(6), 703–716. https://doi.org/10.12989/scs.2018.29.6.703
  • Ghiasi, R., Ghasemi, M. R., & Chan, T. H. T. (2021). Optimum feature selection for SHM of benchmark structures using efficient AI mechanism. Smart Structures and Systems, 27(4), 623–640. https://doi.org/10.12989/sss.2021.27.4.623
  • Ghiasi, R., Ghasemi, M. R., & Noori, M. (2018). Comparative studies of metamodeling and AI-based techniques in damage detection of structures. Advances in Engineering Software, 125, 101–112. https://doi.org/10.1016/j.advengsoft.2018.02.006
  • Ghiasi, R., & Malekjafarian, A. (2022, August). An advanced binary slime mould algorithm for feature subset selection in structural health monitoring data. Civil Engineering Research in Ireland 2022 (CERI2022).
  • Ghiasi, R., Noori, M., Kuok, S.-C., Silik, A., Wang, T., Pozo, F., & Altabey, W. A. (2022). Structural assessment under uncertain parameters via the interval optimization method using the slime mold algorithm. Applied Sciences, 12(4), 1876. https://doi.org/10.3390/app12041876
  • Ghiasi, R., Torkzadeh, P., & Noori, M. (2016). A machine-learning approach for structural damage detection using least square support vector machine based on a new combinational kernel function. Structural Health Monitoring, 15(3), 302–316. https://doi.org/10.1177/1475921716639587
  • Gomes, G. F., Mendez, Y. A. D., Alexandrino, P. D. S. L., da Cunha, S. S., & Ancelotti, A. C. (2018). A review of vibration based inverse methods for damage detection and identification in mechanical structures using optimization algorithms and ANN. Archives of Computational Methods in Engineering, 26(4), 883–897. https://doi.org/10.1007/s11831-018-9273-4
  • He, Y., Huang, Z., Liu, D., Zhang, L., & Liu, Y. (2022). A novel structural damage identification method using a hybrid deep learning framework. Buildings, 12(12), 2130. https://doi.org/10.3390/buildings12122130
  • Kashef, S., & Nezamabadi-Pour, H. (2015). An advanced ACO algorithm for feature subset selection. Neurocomputing, 147, 271–279. https://doi.org/10.1016/j.neucom.2014.06.067
  • Krawczyk, B. (2016). Learning from imbalanced data: Open challenges and future directions. Progress in Artificial Intelligence, 5(4), 221–232. https://doi.org/10.1007/s13748-016-0094-0
  • Kullaa, J. (2011). Distinguishing between sensor fault, structural damage, and environmental or operational effects in structural health monitoring. Mechanical Systems and Signal Processing, 25(8), 2976–2989. https://doi.org/10.1016/j.ymssp.2011.05.017
  • Kullaa, J. (2013). Detection, identification, and quantification of sensor fault in a sensor network. Mechanical Systems and Signal Processing, 40(1), 208–221. https://doi.org/10.1016/j.ymssp.2013.05.007
  • Kullaa, J. (2018). Wooden bridge data. https://users.metropolia.fi/~kullj/JrkwXyZGkhc/
  • Li, S., Chen, H., Wang, M., Asghar, A., & Mirjalili, S. (2020). Slime mould algorithm: A new method for stochastic optimization. Future Generation Computer Systems, 111, 300–323. https://doi.org/10.1016/j.future.2020.03.055
  • Lowe, D., & Broomhead, D. (1988). Multivariable functional interpolation and adaptive networks. Complex Systems, 2(3), 321–355. https://www.complex-systems.com/abstracts/v02_i03_a05/
  • Malan, N. S., & Sharma, S. (2019). Feature selection using regularized neighbourhood component analysis to enhance the classification performance of motor imagery signals. Computers in Biology and Medicine, 107, 118–126. https://doi.org/10.1016/j.compbiomed.2019.02.009
  • Malekjafarian, A., Golpayegani, F., Moloney, C., & Clarke, S. (2019). A machine learning approach to bridge-damage detection using responses measured on a passing vehicle. Sensors, 19(4035), 4035. https://doi.org/10.3390/s19184035
  • Malekjafarian, A., OBrien, E. J., Quirke, P., Cantero, D., & Golpayegani, F. (2021). Railway track loss-of-stiffness detection using bogie filtered displacement data measured on a passing train. Infrastructures, 6(6), 93. https://doi.org/10.3390/infrastructures6060093
  • Mirjalili, S., & Lewis, A. (2013). S-shaped versus V-shaped transfer functions for binary particle swarm optimization. Swarm and Evolutionary Computation, 9, 1–14. https://doi.org/10.1016/j.swevo.2012.09.002
  • Naseri, T. S., & Gharehchopogh, F. S. (2022). A feature selection based on the farmland fertility algorithm for improved intrusion detection systems. Journal of Network and Systems Management, 30(3), 40. https://doi.org/10.1007/s10922-022-09653-9
  • Paniri, M., Dowlatshahi, M. B., & Nezamabadi-Pour, H. (2021). Ant-TD: Ant colony optimization plus temporal difference reinforcement learning for multi-label feature selection. Swarm and Evolutionary Computation, 64, 100892. https://doi.org/10.1016/j.swevo.2021.100892
  • Pashaei, E., & Pashaei, E. (2021). Gene selection using hybrid dragonfly black hole algorithm: A case study on RNA-seq COVID-19 data. Analytical Biochemistry, 627, 114242. https://doi.org/10.1016/j.ab.2021.114242
  • Pashaei, E., & Pashaei, E. (2022a). An efficient binary chimp optimization algorithm for feature selection in biomedical data classification. Neural Computing and Applications, 34(8), 6427–6451. https://doi.org/10.1007/s00521-021-06775-0
  • Pashaei, E., & Pashaei, E. (2022b). Hybrid binary arithmetic optimization algorithm with simulated annealing for feature selection in high ‑ dimensional biomedical data. The Journal of Supercomputing, 78(13), 15598–15637. https://doi.org/10.1007/s11227-022-04507-2
  • Pashaei, E., & Pashaei, E. (2023). Hybrid binary COOT algorithm with simulated annealing for feature selection in high-dimensional microarray data. Neural Computing and Applications, 35(1), 353–374. https://doi.org/10.1007/s00521-022-07780-7
  • Qi, A., Zhao, D., Yu, F., Heidari, A. A., Chen, H., & Xiao, L. (2022). Directional mutation and crossover for immature performance of whale algorithm with application to engineering optimization. Journal of Computational Design and Engineering, 9(2), 519–563. https://doi.org/10.1093/jcde/qwac014
  • Saidi, R., Bouaguel, W., & Essoussi, N. (2019). Hybrid feature selection method based on the genetic algorithm and pearson correlation coefficient. In A. Hassanien (Ed.), Machine Learning Paradigms: Theory and Application. Studies in Computational Intelligence (Vol. 801). Springer. https://doi.org/10.1007/978-3-030-02357-7_1
  • Santos, A., Figueiredo, E., Silva, M. F. M., Sales, C. S., & Costa, J. C. W. A. C. W. A. (2016). Machine learning algorithms for damage detection: Kernel-based approaches. Journal of Sound and Vibration, 363, 584–599. https://doi.org/10.1016/j.jsv.2015.11.008
  • Saremi, S., Mirjalili, S., & Lewis, A. (2015). How important is a transfer function in discrete heuristic algorithms. Neural Computing and Applications, 26(3), 625–640. https://doi.org/10.1007/s00521-014-1743-5
  • Silik, A., Noori, M., Altabey, W. A., Dang, J., Ghiasi, R., & Wu, Z. (2022). Optimum wavelet selection for nonparametric analysis toward structural health monitoring for processing big data from sensor network: A comparative study. Structural Health Monitoring, 21(3), 803–825. https://doi.org/10.1177/14759217211010261
  • Silik, A., Noori, M., Altabey, W. A., & Ghiasi, R. (2021). Selecting optimum levels of wavelet multi-resolution analysis for time-varying signals in structural health monitoring. Structural Control and Health Monitoring, 28(8). https://doi.org/10.1002/stc.2762
  • Sivanandam, S. N., & Deepa, S. N. (2008). Genetic algorithms. In Introduction to genetic algorithms (pp. 15–37). Springer Berlin Heidelberg. https://doi.org/10.1007/978-3-540-73190-0_2
  • Soleimani-Babakamali, M. H., Sepasdar, R., Nasrollahzadeh, K., Lourentzou, I., & Sarlo, R. (2022). Toward a general unsupervised novelty detection framework in structural health monitoring. Computer-Aided Civil and Infrastructure Engineering, 37(9), 1128–1145. https://doi.org/10.1111/mice.12812
  • Soleimanian, F., Alaettin, G., Turgay, U., Bahman, I., & Gultekin, A. (2023). Slime mould algorithm: A comprehensive survey of its variants and applications. Archives of Computational Methods in Engineering. https://doi.org/10.1007/s11831-023-09883-3
  • Thaher, T., Heidari, A. A., Mafarja, M., Dong, J. S., & Mirjalili, S. (2020). Binary harris hawks optimizer for high-dimensional, low sample size feature selection. In S. Mirjalili, H. Faris, & I. Aljarah (Eds.), Evolutionary Machine Learning Techniques. Algorithms for Intelligent Systems. Springer. https://doi.org/10.1007/978-981-32-9990-0_12
  • Too, J., Abdullah, A. R., & Saad, N. M. (2019). A new quadratic binary harris hawk optimization for feature selection. Electronics, 8(10), 1–27. https://doi.org/10.3390/electronics8101130
  • Urbanowicz, R. J., Meeker, M., La Cava, W., Olson, R. S., & Moore, J. H. (2018). Relief-based feature selection: Introduction and review. Journal of Biomedical Informatics, 85, 189–203. https://doi.org/10.1016/j.jbi.2018.07.014
  • Varaee, H., & Ghasemi, M. R. (2017). Engineering optimization based on ideal gas molecular movement algorithm. Engineering with Computers, 33(1), 71–93. https://doi.org/10.1007/s00366-016-0457-y
  • Wu, Y., Wang, H., Zhang, B., & Du, K.-L. (2012). Using radial basis function networks for function approximation and classification. International Scholarly Research Notices, 2012, 1–34. https://doi.org/10.5402/2012/324194
  • Xue, Y., Tang, Y., Xu, X., Liang, J., & Neri, F. (2021). Multi-objective feature selection with missing data in classification. IEEE Transactions on Emerging Topics in Computational Intelligence, 6(2), 355–364. https://doi.org/10.1109/TETCI.2021.3074147
  • Xue, Y., Xue, B., & Zhang, M. (2019). Self-adaptive particle swarm optimization for large-scale feature selection in classification. ACM Transactions on Knowledge Discovery from Data (TKDD), 13(5), 1–27. https://doi.org/10.1145/3340848
  • Zhao, Z., Anand, R., & Wang, M. (2019). Maximum relevance and minimum redundancy feature selection methods for a marketing machine learning platform. In 2019 IEEE International Conference on Data Science and Advanced Analytics (DSAA) (pp. 442–452). IEEE. https://doi.org/10.1109/DSAA.2019.00059