Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal / Special Issue
Modeling of the Human Cardiovascular System: Implementing a Sliding Mode Observer for Fault Detection and Isolation
Previous Article in Journal
Improved Mechanical Characterization of Soft Tissues Including Mounting Stretches
Previous Article in Special Issue
H State and Parameter Estimation for Lipschitz Nonlinear Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Computational Cost Reduction in Multi-Objective Feature Selection Using Permutational-Based Differential Evolution

by
Jesús-Arnulfo Barradas-Palmeros
1,*,
Efrén Mezura-Montes
1,
Rafael Rivera-López
2,
Hector-Gabriel Acosta-Mesa
1 and
Aldo Márquez-Grajales
3
1
Artificial Intelligence Research Institute, Universidad Veracruzana, Xalapa 91097, Veracruz, Mexico
2
Departamento de Sistemas y Computación, Instituto Tecnológico de Veracruz, Veracruz 91897, Veracruz, Mexico
3
INFOTEC Center for Research and Innovation in Information and Communication Technologies, Pocitos, Aguascalientes 20326, Aguascalientes, Mexico
*
Author to whom correspondence should be addressed.
Math. Comput. Appl. 2024, 29(4), 56; https://doi.org/10.3390/mca29040056
Submission received: 3 June 2024 / Revised: 7 July 2024 / Accepted: 9 July 2024 / Published: 13 July 2024
(This article belongs to the Special Issue Numerical and Evolutionary Optimization 2024)

Abstract

:
Feature selection is a preprocessing step in machine learning that aims to reduce dimensionality and improve performance. The approaches for feature selection are often classified according to the evaluation of a subset of features as filter, wrapper, and embedded approaches. The high performance of wrapper approaches for feature selection is associated at the same time with the disadvantage of high computational cost. Cost-reduction mechanisms for feature selection have been proposed in the literature, where competitive performance is achieved more efficiently. This work applies the simple and effective resource-saving mechanisms of the fixed and incremental sampling fraction strategies with memory to avoid repeated evaluations in multi-objective permutational-based differential evolution for feature selection. The selected multi-objective approach is an extension of the DE-FSPM algorithm with the selection mechanism of the GDE3 algorithm. The results showed high resource savings, especially in computational time and the number of evaluations required for the search process. Nonetheless, it was also detected that the algorithm’s performance was diminished. Therefore, the results reported in the literature on the effectiveness of the strategies for cost reduction in single-objective feature selection were only partially sustained in multi-objective feature selection.

1. Introduction

Feature selection (FS) is a dimensionality reduction preprocessing task in machine learning (ML) that deals with selecting the most relevant features in a dataset and discarding the noisy and irrelevant ones. By preferring only the features with meaningful information, the learning process complexity is reduced, and an ML algorithm’s performance in classification or clustering tasks is expected to be improved [1]. FS is an NP-Hard combinatorial problem with an exponentially growing search space of 2 n possible solutions for a dataset with n features. Given the complexity of the problem, nature-inspired metaheuristics, including Evolutionary Computation (EC) algorithms, constitute popular and effective methods for FS [2].
Dimensionality reduction methods have gained attention given the growing volumes of data with high dimensions generated nowadays in applications from different fields. By reducing the data dimensions, the ML algorithm training process is expected to be more efficient and fruitful [3,4]. An alternative feature reduction approach is feature extraction or feature reduction. It uses an algorithm to generate new features that combine and represent the information from the dataset features. In contrast, FS consists of selecting some of the original features in the dataset. An advantage of FS is that the interpretability of the data is not affected [5,6]. This work focuses on FS.
As presented in [3,7,8], the FS mechanism of assessing and comparing the goodness of a subset of features allows the classification of FS approaches into filter, wrapper, and embedded. Filters are the fastest approach, using an evaluation metric independent of the classification or clustering algorithm. Filters usually focus on the relevance of each feature or maximize the interaction between selected features. Wrapper approaches use an ML algorithm to determine the quality of a feature subset. The features that increase the performance of the ML algorithm are expected to be selected. Wrappers are the most resource-demanding approach, given the required several runs of the ML algorithm. Nonetheless, wrappers generally obtain the highest performance. Finally, embedded approaches incorporate the FS process in the training process of the ML algorithm. Decision trees are an example of an embedded approach that selects the most relevant features for classification in the training process. Embedded approaches are expected to be more complex than filters and faster than wrappers.
Differential Evolution (DE) is an evolutionary algorithm (EA) that is highlighted for its simplicity and robustness when applied to optimization problems [9]. As presented in [10], various works have extended the DE algorithm’s classic single-objective form to a multi-objective one. For the FS problem, various multi-objective DE proposals can be found in the literature using two common types of solution representation: a real-value codification with a threshold to determine if a feature is selected and binary adaptations of DE. Binary DE versions require changes in the mutation operator of DE, whereas using a real-value representation with a threshold allows the application of basic operators without changes. Examples of multi-objective DE for feature selection are found in [11,12] for filters with real-value codification, in [13,14] for wrappers with real-value codification, and in [15,16] for wrappers with binary representation. Alternative multi-objective DE approaches for FS can be found in [17] for unsupervised FS and [18] for FS using reinforcement learning.
An alternative and effective solution representation for DE applied for the FS problem is presented in [19] where the permutational-based DE algorithm for FS (DE-FSPM) is proposed. Permutations represent the individuals, and the mutation operator is modified to work with the selected representation. The reported results from the DE-FSPM show that it outperformed other metaheuristic and classic approaches for FS. In this manner, the effectiveness of using the DE adaptation to the permutational space was proved. In [20], the DE-FSPM algorithm is extended to a multi-objective version by combining it with the Generalized Differential Evolution 3 (GDE3) algorithm. The results show that the proposal effectively found a set of solutions that represent a trade-off between the objectives of minimizing the prediction error of a classifier and the number of selected features. Nonetheless, the single objective version of the DE-FSPM algorithm finds subsets with lower classification errors.
Cost-reduction approaches for the FS process are proposed in [21,22,23], where the search process for the most relevant subset of features is conducted using a fraction of the dataset instances, selected with random sampling, to reduce the evaluation cost. In [21], using the randomly selected group of dataset instances is tested in filter approaches for FS and feature extraction. The authors define a fixed number of instances to be sampled (100, 250, 500, 1500, and 2000) and select a subset of features with a predefined size of ten. Six methods are used with large-scale datasets, resulting in an execution time reduction when using the reduced subset of instances with minimal impact on the performance of the feature reduction method.
In [22], the approach of using random sampling for reducing the number of instances in the search process for feature selection, and therefore the computational cost, is extended for wrapper approaches for FS. The authors proposed three different random sampling-based strategies to reduce the number of dataset instances in the search process: the fixed, the incremental, and the evolving sampling fraction. In addition, the Success History parameter adaptation for DE from [24] was adapted to the FS problem with the DE-FSPM algorithm. The results were promising in maintaining the performance of the FS procedure, but the resource savings were scarce in computational time.
An extension of the work from [22] is presented in [23]. A memory mechanism to avoid repeated evaluations is added to work with the fixed and incremental sampling fraction strategies applied to the DE-FSPM algorithm. The memory attempts to solve the problem of wasted resources associated with evaluating duplicated individuals in DE presented in [25]. The method could perform similarly to the DE-FSPM original algorithm using the fixed sampling fraction strategy with memory to avoid repeated evaluations. The proposed approach reduced the average of 35.15% of the computational time and detected that an average of 35.35% of the evaluations could be avoided.
This work is based on the future work stated in [22,23], where the resource-saving mechanisms of the sampling strategies and memory to avoid repeated evaluations were proposed. The main contributions of this paper are the following:
  • We incorporate sampling strategies and memory to avoid repeated evaluations into the GDE3-based multi-objective version of the DE-FSPM algorithm.
  • We introduce two novel proposals: GDE3fix and GDE3inc. The former utilizes the fixed sampling fraction strategy, while the latter incorporates the incremental sampling fraction strategy. Both proposals use memory to avoid repeated evaluations.
  • We test the robustness of the cost-reduction mechanisms in multi-objective FS. The main goal was to determine if the results from the single-objective approach, where the computational cost was reduced without diminishing algorithm performance, are maintained in the multi-objective version of the algorithm.
  • We thoroughly analyze the effects of the proposals in terms of computational time consumption, the number of evaluations performed by the algorithm, and the number of instances used for an evaluation. Additionally, future work is described, including possible areas for improvement.
The rest of this document is organized into four sections. In Section 2, the details of the multi-objective FS process are introduced. After that, the DE algorithm is described along with its adaptation to the permutational codification of solutions and its extension to multi-objective optimization. Additionally, the cost-reduction mechanisms are presented in detail. Section 3 presents the experimentation details and results. Finally, Section 4 and Section 5 include the analysis of the results, the conclusions, and future research directions.

2. Materials and Methods

2.1. Multi-Objective Feature Selection

As presented in [6], the FS process is guided by two main objectives: maximizing the classification accuracy and minimizing the number of selected features. These objectives conflict and different feature subsets can represent trade-offs between obtaining a higher accuracy performance or a smaller subset of features. Usually, the classification accuracy maximization is transformed into the minimization of the classification error. This way, the goal is to minimize both objectives. Given that there are no constraints to the FS problem, in [26], the multi-objective FS problem is modeled following Equation (1). x represents a subset of selected features.
m i n i m i z e F ( x ) = [ f 1 ( x ) , f 2 ( x ) ]
f 1 ( x ) represents the classification error and is calculated using Equation (2). In the equation, the True Positive ( T P ), False Positive ( F P ), True Negative ( T N ), and False Negative ( F N ) values from the confusion matrix are required. A cross-validation (CV) approach is followed in our proposal to calculate the error rate. f 2 ( x ) corresponds to the number of selected features as presented in Equation (3). d is the number of selected features, and n is the total number of features in the dataset. f 1 ( x ) and f 2 ( x ) are in the range of [0, 1].
f 1 ( x ) = F P + F N T P + F P + T N + F N
f 2 ( x ) = d n
As presented in [27], a multi-objective optimization algorithm returns a set of solutions instead of a single solution as in single-objective optimization. The concept of Pareto optimal solutions is presented in multi-objective optimization. A solution is Pareto optimal if it is not dominated by any other solution. The weak dominance (⪯) is defined in Equation (4), where a solution x 1 weakly dominates x 2 ( x 1 x 2 ) if, and only if, all its objective values are less than or equal to the objective values of x 2 . The dominance (≺) of a solution x 1 over x 2 ( x 1 x 2 ) is given by Equation (5), adding to the weak dominance the condition that at least one objective value f m ( x 1 ) must be less than f m ( x 2 ) . In the case of the previously stated modeling of the FS problem, m = 2 since the objectives are the classification error and the normalized number of selected features. The Pareto optimal solutions compose the Pareto optimal set. The visualization of the solutions in the Pareto optimal set is the Pareto front.
x 1 x 2 i f f m { 1 , 2 , , M } : f m ( x 1 ) f m ( x 2 )
x 1 x 2 i f f x 1 x 2 m { 1 , 2 , , M } : f m ( x 1 ) < f m ( x 2 )

2.2. Differential Evolution

DE is a population-based metaheuristic algorithm for continuous optimization proposed in [28]. N individuals constitute the population. A vector of the size of the dimensions of the problem represents an individual ( x i ) in the population. The initial population consists of N individuals randomly created. Each individual has an associated fitness value provided by an evaluation using the selected fitness function that will guide the search process. The population is evolved in an iterative process where in each generation g, the mutation and crossing procedures are applied for each target vector x i to generate the noise ( v i ) and trial ( u i ) vectors. A user-defined parameter is the maximum number of generations G to evolve the population.
As presented in [29], the basic DE version is called D E / r a n d / 1 / b i n . r a n d and 1 come from calculating v i , where Equation (6) is applied. r a n d refers to the random selection of the r 0 , r 1 , and r 2 individuals from the population different from themselves and the x i that is being considered for mutation and crossing. The value 1 is given by the calculation of one vector difference. The scaling factor F is a scalar value defined by the user. b i n comes from the binomial distribution associated with the uniform crossover operator presented in Equation (7). The crossover rate C R is another user-defined parameter that controls the probability of choosing u i , j from v i or x i . To ensure that u i is not copied from x i , a random position J r a n d is chosen, where u i , j is guaranteed to be v i , j . Alternative DE variants have been proposed in the literature, changing the calculation of v i [9]. Some examples include D E / b e s t / 1 / , D E / r a n d / 2 / , and D E / c u r r e n t - t o - b e s t / . An alternative crossover mechanism is the exponential crossover ( e x p ).
v i = r 0 + F ( r 1 r 2 )
u i , j = v i , j if ( r a n d j C R ) o r ( j = J r a n d ) ; j = 1 , , | x i | x i , j otherwise
After u i is computed and evaluated, a binary tournament is performed to determine the individual with better fitness value comparing x i and u i . The winner is included in the population for the next generation of the evolutionary process. The previous process is an elitist selection mechanism that guarantees that the best solution found in the search process is never discarded. Once the G generations are computed, the process ends, and the individual with the highest fitness value in the population is returned as the best solution for the problem.

2.3. Permutational-Based Differential Evolution for Feature Selection

In [19], the DE-FSPM algorithm is proposed, modifying the DE algorithm to be applied to the FS problem. The first modification is that each individual is represented by a permutation with the indexes of the features in the dataset and a number zero used in the decoding process. The indexes that appear before the zero in the permutation are considered the selected features by the individual. Given the alternative encoding of the individuals, the DE procedure is adapted to the permutational space. The main changes are applied to the v i calculation. Equations (8) and (9) are used instead of Equation (6).
r 1 P r 2
v i P F r 0
Equation (8) presents the calculation of the permutation matrix P that maps r 1 and r 2 . After calculating P , a scaled permutation matrix P F is required. P F is used to apply some changes to r 0 in Equation (9); the parameter F controls how disruptive the mutation is. The process of calculating P F is presented in [30]. For each row i in P , if there is a 0 in the position P [ i , i ] and a random number r a n d i is greater than F, the row i is swapped with the row j where the position P [ j , i ] is a 1. When the swap is produced, a number 1 is included in the diagonal of the matrix. The positions in the diagonal of P F with a value of 1 represent no changes to r 0 when Equation (9) is applied. Greater F values will produce little change to P due to the low probability of r a n d i being greater than F. If F = 1 , no changes will be applied to P . By contrast, smaller F values will allow the setting of more elements in the P F diagonal as 1. If F = 0 , the resulting P F will be the diagonal matrix and v i = r 0 .
The uniform crossover operator of DE presented in Equation (7) is maintained in the DE-FSPM algorithm. This way, some elements in the permutation of u i come from v i and others from x i . As a result, u i could possess repeated elements requiring the application of a repair mechanism. The repair mechanism proposed in [19] removes all the repeated elements in u i ; the remaining elements are moved to the left to take the empty spaces left by the removed elements. Finally, the permutation is completed with the missing elements in the order they appear in x i . The mutation and crossover process of the DE-FSPM algorithm is illustrated with an example in Figure 1.
The accuracy metric guides the FS process in the DE-FSPM algorithm. A five-fold stratified cross-validation (CV) evaluation using the k-nearest-neighbors (KNN) algorithm is used as the fitness function. To determine the k value for the KNN algorithm, at the beginning of the search process, different values of k from 1 to 20 with a step of 2 are tested using all the dataset features, selecting the k with the highest performance in a ten-fold CV. After selecting k, the dataset is evaluated using all features with a ten-fold stratified CV. The ten-fold stratified CV evaluation is conducted again at the end of the search process with the selected subset of features.
The DE-FSPM algorithm requires a preprocessing step for the datasets used for the feature selection process. First, the missing values are imputed, considering the mean for numerical features and the mode for categorical features. The next step is converting the categorical features into numerical features. Finally, all features are normalized following a min-max normalization.

2.4. Generalized Differential Evolution 3 Algorithm

The GDE3 algorithm was proposed in [31], extending the capabilities of the DE algorithm to deal with problems with M objectives and K constraints. As presented in [10], the changes to the DE algorithm are applied in the selection procedure. Earlier GDE versions considered the Pareto dominance and crowding distance measurement when comparing x i and u i . GDE3’s population can grow temporally when specific criteria are met (as explained later), keeping both x i and u i . At the end of each generation, if the population grew, the Fast-Non-Dominated Sort and Crowding Distance (CD) assignment algorithms from [32] are applied to reduce the population to the size N. The Fast-Non-Dominated Sort assigns the non-dominated solutions to the front number 1. Front number 2 comprises the non-dominated solutions from the rest, and so on until all solutions are included in a front. The fronts formed by the algorithm are progressively added to the population for the next generation until one front does not fit entirely due to a surpass of N. Then, CD is applied to rank the individuals in the front. The ones with the highest CD value are selected.
In [27], it is explained that the changes applied to the DE algorithm in the generalized versions were intended to be as little as possible. As mentioned earlier, the modification is presented in the selection mechanism, and the rest is the same as the D E / r a n d / 1 / b i n procedure. The GDE3 algorithm is equivalent to the DE original procedure when applied to problems with M = 1 and K = 0 , i.e., problems with one objective and no constraints. To deal with constraint satisfaction problems ( K > 0 ), the selection mechanism of GDE3 is based on the concept of constraint-domination ( c ), a variant of the domination (≺) concept presented in Section 2.1. x 1 c x 2 when:
  • x 1 and x 2 are unfeasible, but x 1 violates the constraints less.
  • x 1 is feasible and x 2 is unfeasible.
  • x 1 and x 2 are feasible and x 1 x 2 .
The selection mechanism defined in [31] for the GDE3 algorithm consists of the following three cases:
  • When x i and u i are infeasible. The one that violates the least number of constraints is selected.
  • If x i is feasible and u i is unfeasible, x i is selected. On the other hand, if u i is feasible and x i is unfeasible, u i is selected.
  • When x i and u i are feasible, u i is selected if u i x i ; x i is selected if x i u i ; and both x i and u i are selected if x i u i and u i x i .
The third case is the only one in which the GDE3 algorithm allows population growth, given that both x i and u i can be selected. Since the FS problem presents no constraints ( k = 0 ), only the third case in the GDE3 selection mechanism is applied. In [20], the DE-FSPM algorithm is extended to work with two objectives using the optimization framework of the GDE3 algorithm. The objectives considered are the classification error and the number of selected features. The selection mechanism is the only modification to the DE-FSPM algorithm. In our research proposal, the number of selected features is normalized by dividing it by the total number of features in the dataset as presented in Section 2.1. The GDE3-based multi-objective version of the DE-FSPM algorithm is presented in Algorithm 1. As observed in Algorithm 1, the procedure returns the Pareto optimal solutions. Hence, a mechanism must be established to choose one of the returned solutions.
Algorithm 1 The multi-objective D E F S P M algorithm
  • Require:  D E F S P M G D E 3 ( C R , F , N , G )
  •    Input: The crossover rate ( C R ), the scale factor (F), the population size (N), and the number of generations (G).
  •    Output: Pareto optimal solutions in the current population.
  •  
  •     X 0
  •    for each   i { 1 , , N } do
  •         x i A permutation chosen at random from the solution space.
  •         X 0 X 0 { x i }
  •    end for
  •    for each   g { 1 , , G } do
  •         X g
  •         m 0
  •        for each  x i X g  do
  •             v i Mutated vector using Equations (8) and (9).
  •             u i Trial vector calculated using Equation (7) and the repair procedure.
  •             X g X g { u i } if u i x i { x i } otherwise
  •            if  u i x i x i u i  then
  •                X g X g { u i }
  •                m m + 0
  •            end if
  •        end for
  •        if  m > 0  then
  •             F R fast-non-dominated-sort ( X g )
  •             X g
  •             j 1
  •            while  | X g | < N  do
  •               if  | X g | + | F R j | N  then
  •                    X g X g F R j
  •               else
  •                    C D j Crowding-distance ( F R j )
  •                    s o r t ( F R j , C D j )                               ▹ Sort F R j in descending C D j order
  •                    X g X g F R j [ 0 : N | X g | ]
  •               end if
  •                j j + 1
  •            end while
  •        end if
  •    end for
  •     F R fast-non-dominated-sort ( X g )
  •    return  F R 1

2.5. Cost-Reduction Mechanisms for Feature Selection

The considered mechanisms for computational cost reduction are the fixed and incremental sampling fraction strategies from [22] and their incorporation with memory to avoid repeated evaluations from [23]. Both proposals were applied to the DE-FSPM algorithm that considers only one objective in its optimization process. Nonetheless, the mechanisms can also be applied to multi-objective optimization processes. The incorporation of both mechanisms into the multi-objective feature selection process is described next.

2.5.1. Fixed Sampling Fraction with Memory

The FS search process is conducted with a fraction of the dataset instances in the fixed sampling fraction strategy. The user defines two parameters: the initial sampling fraction S and the number of blocks in the search B. S is used at the beginning of the search process to apply random sampling and select part of the dataset instances. At that step, the memory used to avoid repeated evaluations is empty. When an individual is decoded for evaluation, the procedure searches in memory if there is a stored fitness value associated with the subset of features represented by the individual. The stored fitness value is returned if a coincidence is found in memory. Otherwise, the evaluation is performed using the fitness function. In the case of a multi-objective process, the memory stores the value obtained from the assessment of the individual in each of the considered objectives.
The parameter B divides the search process’s generations into groups. In the fixed sampling fraction, the memory is reset at the beginning of each block. The previous block is used to control the size of the memory. For example, if G is set as 100, B is set as five. The memory is reset at generations 1, 21, 41, 61, and 81. The memory mechanism is expected to have a more considerable impact on the last blocks of the search process due to the algorithm’s convergence, where population diversity diminishes and the probability of finding a duplicate individual increases.

2.5.2. Incremental Sampling Fraction with Memory

When the incremental sampling fraction strategy is applied, the process has the same start conditions as the fixed sampling fraction strategy. The user must also define the parameters S and B. Random sampling is applied at the beginning of the process, considering S to select part of the dataset instances. The memory mechanism is also used to avoid repeated evaluations. The characteristic aspect of the incremental sampling fraction strategy is that, at the beginning of each block of generations, not only is the memory reset, but more instances are proportionally added to the search process. In the last block of generations, all dataset instances are used. This way, fewer instances are considered in early generations, expecting that the algorithm will find promising areas of the search space with less costly evaluations. Then, late generations will likely be able to consider feature subsets with better generalization capabilities and avoid overfitting the selected feature subset to a fraction of dataset instances.
Given that the conditions of the FS problem change when more instances are considered, the population is reevaluated at the beginning of each block. The previous point is a drawback of the strategy due to the extra evaluations. Figure 2 presents a diagram of the multi-objective version of the DE-FSPM algorithm with the resource-saving modifications in the search process. It is seen that the incremental sampling fraction strategy requires the reevaluation process when a new block starts, while the fixed sampling fraction only requires resetting the memory.

3. Results

To test the effectiveness of the cost-reduction mechanisms from Section 2.5, the experimentation approach from [23] is adopted. The algorithm proposals are run 30 times for each of the eighteen datasets selected for experimentation. The code was implemented using Python, and the experiments were run in the virtual environments provided by Google Colab Pro. The datasets are selected from the UCI Machine Learning Repository [33], and their details are presented in Table 1.
The proposals are identified as:
  • GDE3: for the GDE3 version of the DE-FSPM algorithm.
  • GDE3fix: for the GDE3 version of the DE-FSPM algorithm using the fixed sampling fraction strategy with memory to avoid repeated evaluations.
  • GDE3inc: for the GDE3 version of the DE-FSPM algorithm using the incremental sampling fraction strategy with memory to avoid repeated evaluations.
The parameter configuration is adopted from [19], where the parameters G and N are defined for the DE-FSPM algorithm. F and C R are obtained in the fine-tuning of the GDE3 version of the DE-FSPM algorithm from [20]. Finally, the parameters S and B for the cost-reduction mechanisms are presented in [23]. The details are provided next.
  • G: 200 as the maximum number of generations.
  • N: 5 times the number of features in the dataset. The value is bound to have at least 200 individuals and at most 450.
  • F: 0.8305.
  • C R : 0.7049.
  • S: 0.6 for GDE3fix and GDE3inc.
  • B: 10 for GDE3fix and GDE3inc.
Three criteria were considered when selecting a solution from the resulting Pareto front of each method. First, the solution with the best performance in f 1 ( x ) represents the lowest classification error obtained by a solution. After that, the solution in the knee point of the front is considered a middle point in the trade-off between f 1 ( x ) and f 2 ( x ) . The knee point is calculated as the solution in the Pareto front with the smallest Euclidian distance concerning the point (0,0). Finally, the solution with better performance minimizing f 2 ( x ) is selected.
Three key aspects were monitored to assess the computational resources saved by the cost-reduction mechanism: the execution time, the number of evaluations used by the method, and the number of instances avoided in evaluation. The execution time is the time required on average by the FS procedure in the runs of each one of the proposals. The number of evaluations refers to the times an individual is evaluated with f 1 ( x ) and f 2 ( x ) . In the GDE3fix and GDE3inc proposals, the memory to avoid repeated evaluations is used to reduce the number of evaluations. Finally, the effect of using the fixed and incremental sampling strategies and the memory mechanism is measured with the sum of the number of instances used while evaluating an individual. Additionally, the hypervolume and spacing indicators for multi-objective optimization are reported.
The results are divided into three subsections of the document. First, in Section 3.1, the proposals are compared in terms of accuracy and the number of selected features. Then, in Section 3.2, the resource savings achieved by the GDE3fix and GDE3inc proposals are presented. Finally, Section 3.3 shows the performance of the proposals measured by the selected multi-objective indicators.

3.1. Accuracy Performance and Number of Selected Features

Table 2 presents the average results of the proposals when the solution selected from the Pareto front is the one with the minimum classification error. For this selection, the priority is given to f 1 ( x ) . For comparison, the table includes the accuracy and the number of selected features performance of the data without FS, and the results reported in [23] for the DE-FSPM algorithm. It is seen that the single-objective version of the DE-FSPM algorithm achieves a higher accuracy performance. Nonetheless, in almost all cases, the GDE3-based proposals attained the goal of reducing the data dimensionality and increasing the accuracy of using the dataset without feature selection.
The next comparison point is when the solution is selected from the Pareto front in the knee point. These solutions represent the trade-off between the accuracy performance and the number of selected features. Table 3 presents the results. In this case, the accuracy performance of the dataset without FS is presented as a reference. In all cases, the procedure reduced the dimensionality of the data. However, in most cases, the accuracy results are poorer than the accuracy performance achieved by the method without FS.
The third selection strategy was to prioritize the dimensional reduction and select the solution in the Pareto front with better performance in f 2 ( x ) . Table 4 presents the results of choosing the solution with the smallest feature subset. The accuracy performance of the dataset without FS is presented as a reference. It is seen that, in this case, the dimensionality reduction is maximum. All the proposals found a subset with only one feature, but the accuracy was severely affected. It is worth noticing the cases where only one feature obtains a performance close to the accuracy using all features, such as the arrhythmia, Australian, cylinder-b, CRX, and vote datasets.

3.2. Computational Cost Reduction

The effects of the cost-reduction mechanisms on the computational time, the number of instances used for evaluation, and the number of evaluations are presented in Table 5 and Table 6. Table 5 presents the average execution time of the three proposals with the percentage of reduction obtained using cost-reduction mechanisms. Both cost-reduction proposals could significantly reduce the computational time, but GDE3fix achieved a slightly higher time reduction than GDE3inc.
Table 6 shows that the memory mechanism can avoid around 80% of the evaluations in the search process. In reducing the number of instances used for evaluation, it is seen that GDE3fix achieves a higher reduction than GDE3inc. This behavior is expected given that GDE3inc proportionally incorporates more instances in the search process.
Figure 3 summarizes the effect of incorporating the fixed and incremental sampling fraction strategies with memory to avoid repeated evaluations in the multi-objective version of the DE-FSPM algorithm. It is seen that, despite reducing the required evaluations in a similar percentage, the GDE3inc proposal achieved a higher reduction in the computational cost of the FS procedure and in the number of instances used during evaluation. A higher reduction in the number of instances was expected, given that GDE3fix uses the same amount of instances throughout the FS process, while GDE3inc uses more instances in the final generations.

3.3. Multi-Objective Optimization Indicators

Finally, Table 7 presents the performance of the proposals in the hypervolume and spacing indicators for multi-objective optimization. In most cases, the GDE3 proposal achieved a larger hypervolume value, but the GDE3fix and GDE3 proposals presented a better spacing value. The values from the spacing are expected to be small in this problem due to the discrete nature of the Pareto front. The solutions select feature subsets with integer sizes, distributing them in the front.

4. Discussion

Statistical tests were conducted as suggested in [34] to evaluate if the resource-saving mechanisms could maintain the performance of the GDE3 algorithm. First, the Friedman test was applied with the accuracy results of the GDE3, GDE3fix, and GDE3inc proposals from Table 2, obtaining a p-value of 1.52 × 10 8 that indicates significant differences in the means of the proposals. The post hoc Nemenyi test was then applied to compare the proposals, and the results showed that significant differences were presented in all cases.
An alternative comparison is performed using the hypervolume indicator results to see if the cost-saving mechanisms are affecting the multi-objective search capabilities of the algorithm. The statistical tests were rerun considering the hypervolume values of the GDE3, GDE3fix, and GDE3inc proposals from Table 7. The Friedman test resulted in a p-value of 3.84 × 10 5 , showing significant differences among the proposals. The Nemenyi post hoc test indicated significant differences for the GDE3 and the GDE3fix proposals and for the GDE3 and the GDE3inc procedures.
In contrast with the findings from [23], incorporating the fixed sampling fraction proposal with memory to avoid repeated evaluations resulted in diminishing the performance of the multi-objective FS method. Nonetheless, the fixed and incremental sampling strategies with memory achieved considerably higher resource savings. The time reduction observed in the single-objective approach was reported as 35.16% and 48.61%, respectively, for the fixed and incremental proposals. Whereas in the multi-objective approach, the time reduction was 80.92% and 78.49%, for the fixed and incremental proposals, respectively. The fixed proposal achieved a higher reduction in this case.
An interesting thing to analyze is the number of evaluations the memory mechanism avoids. The reduction of 80% in the number of evaluations clearly indicates that the adaptation of the DE-FSPM algorithm to multi-objective optimization using GDE3 is finding a high number of duplicate individuals. Population diversity and the crowdedness of the solutions are aspects to be considered in future attempts to improve the algorithm’s performance.
Another future direction of implementing the cost-reduction mechanisms with the multi-objective version of the DE-FSPM algorithm is to find an adequate compromise between performance and resource savings. As seen in [23], the algorithm’s performance is not affected, but the reported resource savings are fewer than the ones found in this work despite using the same configuration of the mechanisms. The previous observation suggests that the cost-reduction mechanisms require a more specific parameter configuration for their application in different FS approaches. Additionally, a comparison among the subsets of features selected by each proposal can provide valuable insights into the differences in the results of the FS process. Applying the FS processes in a real-life application, like the one from [35], will result in a deeper comparison of the results when collaborating with an expert in the field.

5. Conclusions

In this work, fixed and incremental sampling fraction strategies with memory to avoid repeated evaluations were implemented as cost-reduction mechanisms in a multi-objective permutational-based DE approach for FS. The proposed approach for multi-objective FS presented some limitations. The main one is in the method’s performance in the accuracy and multi-objective indicators, which were reduced by the use of cost-reduction mechanisms. The proposals presented limited exploration capabilities, evidenced by the number of evaluations they avoid due to the finding of repeated individuals. The results exhibit high savings of computational resources at the expense of reducing the method’s performance. The previous observation shows that the findings in the single-objective version of the algorithm, where the fixed sampling fraction proposal achieved the goal of reducing the computational cost of a wrapper approach for FS without diminishing its performance, is not maintained in a multi-objective FS approach.
The multi-objective approach for FS using the permutational version of GDE3 does not achieve accuracy results as high as the single-objective version of the DE-FSPM algorithm. Nevertheless, the advantage of using a multi-objective approach is that more than one solution to the problem is found, allowing the user to choose the most convenient subset of features among the options in the Pareto front. The GDE3-based proposals found solutions with smaller feature subsets and involved different compromises between the accuracy performance and the number of selected features. The previous results are an advantage of the multi-objective approach. However, it requires an additional step in the process that can be performed with an automatic technique or by following the recommendation of an expert in the data field in the context of using the FS technique in a real-life application.
The parameter configuration used affects the algorithm’s search capabilities. In DE, the F and C R parameters control the algorithm’s exploration/exploitation capabilities. Finding an adequate set of parameters for every dataset in the FS process is a complicated task. A parameter adaptation scheme can help increase the algorithm’s search capabilities in future work. Another critical aspect for future improvements is considering mechanisms to improve the population’s diversity. Given the high number of repeated individuals found, the procedure appears to suffer from stagnation. Following the previous point, another area to explore in future changes to the algorithm is the mechanism used to reduce the population when it grows with the selection mechanism of the GDE3 algorithm. Finding more effective selection criteria for the FS problem will be helpful.
Future experimentation, in which the resource-saving mechanisms are applied to different single- and multi-objective approaches for FS, will provide more insights into how robust the mechanisms are for computational cost reduction. As presented in this work, finding a balance in the severity of the proposed resource savings is necessary if the goal is to maintain the algorithm’s performance. Consequently, additional experimentation focusing on the effect of the parameters S and B will be helpful. A further aspect to be considered in future experimentation is the mechanism used to select a solution among the Pareto optimal ones that the algorithm is returning. A more complex selection procedure can provide options for considering a different trade-off between the objectives than just selecting the solution that performs best in one objective or the knee point in the front.

Author Contributions

Conceptualization, J.-A.B.-P., R.R.-L., E.M.-M., H.-G.A.-M. and A.M.-G.; methodology, R.R.-L. and E.M.-M.; software, J.-A.B.-P.; validation, R.R.-L., E.M.-M. and H.-G.A.-M.; formal analysis, J.-A.B.-P. and A.M.-G.; investigation, J.-A.B.-P.; resources, R.R.-L. and E.M.-M.; writing—original draft preparation, J.-A.B.-P.; writing—review and editing, R.R.-L., E.M.-M., H.-G.A.-M. and A.M.-G.; visualization, J.-A.B.-P. and A.M.-G.; supervision, E.M.-M. and R.R.-L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available in the UCI Machine Learning repository at https://archive.ics.uci.edu/, reference number [33].

Acknowledgments

The first author (CVU 1142850) acknowledges support from the Mexican National Council of Humanities, Science, and Technology (CONAHCYT) with a scholarship to pursue PhD studies at Universidad Veracruzana.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
FSFeature Selection
DEDifferential Evolution
EAEvolutionary Algorithm
CVCross-validation
GDE3Generalized Differential Evolution 3
KNNK-nearest-neighbors
DE-FSPMPermutational-based Differential Evolution Algorithm for Feature Selection

References

  1. Sharma, M.; Kaur, P. A Comprehensive Analysis of Nature-Inspired Meta-Heuristic Techniques for Feature Selection Problem. Arch. Comput. Methods Eng. 2021, 28, 1103–1127. [Google Scholar] [CrossRef]
  2. Dokeroglu, T.; Deniz, A.; Kiziloz, H.E. A comprehensive survey on recent metaheuristics for feature selection. Neurocomputing 2022, 494, 269–296. [Google Scholar] [CrossRef]
  3. Abdulwahab, H.M.; Ajitha, S.; Saif, M.A.N. Feature selection techniques in the context of big data: Taxonomy and analysis. Appl. Intell. 2022, 52, 13568–13613. [Google Scholar] [CrossRef]
  4. Brezočnik, L.; Fister, I.; Podgorelec, V. Swarm Intelligence Algorithms for Feature Selection: A Review. Appl. Sci. 2018, 8, 1521. [Google Scholar] [CrossRef]
  5. Agrawal, P.; Abutarboush, H.F.; Ganesh, T.; Mohamed, A.W. Metaheuristic Algorithms on Feature Selection: A Survey of One Decade of Research (2009–2019). IEEE Access 2021, 9, 26766–26791. [Google Scholar] [CrossRef]
  6. Xue, B.; Zhang, M.; Browne, W.N.; Yao, X. A Survey on Evolutionary Computation Approaches to Feature Selection. IEEE Trans. Evol. Comput. 2016, 20, 606–626. [Google Scholar] [CrossRef]
  7. Dhal, P.; Azad, C. A comprehensive survey on feature selection in the various fields of machine learning. Appl. Intell. 2022, 52, 4543–4581. [Google Scholar] [CrossRef]
  8. Theng, D.; Bhoyar, K.K. Feature selection techniques for machine learning: A survey of more than two decades of research. Knowl. Inf. Syst. 2024, 66, 1575–1637. [Google Scholar] [CrossRef]
  9. Ahmad, M.F.; Isa, N.A.M.; Lim, W.H.; Ang, K.M. Differential evolution: A recent review based on state-of-the-art works. Alex. Eng. J. 2022, 61, 3831–3872. [Google Scholar] [CrossRef]
  10. Mezura-Montes, E.; Reyes-Sierra, M.; Coello, C.A.C. Multi-objective Optimization Using Differential Evolution: A Survey of the State-of-the-Art. In Advances in Differential Evolution; Chakraborty, U.K., Ed.; Springer: Berlin/Heidelberg, Germany, 2008; pp. 173–196. [Google Scholar] [CrossRef]
  11. Hancer, E.; Xue, B.; Zhang, M. Differential evolution for filter feature selection based on information theory and feature ranking. Knowl.-Based Syst. 2018, 140, 103–119. [Google Scholar] [CrossRef]
  12. Hancer, E.; Xue, B.; Zhang, M. An evolutionary filter approach to feature selection in classification for both single- and multi-objective scenarios. Knowl.-Based Syst. 2023, 280, 111008. [Google Scholar] [CrossRef]
  13. Xue, B.; Fu, W.; Zhang, M. Multi-objective Feature Selection in Classification: A Differential Evolution Approach. In Proceedings of the Simulated Evolution and Learning; Dick, G., Browne, W.N., Whigham, P., Zhang, M., Bui, L.T., Ishibuchi, H., Jin, Y., Li, X., Shi, Y., Singh, P., et al., Eds.; Springer: Cham, Switzerland, 2014; pp. 516–528. [Google Scholar]
  14. Wang, P.; Xue, B.; Liang, J.; Zhang, M. Differential Evolution-Based Feature Selection: A Niching-Based Multiobjective Approach. IEEE Trans. Evol. Comput. 2023, 27, 296–310. [Google Scholar] [CrossRef]
  15. Bidgoli, A.A.; Ebrahimpour-Komleh, H.; Rahnamayan, S. A Novel Multi-objective Binary Differential Evolution Algorithm for Multi-label Feature Selection. In Proceedings of the 2019 IEEE Congress on Evolutionary Computation (CEC), Wellington, New Zealand, 10–13 June 2019; pp. 1588–1595. [Google Scholar] [CrossRef]
  16. Wang, P.; Xue, B.; Liang, J.; Zhang, M. Feature Selection Using Diversity-Based Multi-objective Binary Differential Evolution. Inf. Sci. 2023, 626, 586–606. [Google Scholar] [CrossRef]
  17. Hancer, E. A new multi-objective differential evolution approach for simultaneous clustering and feature selection. Eng. Appl. Artif. Intell. 2020, 87, 103307. [Google Scholar] [CrossRef]
  18. Yu, X.; Hu, Z.; Luo, W.; Xue, Y. Reinforcement learning-based multi-objective differential evolution algorithm for feature selection. Inf. Sci. 2024, 661, 120185. [Google Scholar] [CrossRef]
  19. Rivera-López, R.; Mezura-Montes, E.; Canul-Reich, J.; Cruz-Chávez, M.A. A permutational-based Differential Evolution algorithm for feature subset selection. Pattern Recognit. Lett. 2020, 133, 86–93. [Google Scholar] [CrossRef]
  20. Mendoza-Mota, J.A. Selección de Atributos con un Enfoque Evolutivo Multiobjetivo. Master’s Thesis, Laboratorio Nacional de Informática Avanzada, Xalapa-Enríquez, Mexico, 2021. [Google Scholar]
  21. Malekipirbazari, M.; Aksakalli, V.; Shafqat, W.; Eberhard, A. Performance comparison of feature selection and extraction methods with random instance selection. Expert Syst. Appl. 2021, 179, 115072. [Google Scholar] [CrossRef]
  22. Barradas-Palmeros, J.A.; Rivera-López, R.; Mezura-Montes, E.; Acosta-Mesa, H.G. Experimental Study of the Instance Sampling Effect on Feature Subset Selection Using Permutational-Based Differential Evolution. In Proceedings of the Advances in Computational Intelligence, MICAI 2023 International Workshops; Calvo, H., Martínez-Villaseñor, L., Ponce, H., Zatarain Cabada, R., Montes Rivera, M., Mezura-Montes, E., Eds.; Springer: Cham, Switzerland, 2024; pp. 409–421. [Google Scholar] [CrossRef]
  23. Barradas-Palmeros, J.A.; Mezura-Montes, E.; Rivera-López, R.; Acosta-Mesa, H.G. Computational Cost Reduction in Wrapper Approaches for Feature Selection: A Case of Study Using Permutational-Based Differential Evolution (In press). In Proceedings of the 2024 IEEE Congress on Evolutionary Computation (CEC), Yokohama, Japan, 30 June–5 July 2024. [Google Scholar]
  24. Tanabe, R.; Fukunaga, A. Success-history based parameter adaptation for Differential Evolution. In Proceedings of the 2013 IEEE Congress on Evolutionary Computation, Cancun, Mexico, 20–23 June 2013; pp. 71–78. [Google Scholar] [CrossRef]
  25. Kitamura, T.; Fukunaga, A. Duplicate Individuals in Differential Evolution. In Proceedings of the 2022 IEEE Congress on Evolutionary Computation (CEC), Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar] [CrossRef]
  26. Al-Tashi, Q.; Abdulkadir, S.J.; Rais, H.M.; Mirjalili, S.; Alhussian, H. Approaches to Multi-Objective Feature Selection: A Systematic Literature Review. IEEE Access 2020, 8, 125076–125096. [Google Scholar] [CrossRef]
  27. Kukkonen, S.; Coello, C.A. Generalized Differential Evolution for Numerical and Evolutionary Optimization. In NEO 2015: Results of the Numerical and Evolutionary Optimization Workshop NEO 2015 Held at September 23–25 2015 in Tijuana, Mexico; Schütze, O., Trujillo, L., Legrand, P., Maldonado, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 253–279. [Google Scholar] [CrossRef]
  28. Storn, R.; Price, K. Differential Evolution—A Simple and Efficient Heuristic for global Optimization over Continuous Spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  29. Eiben, A.E.; Smith, J.E. Introduction to Evolutionary Computing; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar] [CrossRef]
  30. Price, K.V.; Storn, R.M.; Lampinen, J.A. Differential Evolution: A Practical Approach to Global Optimization; Springer: Berlin/Heidelberg, Germany, 2005. [Google Scholar] [CrossRef]
  31. Kukkonen, S.; Lampinen, J. GDE3: The third evolution step of generalized differential evolution. In Proceedings of the 2005 IEEE Congress on Evolutionary Computation, Edinburgh, UK, 2–5 September 2005; Volume 1, pp. 443–450. [Google Scholar] [CrossRef]
  32. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef]
  33. Kelly, M.; Longjohn, R.; Nottingham, K. The UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu (accessed on 25 July 2023).
  34. Derrac, J.; García, S.; Molina, D.; Herrera, F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evol. Comput. 2011, 1, 3–18. [Google Scholar] [CrossRef]
  35. Vargas-Moreno, I.; Rodríguez-Landa, J.F.; Acosta-Mesa, H.G.; Fernández-Demeneghi, R.; Oliart-Ros, R.; Hernández Baltazar, D.; Herrera-Meza, S. Effects of Sterculia Apetala Seed Oil on Anxiety-like Behavior and Neuronal Cells in the Hippocampus in Rats. J. Food Nutr. Res. 2023, 11, 211–222. [Google Scholar] [CrossRef]
Figure 1. An example of the DE-FSPM algorithm mutation and crossover procedures. The decoding of an individual x i is presented with an example of calculating u i and applying the repair mechanism to maintain a valid permutation. The random values required by the process are also shown. The parameters considered for the example are F = 0.6 and C R = 0.4 . After calculating u i , it is evaluated, and its associated fitness value is compared with the x i fitness value. The most fitted one is included in the next-generation population. Different colors are used in the figure to represent the x i , v i , and u i vectors.
Figure 1. An example of the DE-FSPM algorithm mutation and crossover procedures. The decoding of an individual x i is presented with an example of calculating u i and applying the repair mechanism to maintain a valid permutation. The random values required by the process are also shown. The parameters considered for the example are F = 0.6 and C R = 0.4 . After calculating u i , it is evaluated, and its associated fitness value is compared with the x i fitness value. The most fitted one is included in the next-generation population. Different colors are used in the figure to represent the x i , v i , and u i vectors.
Mca 29 00056 g001
Figure 2. Diagram of the multi-objective DE-FSPM algorithm with the cost-reduction mechanisms applied in the search process. S and B are considered 0.5 and 6, respectively, as examples of their effect on managing the dataset instances in the search process.
Figure 2. Diagram of the multi-objective DE-FSPM algorithm with the cost-reduction mechanisms applied in the search process. S and B are considered 0.5 and 6, respectively, as examples of their effect on managing the dataset instances in the search process.
Mca 29 00056 g002
Figure 3. Resource consumption reduction achieved by the proposals GDE3fix and GDE3inc. The plot presents the average reductions concerning the GDE3 procedure in terms of computational time, the number of evaluations, and the number of instances used when evaluating the individuals.
Figure 3. Resource consumption reduction achieved by the proposals GDE3fix and GDE3inc. The plot presents the average reductions concerning the GDE3 procedure in terms of computational time, the number of evaluations, and the number of instances used when evaluating the individuals.
Mca 29 00056 g003
Table 1. Details of the datasets selected for experimentation.
Table 1. Details of the datasets selected for experimentation.
DatasetFeaturesInstancesClasses
Arrhythmia27945216
Audiology6922624
Australian146902
Cylinder-b395402
CRX156902
Dermatology343666
German-c2010002
Hill valley10012122
Ionosphere343512
M-libras9036015
Musk 11684762
Parkinsons221952
Sonar602082
Soybean3568319
SPECTF442672
Vehicle188464
Vote164352
WDBC305692
Table 2. Accuracy and feature selection results of selecting the solution with the best performance in accuracy from the Pareto front. The best result is marked in bold, and the best result considering only the GDE3, GDE3fix, and GDE3inc proposals is underlined.
Table 2. Accuracy and feature selection results of selecting the solution with the best performance in accuracy from the Pareto front. The best result is marked in bold, and the best result considering only the GDE3, GDE3fix, and GDE3inc proposals is underlined.
DatasetWoFSDE-FS(PM)GDE3GDE3fixGDE3inc
SF Acc SF Acc SF Acc SF Acc SF Acc
Arrhtythmia27958.7910.975.6610.570.7316.066.3712.166.75
Audiology6976.4725.185.6342.982.4343.779.7244.981.00
Australian1485.986.786.829.586.486.786.297.686.47
Cylinder-b3974.513.984.192.283.5110.280.972.382.65
CRX1586.139.087.4910.887.178.086.169.086.94
Dermatology3496.8919.198.3525.798.0223.997.1726.697.65
German-c2075.359.177.5315.476.1412.674.3414.875.65
Hill valley10064.549.971.6410.569.5214.967.2416.668.48
Ionosphere3486.976.594.826.693.455.790.528.092.13
M-libras9086.0624.088.8924.987.8639.586.2347.387.19
Musk 116885.6536.194.2342.491.2352.587.9860.589.45
Parkinsons2295.8410.998.849.698.2413.395.4412.897.39
Sonar6086.7022.493.7925.290.5625.187.0034.389.03
Soybean3591.8517.394.9121.394.3023.392.6922.693.61
SPECTF4477.685.584.226.983.0015.379.349.279.93
Vehicle1870.089.974.3111.173.1911.671.6211.972.15
Vote1693.505.296.594.796.544.395.382.595.93
WDBC3097.0718.297.5523.597.2617.696.8222.597.17
Table 3. Accuracy and feature selection results of selecting the solution in the knee point from the Pareto front. The best result is marked in bold, and the best result considering only the GDE3, GDE3fix, and GDE3inc proposals is underlined.
Table 3. Accuracy and feature selection results of selecting the solution in the knee point from the Pareto front. The best result is marked in bold, and the best result considering only the GDE3, GDE3fix, and GDE3inc proposals is underlined.
DatasetWoFSDE-FS(PM)GDE3GDE3fixGDE3inc
SF Acc SF Acc SF Acc SF Acc SF Acc
Arrhtythmia27958.7910.975.669.270.5411.066.127.766.81
Audiology6976.4725.185.639.873.0811.468.0813.072.95
Australian1485.986.786.821.085.511.085.511.085.42
Cylinder-b3974.513.984.192.083.302.281.692.182.41
CRX1586.139.087.491.084.571.083.821.084.86
Dermatology3496.8919.198.354.490.424.587.984.790.87
German-c2075.359.177.531.871.131.970.841.971.69
Hill valley10064.549.971.646.969.465.566.156.868.11
Ionosphere3486.976.594.822.289.172.487.332.789.20
M-libras9086.0624.088.898.586.929.684.598.885.44
Musk 116885.6536.194.2312.388.6911.585.1612.886.91
Parkinsons2295.8410.998.842.093.481.989.472.192.04
Sonar6086.7022.493.795.886.325.079.355.983.19
Soybean3591.8517.394.915.585.945.983.555.983.96
SPECTF4477.685.584.222.383.302.379.762.079.95
Vehicle1870.089.974.312.265.882.465.822.866.93
Vote1693.505.296.591.095.131.095.631.095.63
WDBC3097.0718.297.552.095.311.893.862.094.79
Table 4. Accuracy and feature selection results of selecting the solution with the best performance in the number of selected features from the Pareto front. The best result is marked in bold, and the best result considering only the GDE3, GDE3fix, and GDE3inc proposals is underlined.
Table 4. Accuracy and feature selection results of selecting the solution with the best performance in the number of selected features from the Pareto front. The best result is marked in bold, and the best result considering only the GDE3, GDE3fix, and GDE3inc proposals is underlined.
DatasetWoFSDE-FS(PM)GDE3GDE3fixGDE3inc
SF Acc SF Acc SF Acc SF Acc SF Acc
Arrhtythmia27958.7910.975.66157.64156.04156.42
Audiology6976.4725.185.63129.13127.68127.57
Australian1485.986.786.82185.51185.51185.29
Cylinder-b3974.513.984.19174.88174.56174.86
CRX1586.139.087.49184.56183.28184.78
Dermatology3496.8919.198.35147.10143.27146.09
German-c2075.359.177.53168.08167.65169.60
Hill valley10064.549.971.64155.56153.45154.77
Ionosphere3486.976.594.82179.17177.97178.77
M-libras9086.0624.088.89125.88123.49125.29
Musk 116885.6536.194.23163.07161.74162.44
Parkinsons2295.8410.998.84180.61178.30179.46
Sonar6086.7022.493.79167.07165.22167.20
Soybean3591.8517.394.91129.56130.51128.89
SPECTF4477.685.584.22178.33178.15178.50
Vehicle1870.089.974.31150.43148.79149.58
Vote1693.505.296.59195.38195.41195.54
WDBC3097.0718.297.55191.07190.35190.73
Table 5. Time reduction obtained when applying the cost-reduction mechanisms to the multi-objective feature selection process of the DE-FSPM algorithm with GDE3. The larger reduction in computational time between the GDE3fix and GDE3inc proposals is marked in bold.
Table 5. Time reduction obtained when applying the cost-reduction mechanisms to the multi-objective feature selection process of the DE-FSPM algorithm with GDE3. The larger reduction in computational time between the GDE3fix and GDE3inc proposals is marked in bold.
DatasetGDE3GDE3fixGDE3inc
Arrhythmia8038.5082465.11569.33%2514.49468.72%
Audiology3991.2531035.45674.06%1170.44070.67%
Australian3032.478233.61892.30%228.90292.45%
Cylinder-b2696.103429.12884.08%364.85486.47%
CRX2939.885287.54590.22%239.61891.85%
Dermatology2561.947544.80778.73%818.31668.06%
German-c4036.628423.84689.50%524.60187.00%
Hill valley11,660.1992057.72382.35%2662.47177.17%
Ionosphere2171.877353.16183.74%372.97282.83%
M-libras5755.8651765.32269.33%2640.83254.12%
Musk 17632.8183085.92459.57%3704.49551.47%
Parkinsons1853.851299.76783.83%348.36781.21%
Sonar3417.529782.01477.12%849.62275.14%
Soybean2844.210820.75871.14%700.03275.39%
SPECTF2627.962315.10588.01%331.94187.37%
Vehicle3228.412423.96986.87%420.57186.97%
Vote2278.412173.89792.37%164.08892.80%
WDBC3075.011492.20983.99%518.76883.13%
Average 80.92% 78.49%
Table 6. Percentages of avoided evaluations and reduction in the number of instances used for evaluation. The proposal representing more significant savings in each case is marked in bold.
Table 6. Percentages of avoided evaluations and reduction in the number of instances used for evaluation. The proposal representing more significant savings in each case is marked in bold.
DatasetEvaluationsInstances
GDE3fix GDE3inc GDE3fix GDE3inc
Arrhythmia73.05%75.94%83.84%75.94%
Audiology77.64%74.12%86.59%74.12%
Australian90.91%93.56%94.55%93.56%
Cylinder-b83.24%88.07%89.94%88.07%
CRX89.43%93.00%93.66%93.00%
Dermatology75.43%68.12%85.23%68.12%
German-c86.74%86.16%92.04%86.16%
Hill valley77.46%74.32%86.48%74.32%
Ionosphere84.12%84.37%90.46%84.37%
M-libras69.98%57.58%81.99%57.58%
Musk 151.83%51.52%71.06%51.52%
Parkinsons84.35%84.11%90.61%84.11%
Sonar77.29%78.07%86.36%78.07%
Soybean71.12%73.27%82.66%73.27%
SPECTF89.09%89.84%93.46%89.84%
Vehicle85.66%86.55%91.39%86.55%
Vote92.38%95.01%95.43%95.01%
WDBC81.10%82.55%88.67%82.55%
Average80.04%79.79%88.02%79.79%
Table 7. Proposal’s results for the hypervolume and spacing indicators. The method with the best performance in each case is marked in bold.
Table 7. Proposal’s results for the hypervolume and spacing indicators. The method with the best performance in each case is marked in bold.
DatasetHypervolumeSpacing
GDE3 GDE3fix GDE3inc GDE3 GDE3fix GDE3inc
Arrhythmia0.70930.68160.66600.00020.00000.0001
Audiology0.78700.73520.74220.00270.00030.0003
Australian0.81020.81410.80610.00050.00020.0004
Cylinder-b0.83480.78090.79530.00060.00080.0000
CRX0.81670.82010.81050.00050.00030.0007
Dermatology0.91820.91220.90830.00300.00040.0003
German-c0.72820.73010.71300.01610.00040.0005
Hill valley0.69260.64570.67120.00030.00010.0001
Ionosphere0.90860.90280.88760.00080.00010.0002
M-libras0.86630.82030.85080.00530.00040.0005
Musk 10.90440.88220.88520.00120.00080.0004
Parkinsons0.93660.92080.91510.00370.00030.0011
Sonar0.89550.88300.86380.00190.00020.0004
Soybean0.87070.85310.85010.00410.00040.0004
SPECTF0.83370.84010.79360.00260.00090.0002
Vehicle0.68320.67220.66180.00330.00050.0003
Vote0.90480.90280.89920.00010.00010.0000
WDBC0.94420.94360.93850.00380.00070.0003
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Barradas-Palmeros, J.-A.; Mezura-Montes, E.; Rivera-López, R.; Acosta-Mesa, H.-G.; Márquez-Grajales, A. Computational Cost Reduction in Multi-Objective Feature Selection Using Permutational-Based Differential Evolution. Math. Comput. Appl. 2024, 29, 56. https://doi.org/10.3390/mca29040056

AMA Style

Barradas-Palmeros J-A, Mezura-Montes E, Rivera-López R, Acosta-Mesa H-G, Márquez-Grajales A. Computational Cost Reduction in Multi-Objective Feature Selection Using Permutational-Based Differential Evolution. Mathematical and Computational Applications. 2024; 29(4):56. https://doi.org/10.3390/mca29040056

Chicago/Turabian Style

Barradas-Palmeros, Jesús-Arnulfo, Efrén Mezura-Montes, Rafael Rivera-López, Hector-Gabriel Acosta-Mesa, and Aldo Márquez-Grajales. 2024. "Computational Cost Reduction in Multi-Objective Feature Selection Using Permutational-Based Differential Evolution" Mathematical and Computational Applications 29, no. 4: 56. https://doi.org/10.3390/mca29040056

Article Metrics

Back to TopTop