Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Surfaces with Constant Negative Curvature
Previous Article in Journal
Ollivier–Ricci Curvature Based Spatio-Temporal Graph Neural Networks for Traffic Flow Forecasting
Previous Article in Special Issue
Dynamic Dual-Threshold Virtual Machine Merging Method Based on Three-Way Decision
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Forward Greedy Searching to κ-Reduct Based on Granular Ball

School of Computer, Jiangsu University of Science and Technology, Zhenjiang 212100, China
*
Author to whom correspondence should be addressed.
Symmetry 2023, 15(5), 996; https://doi.org/10.3390/sym15050996
Submission received: 23 March 2023 / Revised: 22 April 2023 / Accepted: 25 April 2023 / Published: 27 April 2023
(This article belongs to the Special Issue Recent Advances in Granular Computing for Intelligent Data Analysis)

Abstract

:
As a key part of data preprocessing, namely attribute reduction, is effectively applied in the rough set field. The purpose of attribute reduction is to prevent too many attributes from affecting classifier operations and reduce the dimensionality of data space. Presently, in order to further improve the simplification performance of attribute reduction, numerous researchers have proposed a variety of methods. However, given the current findings, the challenges are: to reasonably compress the search space of candidate attributes; to fulfill multi-perspective evaluation; and to actualize attribute reduction based on guidance. In view of this, forward greedy searching to κ -reduct based on granular ball is proposed, which has the following advantages: (1) forming symmetrical granular balls to actualize the grouping of the universe; (2) continuously merging small universes to provide guidance for subsequent calculations; and (3) combining supervised and unsupervised perspectives to enrich the viewpoint of attribute evaluation and better improve the capability of attribute reduction. Finally, based on three classifiers, 16 UCI datasets are used to compare our proposed method with six advanced algorithms about attribute reduction and an algorithm without applying any attribute reduction algorithms. The experimental results indicate that our method can not only ensure the result of reduction has considerable performance in the classification test, but also improve the stability of attribute reduction to a certain degree.

1. Introduction

With the rapid increase in data, attributes have become redundant and uncertain. Uncertainty mainly consists of the five following aspects: incompleteness, inconsistency, incompatibility, fuzziness and randomness. Consequently, finding valuable information from high-dimensional data is a challenge for the research field.
With the intention of effectively disposing of ambiguous, incomplete, and inaccurate data, Polish scholar Pawlak first put forward rough set theory [1] in 1982, which has been extensively adopted in data mining, pattern recognition, decision analysis [2,3,4], and other domains. Based on rough set theory, many extensions and improvements have been proposed, such as neighborhood rough set [5], fuzzy rough set [6], decision-theoretic rough set [7], and Pythagorean fuzzy set [8]. Attribute reduction [9,10,11,12,13], as a common dimensional reduction method, can efficaciously remove redundant components in information systems, choose the optimal minimum attribute subset and further improve the effectiveness of data knowledge discovery. Obviously, attribute reduction has grown to be a paramount research branch of rough set theory.
Generally speaking, simplified searching strategies can be split into two general classes: exhaustive search and heuristic search [14]. In the process of data analysis, the final reduction result is directly related to the given constraint which can be implemented by constructing different measurement criteria.
As a mature heuristics-based search, The forward greedy strategy has a wide range of measures [15,16,17], such as approximation quality and conditional entropy [18,19,20,21]. These measures are especially used for assessing attributes and exporting reduction results. However, in the exploration of attribute reduction, many researchers only take into account the single view measure to determine the constraint. For instance, Jiang et al. [22] studied the supervised neighborhood attribute reduction; Zhang et al. [23] investigated a semi-supervised attribute reduction method that combined the collaborative learning theory; and Yuan et al. [24] introduced a fuzzy complementary entropy measure and proposed an unsupervised attribute reduction algorithm for mixed data. In order to fully consider the diversity of evaluation, multi-view measures are necessary to be proposed in attribute reduction.
The neighborhood rough set provides a flexible granular representation, but it requires determining the size of a neighborhood radius through grid search frequently, which is time-consuming. To overcome this problem, many strategies for determining the radius without parameters have been introduced. As an example, Xia et al. [25] put forward the concept of granular balls and enhanced the efficacy of a classifier based on granular computing by generating granular balls; Zhou et al. [26] proposed the concept of a gap neighborhood when resolving the problem of online feature selection, which can automatically determine the neighborhood size according to the distance difference between samples.
Through the above discussion, in order to effectively obtain the salient features of multiple views [25,26,27] and improve the classification performance of attribute reduction, we propose a new strategy in this paper: forward greedy searching to κ -reduct based on granular ball. The key of our strategy includes three phases: (1) grouping the samples in the whole universe (the universe is a finite set of all samples) based on generated symmetric granular balls; (2) attribute search for sample groups based on guidance; and (3) attribute reduction based on multiple perspectives. The first stage is to automatically create granular balls in accordance with the distribution of data itself, and merge granular balls which have small number of samples. Therefore, it can realize the division of the universe. In addition, the second stage is guidance-based evaluations, aiming to compress the search space related to candidate attributes. Therefore, the time needed for attribute reduction can be made faster because fewer candidate attributes need to be evaluated. Finally, the third stage blends supervised and unsupervised perspectives [28] and uses the quality-to-entropy ratio as a measure to attribute reduction. Therefore, it is feasible to identify attributes and labels with accuracy, and quantitatively characterize the uncertainty of data itself [29,30].
To sum up, the main contributions of our research are: (1) decreasing the size of samples by grouping different granular balls which adaptively generate; (2) enhancing attribute reduction efficiencies by achieving guidance-based search; and (3) utilizing the quality-to-entropy ratio which combines two perspectives to improve the accuracy of recognizing eligible attributes.
The remainder of this paper is organized as follows. Section 2 introduces the basic concepts of rough set, granular computing, and attribute reduction. Section 3 describes the fundamental framework and specific procedures of the new proposed method. Comparative experimental results of datasets and analysis are reported in Section 4. Finally, Section 5 is a summary of the algorithm and points for further work.

2. Preliminaries

2.1. Neighborhood Rough Set

Formally, a decision system can be defined as a binary group, expressed as D S = U , A T d : the universe of discourse U = { x 1 , x 2 , , x n } is a set of non-empty finite samples; A T = { a 1 , a 2 , a m } is the set of all conditional attributes; and d is the decision attribute. According to the decision values of all samples, it is not difficult to obtain a partition like U / I N D ( d ) = { X 1 , X 2 , , X n } which induced by decision attribute d on universe U: I N D ( d ) = { ( x i , x j ) U × U : d ( x i ) = d ( x j ) } ; x i U , d ( x i ) is the label of sample x i . It is especially worth noting that I N D ( d ) is a relation of equivalence with symmetry, reflexivity, and transitivity. The following definitions are the form of conventional rough sets.
Definition 1.
For a given decision system D S , a given radius δ 0 , A A T , x i U , the neighborhood of δ A ( x i ) is defined as:
δ A x i = x U : Δ A x i , x j δ ,
in which Δ A x i , x j represents the distance function between sample x i and x j with respect to A.
Immediately, from Definition 1, it can be obviously known that the size of the generated neighborhood relies on the value of given δ , i.e., the neighborhood becomes larger as the value of δ increases.
As the fundamental units of the neighborhood rough set, the specific definitions of upper and lower approximations are given in the following Definition 2.
Definition 2.
For a given decision system D S , X U , A A T , the lower and upper approximations of X are defined as
N ̲ X = x i U : δ A x i X ,
N ¯ X = x i U : δ A x i X .
The neighborhood rough set is built on the foundation of the standard rough set. It can not only deal with complex data, but also possess a multi-granularity structure by giving various radii. However, finding the appropriate radius generally requires a large number of trials or a certain parameter searching strategy, which is very time-consuming.

2.2. Granular Ball Computing

Considering that using the neighborhood relationship takes significant time to obtain the optimal radius, Xia et al. [31] proposed the concept of granular ball. In rough set theory, granule is the division of a sample set, and granular ball is based on the concept of granule. Xia et al. [25] regard hyper-ball with a completely symmetrical structure as granular ball.
The granular ball has a straightforward geometric shape with two parameters, i.e., center and radius. Compared with the neighborhood, the granular ball method has higher searching efficiency and robustness. The detailed definitions are as follows.
Definition 3.
For a given decision system D S , A A T , G B S U , G B S is a granular ball induced by conditional attribute set A if and only if C is the center point of G B S , r is the average of distances from all samples in the granular ball to C. The C and r of the granular ball are expressed as follows
C = 1 G B S i = 1 G B s x i ,
r = 1 G B S i = 1 G B s Δ A x i , C ,
in which G B S indicates the number of samples in the granular ball.
In the following, G B A U is defined as the set of all granular balls induced by conditional attribute set A on universe U.
Definition 4.
For a given decision system D S , A A T , G B S G B A U , d ( G B S ) is recorded as the overall label of G B S , i.e., d ( G B S ) is the label corresponding to samples with the same label and maximum proportion in the granular ball.
Definition 5.
For a given decision system D S , A A T , G B S G B A U , the average purity of G B S is defined as
P u r e ( G B s ) = { x i G B s : d ( x i ) = d ( G B s ) } G B s .
in which d ( x i ) indicates the label of the sample x i .
Furthermore, P u r e ( G B A U ) can be recorded as the mean purity of all granular balls induced by conditional attribute set A.
In the process of generating granular balls, the main idea is using an iterative two-means algorithm. The concrete procedures are given as follows.
(1) Consider the entire universe U as an initial granular ball and set n = 1 (n is the number of existing granular balls).
(2) Cluster each cluster by the two-means algorithm.
(3) Compute the center point of each cluster and the average distance between each cluster’s samples and the center point.
(4) Obtain the granular ball and calculate the granular ball’s purity.
(5) Traverse all currently existent granular balls; if each granular ball’s purity is below the given threshold, end this step; otherwise, return to (2).
On the basis of the aforementioned method of obtaining granular balls, Xia et al. [31] further put forward the concept of granular ball rough set, as shown in Definition 6.
Definition 6.
For a given decision system D S , A A T , X P U U I N D ( d ) I N D ( d ) , according to the conditional attribute set A, the upper and lower approximations of X p are, respectively, defined as
G B A U ¯ X p = x i : G B S G B A U , x i G B S , G B S X p ,
G B A U ̲ X p = x i : G B S G B A U , x i G B S , G B S X p .

2.3. Attribute Reduction

Rough set is a powerful tool to handle fuzzy data, and we need to deal with high-dimensional data through attribute reduction. By searching the minimum attribute subset which satisfies the given constraints, attribute reduction can not only reduce the dimension, but also enhance the generalization performance.
To date, various kinds of attribute reduction have been proposed for different requirements [9,10,11,12,13,18,32], whereas Yao et al. [33] indicated that the majority of them have analogous structures. There are two mainstream learning perspectives, i.e., supervised learning and unsupervised learning. Then, we pick the approximation quality [34] and conditional entropy [19,35,36,37,38,39] as two custom measures to better comprehend and investigate the essence of attribute reduction in terms of the neighborhood rough set.

2.3.1. Supervised Attribute Reduction

Supervised attribute reduction refers to the process of screening attributes using given labels in datasets so as to determine the important subsets of attributes which can best distinguish different categories.
Definition 7.
For a given decision system D S and a radius δ 0 , A A T , the supervised approximation quality of d in terms of A is defined as
γ A ( d ) = U k = 1 q δ A x k | U | ,
in which X is the cardinality of set X.
Apparently, it is not difficult to obtain that γ A ( d ) [ 0 , 1 ] holds. The approximation quality reflects the proportion of samples in the lower approximation of the decision class, and it is used to describe the dependency between attributes. Note that by Definition 7, the degree of dependency increases as the value of approximation quality increases. Generally speaking, the majority of samples in U can be told apart from each other.
Definition 8.
For a given decision system D S and a radius δ 0 , A A T , the supervised conditional entropy of d based on A is defined as
C E A ( d ) = 1 | U | x U δ A ( x ) [ x ] d log δ A ( x ) [ x ] d δ A ( x ) .
It is proven that C E A ( d ) 0 , | U | e holds [19]. As another important measure of the neighborhood rough set, conditional entropy reflects the discriminating performance of conditional attribute set A over decision attribute d. Following Definition 8, it is obvious that, as the value of the conditional entropy decreases, the discrimination of A relative to d increases.
Definition 9.
For a given decision system D S and a constraint condition C ρ U , which is associated with measure ρ on the universe U, A A T , A is deemed as a C ρ U -reduct if and only if
(1) 
A meets C ρ U ,
(2) 
A 1 A , A 1 does not meet C ρ U .
From Definition 9, it is uncomplicated to conclude that A is an ideal and minimal subset which satisfies the constraint condition. Without loss of generality, the constraint is closely related to the used measure. We will discuss it from the two following aspects:
(1)
If the measure is of approximation quality [28,40], the constraint condition may be γ A ( d ) γ A T ( d ) ;
(2)
If the measure is conditional entropy [41], the constraint condition may be C E A ( d ) C E A T ( d ) .

2.3.2. Unsupervised Attribute Reduction

As we all know, supervised attribute reduction depends on the labels of samples to a great extent, so it is time-consuming to obtain the labels of samples. However, unsupervised attribute reduction does not need obtain such labels.
In an unsupervised perspective, if approximate quality or conditional entropy is still needed as a measure, how to make labels for samples is an urgent problem. In order to solve the problem, Yang et al. [42] used the conditional attribute information of samples to construct pseudo-labels. Based on the pseudo-label strategy, it is not arduous to give the following definitions.
Definition 10.
For a given unsupervised decision I S and a radius δ 0 , A A T , a A , the unsupervised approximation quality in terms of A is defined as
γ A = 1 | A | a A γ A { a } d a ,
in which d a is a pseudo-label decision that records conditional attribute a to contain the pseudo-labels of samples.
In analogy with Definition 7, γ A [ 0 , 1 ] apparently holds. The approximate quality in Definition 10 represents the correlation between a set of attributes and a single attribute. Naturally, the higher the value of unsupervised approximation quality is, the greater the degree of such correlation.
Definition 11.
For a given unsupervised decision I S and a radius δ 0 , A A T , a A , the unsupervised conditional entropy with respect to A is defined as
C E A = 1 | A | a A C E A { a } d a ,
in which d a is a derived decision that employs conditional attribute a to contain the pseudo-labels of samples.
Similarly to Definition 8, the C E A 0 , | U | e constant holds in an unsupervised perspective. Undoubtedly, the certainty of the pseudo-label neighborhood judgment system increases as the value of the conditional entropy decreases.
Definition 12.
For a given unsupervised decision I S , a measure ρ and C ρ U is a constraint condition, A A T , A is deemed as a ρ-reduct if and only if:
(1) 
A meets the constraint C ρ U ;
(2) 
A 1 A , A 1 does not meet the constraint C ρ U .
Analogous to Definition 9, the constraint condition determined by ρ will depend on the type of measure. The constraint condition may be γ A γ A T if the unsupervised approximation quality is used as a measure; it may be C E A C E A T if the unsupervised conditional entropy is used as a measure.

3. Proposed Method

3.1. Theoretical Foundations

3.1.1. Quality-to-Entropy Ratio

Many researchers frequently use measures based on single view for attribute reduction, such as the one of supervised and unsupervised attribute reduction mentioned in Section 2.3. However, when considering only one perspective, some limitations may exist and crucial attributes cannot be effectively obtained.
Therefore, we propose a new measure which combines the supervised and unsupervised perspectives. Specifically, the supervised perspective selects the measure of approximation quality [40] and the unsupervised perspective uses conditional entropy [43]. The new measure can quantitatively describe the relationship between attributes and labels, and uncover the internal structure of data itself.
Evidently, from Definitions 7 and 11, it is known that the relationship between these two measures and the importance of attributes is completely opposite. Therefore, we adopt the form of ratio to unify the relationships. In addition, expressing the conditional entropy as an exponential function can significantly increase the relationship between conditional entropy and the importance of the attribute. The specific definition is as follows.
Definition 13.
For a given decision system D S and a radius δ 0 , A A T , the quality-to-entropy ratio is defined as
κ A ( d ) = γ A ( d ) exp C E A ,
in which γ A ( d ) is the approximation quality of d in terms of A as given in Definition 7 and C E A is the unsupervised conditional entropy over A as given in Definition 11.
According to the form of the quality-to-entropy ratio, when the value of γ A ( d ) is higher and the value of C E A is lower, the value of κ A ( d ) is higher. From Section 2.3.1 and Section 2.3.2, it can be seen that the higher the value of γ A ( d ) , the greater the influence of conditional attributes on the discriminant performance of decision d; the lower the value of C E A , the stronger the ability of distinguishing conditional attributes from pseudo-labels. Thus, the higher the value of κ A ( d ) , the stronger the discriminant ability of conditional attributes relative to decision making. To sum up, the conclusion is in line with the ultimate goal of attribute reductions.
Theorem 1.
For a given decision system D S , a radius δ 0 , A A T , κ A ( d ) [ 0 , 1 ] .
Proof. 
γ A ( d ) [ 0 , 1 ] holds in accordance with the property of approximate quality in Definition 7. Similarly, C E A 0 , | U | e holds in accordance with the property of conditional entropy in Definition 11, so we can infer that exp ( C E A ) [ 1 , e | U | e ] holds. Immediately, the quality-to-entropy ratio κ A ( d ) [ 0 , 1 ] holds. Specifically, if γ A ( d ) = 0 and exp ( C E A ) = e | U | e , κ a ( d ) = 0 ; if γ A ( d ) = 1 and exp ( C E A ) = 1 , κ a ( d ) = 1 .    □
Definition 14.
For a given decision system D S and a threshold θ [ 0 , 1 ] , A A T , A is defined as a κ-reduct if and only if
(1) 
κ A ( d ) κ A T ( d ) θ ;
(2) 
A 1 A , κ A 1 ( d ) κ A T ( d ) < θ .
By the above definition, it is observed that, as a minimal subset of attributes, κ -reduct improves the quality-to-entropy ratio. However, the issue of how such reduct can be found urgently needs to be solved. Generally speaking, we need to evaluate the significance of attributes in A T , eliminate low-quality attributes from the reduct pool, and select qualified attributes. Based on the greedy searching for attribute reduction [44,45,46], Definition 15 provides a significance attribute about our proposed quality-to-entropy ratio.
Definition 15.
For a given decision system D S , A A T , a A T A , we define the significance about the quality-to-entropy ratio as
S i g κ a ( d ) = κ A { a } ( d ) κ A ( d ) .
Obviously, S i g κ a ( d ) [ 0 , 1 ] holds. Definition 15 shows that the significance of conditional attributes increases as the value of S i g κ a ( d ) rises. Moreover, the attributes with high significance are likely to be selected and put into the reduct pool. For example, if we assume that S i g κ a 1 ( d ) < S i g κ a 2 ( d ) , a 1 , a 2 A T A , then κ A { a 1 } ( d ) < κ A { a 2 } ( d ) can be known. The result illustrates that we prefer a 2 to a 1 as an element in the candidate attribute subset.

3.1.2. Forward Greedy Searching to κ -Reduct Based on Granular Ball (GBFGS- κ )

Looking back on the above research, regardless of the searching methods mentioned in Section 2.3 or the quality-to-entropy ratio mentioned in Section 3.1.1, it is not difficult to obverse that: (1) searching methods obtain the information granulation based on the whole universe; and (2) in each iteration, the information granulation in the universe needs to be recalculated.
From the above reasons, the efficiency of searching methods depends on the size of the universe. Therefore, in this section, we will propose a forward greedy searching to κ -reduct based on granular ball. This strategy takes granular balls as groups to reduce the number of information granulation which need to be recalculated in each iteration. Note that Figure 1 and Figure 2 describe its basic framework.
The key steps of our proposed strategy will be explained as follows:
(1) The whole universe U is regarded as a granular ball. Cluster each granular ball by the two-means algorithm and repeatedly cluster until all balls’ purity are reached.
(2) Generate a new empty granular ball. Transfer samples into this empty granular ball if the number of samples in an existing granular ball is less than 4.
(3) Record the number of existing granular balls as n, and divide the whole universe U into n mutually asymmetric groups ( U 1 , U 2 , U n ). Granular balls correspond to groups one by one.
(4) Calculate the attribute reduction result A 1 on U 1 firstly by using the quality-to-entropy ratio. Then, the attribute reduction result A 2 on U 1 U 2 will be calculated on the basis of the result A 1 .
(5) When the constraints are satisfied, a reduction result A n is obtained which is the ultimate attribute reduction result of universe U.
Given the foregoing, it is not formidable to conclude that the forward greedy searching to the κ -reduct based on granular ball has the following benefits.
  • Reduction in the number of iterations and time consumption: the merging of granular balls with fewer samples into a new ball reduces the number of iterations required, which decreases the time consumption for the subsequent calculations.
  • Iterative refinement implemented in a part of the universe: based on the guiding idea, our strategy does not need to iteratively refine the information of the whole universe. This helps improve the efficiency of the attribute reduction process.
  • Consideration of diversity evaluation [47,48] and complex constraints [32]: our strategy combines supervised and unsupervised learning methods to identify more significant attributes and eliminate issues that may arise from a single perspective.

3.2. Detailed Algorithm

According to the discussion in Section 3.1, our new strategy firstly generates granular balls by iteratively applying two-means clustering on the whole dataset. Then, we merge granular balls with fewer samples. The specific algorithm of this step is given in Algorithm 1.   
Algorithm 1: Modified generation of granular balls.
Symmetry 15 00996 i001
   Secondly, we select any one of granular balls, quickly identify the data in this granular ball, and obtain the attributes which meet the constraints by κ -reduct. Algorithm 2 explains the above process to us.   
Algorithm 2: Forward greedy searching to κ -reduct (FGS- κ ).
Symmetry 15 00996 i002
   Finally, by constantly merging small universes, the reduction results obtained from previous merged small universes can be used as a guide for subsequent calculations. In other words, attributes are added on the basis of the results which obtained before merging, so as to improve the stability of reduction.
The complete algorithm of our proposed strategy is given in Algorithm 3.
Algorithm 3: Forward greedy searching to κ -reduct based on granular ball (GBFGS- κ ).
Symmetry 15 00996 i003
In line with the process of Algorithm 3, it is not difficult to calculate the time complexity of forward greedy searching to κ -reduct based on a granular ball.
The first and foremost, in the process of generating granular balls, the universe U will be clustered into G B A T U granular balls, resulting in a maximum of ( U × G B A T U ) iterations. There is one more point: when employing the κ -reduct method, pseudo-labels produced by k-means clustering have a time complexity of O ( k · T · U · A T ) , where k is the number of clusters and T is the total number of k-means iterations. Last but not the least, in the process of recursively calculating κ A ( d ) , the worst case may require adding all of the attributes in A T to the reduct pool. Obviously, the total number of times of traversing κ A ( d ) will reach A T + A T 1 + A T 2 + + 2 + 1 . In summary, the time complexity of Algorithm 3 is U × G B A T U + A T × A T 3 .

4. Experimental Analysis

4.1. Datasets

We use 16 UCI datasets for verification to demonstrate the effectiveness of our forward greedy searching to κ -reduct based on the granular ball (GBFGS- κ ). Table 1 provides a thorough explanation of the various datasets.

4.2. Experimental Configuration

All experiments were conducted on a personal computer with Windows 10, Intel Core i7-10510U CPU(2.30 GHz) and 8.00 GB memory. The programming environment is MATLAB R2020a.
In the following experiment, the two-means algorithm was used to iteratively create granular balls, k-means clustering [44,49] was utilized to create pseudo-labels of samples, and quality-to-entropy ratio was the measure used in attribute reduction. It is rather remarkable that the value of k should be consistent with the number of decision classes in the data. In addition, the result of the neighborhood rough set largely depends on the given radius. In order to demonstrate the applicability and universality of our proposed method, all experiments employed 20 radii with a step size of 0.02, which are 0.02, 0.04, …, 0.40.
Moreover, the deduction simplification process was verified by 10-fold cross-validation. That is to say, for each radius, the samples in universe U were divided into ten groups, i.e., U 1 , U 2 , U 10 , then, nine of them were used as training groups and the rest was used as the test group. Repeat 10-fold cross-validation process for 10 times to ensure each group serves as a test group, so as to test the classification performance and obtain a reliable and stable model.
Finally, we used K-nearest neighbor (KNN, K = 3) [50,51], support vector machine (SVM) [52] and classification and regression tree (CART) [53] to compare our proposed method with six progressive algorithms in terms of attribute reduction as well as with the algorithm without applying any attribute reduction methods (No-reduct classification). The performance of the derived reducers was mainly tested from the aspects of classification stability, classification accuracy, reduced stability, and elapsed time. The attribute reduction algorithms used for comparison are as follows:
(1) Dissimilarity Based Searching for Attribute Reduction (DBSAR) [54];
(2) Knowledge Change Rate (KCR) [55];
(3) Attribute Group (AG) [56];
(4) Ensemble Selector For Attribute Reduction (ESAR) [47];
(5) Multi-criterion Neighborhood Attribute Reduction (MNAR) [32];
(6) Robust Attribute Reduction Based On Rough Sets (RARR) [57].

4.3. Comparison of Classification Accuracy

In this section, we will use KNN, SVM and CART to predict the test samples to weigh up the classification accuracy of each algorithm. Immediately, for attribute reduction algorithms, given a decision system D S , the classification accuracy applied to reduction is defined as
A c c r e d = { x i r e d | P r e r e d ( x i ) = d ( x i ) } U ,
in which P r e r e d ( x i ) is the prediction label made using reduct r e d for x i .
Table 2 displays the detailed classification accuracy results for each algorithm on 16 datasets and Figure 3 illustrates the radar charts for each dataset under the three classifiers with three different colors. The following conclusions can easily be reached by observing Table 2 and Figure 3.
(1)
For the majority of datasets, regardless of whether the KNN, SVM, or CART classifier is used, the classification accuracies related to GBSAR- κ outperform other comparison algorithms. Taking the dataset “Parkinson Speech (ID: 7)” as an example, when KNN classifier is adopted, the classification accuracies of GBSAR- κ , DBSAR, BCKCR, AG, ESAR, MMAR, RAAR, and No-reduct classification each are 0.7259, 0.7063, 0.7008, 0.7093, 0.7031, 0.7253, 0.7095, and 0.6984, respectively; when using the SVM classifier, the classification accuracies of GBSAR- κ , DBSAR, BCKCR, AG, ESAR, MMAR, RAAR, and No-reduct classification are 0.6661, 0.6532, 0.6521, 0.6548, 0.6543, 0.6639, 0.6539, and 0.6488, respectively; by employing CART, the classification accuracies of GBSAR- κ , DBSAR, BCKCR, AG, ESAR, MMAR, RAAR, and No-reduct classification are 0.6433, 0.6429, 0.6420, 0.6413, 0.6424, 0.6307, 0.6421, and 0.6419 respectively. Therefore, the simplification derived from our GBSAR- κ can offer an effective categorization performance.
(2)
From the average classification accuracy of each algorithm, the classification accuracy associated with GBSAR- κ is comparable or even more significant than that of DBSAR, BCKCR, AG, ESAR, MMAR, RAAR, and No-reduct classification. When using the KNN classifier, GBSAR- κ ’s classification accuracy is 0.8258, which is at most 32.21% higher than those of others; when the SVM classifier is utilized, GBSAR- κ ’s classification accuracy is 0.7903, which is at most 34.12% higher than those of others; by employing CART, GBSAR- κ ’s classification accuracy is 0.8090, which is at most 27.35% higher than those of others.

4.4. Comparison of Classification Stability

In this section, similarly to Section 4.3, we will evaluate the classification stability of each algorithm under KNN, SVM, and CART based on six advanced attribute reduction algorithms and a classification algorithm without applying any attribute reduction. Higher values of classification stability imply that the predicted label result is more stable and less susceptible to interference from the training samples.
Following the use of three classifiers on 16 datasets, Table 3 and Figure 4 show the classification stability findings of each algorithm. The following conclusions can easily be drawn in Table 3 and Figure 4.
(1)
For most datasets, our GBSAR- κ algorithm plays a leading role in classification stability compared with other algorithms. Moreover, predictions based on the features related to GBSAR- κ gain absolute advantages for some datasets. Consider the dataset “Twonorm (ID: 12)” as an example: when the KNN classifier is used, the classification accuracies of GBSAR- κ , DBSAR, BCKCR, AG, ESAR, MMAR, RAAR, and No-reduct classification are 0.9300, 0.8934, 0.8747, 0.8744, 0.8809, 0.5300, 0.7139, and 0.8772, respectively; when adopting the SVM classifier, the classification accuracies of GBSAR- κ , DBSAR, BCKCR, AG, ESAR, MMAR, RAAR, and No-reduct classification are 0.9693, 0.9333, 0.9140, 0.9116, 0.9224, 0.5582, 0.8164, and 0.9458, respectively; when using CART, the classification accuracies of GBSAR- κ , DBSAR, BCKCR, AG, ESAR, MMAR, RAAR, and No-reduct classification are 0.7723, 0.7600, 0.7531, 0.7491, 0.7564, 0.5216, 0.6803, and 0.7512, respectively. Therefore, from the standpoint of classifier stability, GBSAR- κ can indeed provide a more stable classification performance.
(2)
In terms of average classification accuracy, the classification stability connected with GBSAR- κ is far superior to the other algorithms. Moreover, when employing KNN classifier, the classification stability of GBSAR- κ is 0.9015, which is at most 24.52% higher than those of other methods; the classification stability of GBSAR- κ using SVM classifier is 0.9336, which is at most 14.67% higher than those of others; the classification stability of GBSAR- κ through the use of CART classifier is 0.8301, which is at most 12.92% higher than those of others.

4.5. Comparison of Reduced Stability

In this section, we will show the reduced stability of the attribute reduction corresponding to 16 datasets. The specific results are given in Table 4.
The information shown in Table 4 indicates that the reduced stability of GBSAR- κ is slightly lower than RARR, but still in a leading position. Obviously, compared with DBSAR, BCKCR, AG, ESAR, and MMAR, the average reduced stability value of GBSAR- κ is increased by 17.21%, 27.74%, 46.53%, 10.99%, and 111.21%, while it only decreases by 0.93% compared with RARR.
In general, although the reduced stability of our GBSAR- κ is not inferior to the result of RAAR for many datasets, its result is better than the six advanced algorithms in terms of attribute reduction in some cases. For instance, as far as the dataset “Climate Model Simulation Crashes (ID: 2)” is concerned, the reduced stabilities of GBSAR- κ , DBSAR, BCKCR, AG, ESAR, MMAR, and RAAR are 0.6605, 0.3265, 0.5284, 0.3483, 0.5814, 0.0545, and 0.3773, respectively. Compared with other algorithms, the result of GBSAR- κ is improved by 102.29%, 25.00%, 89.64%, 13.61%, 1111.93%, and 75.06%, respectively.
Therefore, it should be pointed out that using GBSAR- κ is more conducive to selecting attributes which are more suitable for sample changes.

4.6. Comparisons of Elapsed Time

In this section, we will compare the time taken to derive a simplification using different algorithms. The detailed results are reported in Table 5.
Following a thorough analysis of Table 5, it is not difficult to come to the findings that are listed below.
Considering the reduced stability mentioned in Section 4.5 and the reduced length conflict with each other, it can be concluded that the higher the value of reduced stability, the longer the reduced length. Apparently, the reduced length of GBSAR- κ is longer, which indicates that, in the simplification process, our algorithm needs to be strengthened in terms of the time speed.
From the view of the average elapsed time, it is worth mentioning that the value of GBSAR- κ is 56.42% and 65.42% lower than BCKCR and RARR, respectively. Taking the dataset “Pen-Based Recognition of Handwritten Digits (ID: 8)” as an example, the speed-up ratios of GBSAR- κ algorithm reached 0.9777, 7.6960, 1.3714, 1.9355, 0.0891, and 8.4364, respectively, when the elapsed times of GBSAR- κ , DBSAR, BCKCR, AG, ESAR, MMAR, and RAAR each are 175.1957, 171.2967, 1348.3062, 240.2581, 339.0988, 15.6138, and 1478.0267 s. Therefore, the elapsed time of GBSAR- κ for attribute reduction is lower than that of AG and ESAR under some circumstances.
From the above discussion, it is observed that, even though the elapsed time of our new algorithm is better than BCKCR and RARR in some datasets, GBSAR- κ ’s speed performance still has to be improved.

5. Conclusions and Future Perspectives

In this paper, we propose a new searching strategy that differs from conventional algorithms in the following aspects. On the one hand, by automatically generating granular balls, there is no time consumption for radius optimization. On the other hand, guidance-based searching is designed to compress the attribute searching space. In addition, the quality-to-entropy ratio can overcome the limitations and predictability of the single-attribute measure method.
Through experiments on 16 UCI datasets, it is not formidable to reveal that our proposed strategy has quite a positive classification performance and strong stability in the process of exporting reduction.
Further research can be conducted for the two following aspects:
(1)
Using the fused measure may increase the time of selecting the best attribute. Therefore, more accelerators [11] can be added to further improve the efficiency and reduce the time consumption.
(2)
The searching strategy proposed in this paper is a general module. Therefore, other measures based on the rough set can be substituted for the quality-to-entropy ratio, so as to compare the classification performance under various measures.

Author Contributions

Conceptualization, J.C.; methodology, J.C.; software, M.S.; validation, J.S.; formal analysis, Y.F.; investigation, T.X.; resources, J.S.; data curation, T.X.; writing—original draft preparation, M.S.; writing—review and editing, T.X.; visualization, M.S.; supervision, J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Industry-school Cooperative Education Program of the Ministry of Education (Grant No. 202101363034), the Natural Science Foundation of Jiangsu Higher Education (Grant No. 17KJB520007), and the Key Research and Development Program of Zhenjiang-Social Development (Grant No. SH2018005).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pawlak, Z.; Skowron, A. Rudiments of rough sets. Inf. Sci. 2007, 177, 3–27. [Google Scholar] [CrossRef]
  2. Dash, M.; Liu, H. Consistency-based search in feature selection. Artif. Intell. 2003, 151, 155–176. [Google Scholar] [CrossRef]
  3. Dowlatshahi, M.; Derhami, V.; Nezamabadi-pour, H. Ensemble of Filter-Based Rankers to Guide an Epsilon-Greedy Swarm Optimizer for High-Dimensional Feature Subset Selection. Information 2017, 8, 152. [Google Scholar] [CrossRef]
  4. Chen, H.M.; Li, T.R.; Luo, C.; Horng, S.J.; Wang, G.Y. A Decision-Theoretic Rough Set Approach for Dynamic Data Mining. IEEE Trans. Fuzzy Syst. 2015, 23, 1958–1970. [Google Scholar] [CrossRef]
  5. Xu, T.H.; Wang, G.Y.; Yang, J. Finding strongly connected components of simple digraphs based on granulation strategy. Int. J. Approx. Reason. 2020, 118, 64–78. [Google Scholar] [CrossRef]
  6. Qian, Y.; Liang, J.Y.; Pedrycz, W.; Dang, C. An efficient accelerator for attribute reduction from incomplete data in rough set framework. Pattern Recognit. 2011, 44, 1658–1670. [Google Scholar] [CrossRef]
  7. Cheng, K.; Gao, S.; Dong, W.L.; Yang, X.B.; Wang, Q.; Yu, H.L. Boosting label weighted extreme learning machine for classifying multi-label imbalanced data. Neurocomputing 2020, 403, 360–370. [Google Scholar] [CrossRef]
  8. Akram, M.; Nawaz, H.S.; Deveci, M. Attribute reduction and information granulation in Pythagorean fuzzy formal contexts. Expert Syst. Appl. 2023, 222, 119794. [Google Scholar] [CrossRef]
  9. Jiang, Z.H.; Yang, X.B.; Yu, H.L.; Liu, D.; Wang, P.X.; Qian, Y. Accelerator for multi-granularity attribute reduction. Knowl. Based Syst. 2019, 177, 145–158. [Google Scholar] [CrossRef]
  10. Chen, Y.; Yang, X.B.; Li, J.H.; Wang, P.X.; Qian, Y.H. Fusing attribute reduction accelerators. Inf. Sci. 2021, 587, 354–370. [Google Scholar] [CrossRef]
  11. Wang, P.X.; Yang, X.B. Three-Way Clustering Method Based on Stability Theory. IEEE Access 2021, 9, 33944–33953. [Google Scholar] [CrossRef]
  12. Chen, Q.; Xu, T.H.; Chen, J.J. Attribute Reduction Based on Lift and Random Sampling. Symmetry 2022, 14, 1828. [Google Scholar] [CrossRef]
  13. Chen, Y.; Wang, P.X.; Yang, X.B.; Mi, J.S.; Liu, D. Granular ball guided selector for attribute reduction. Knowl. Based Syst. 2021, 229, 107326. [Google Scholar] [CrossRef]
  14. Chen, Z.; Liu, K.Y.; Yang, X.B.; Fujita, H. Random sampling accelerator for attribute reduction. Int. J. Approx. Reason. 2021, 140, 75–91. [Google Scholar] [CrossRef]
  15. Hu, Q.H.; Pedrycz, W.; Yu, D.R.; Lang, J. Selecting Discrete and Continuous Features Based on Neighborhood Decision Error Minimization. IEEE Trans. Syst. Man Cybern. B Cybern. 2010, 40, 137–150. [Google Scholar] [CrossRef] [PubMed]
  16. Liang, J.Y.; Chin, K.S.; Dang, C.Y.; Yam, R.C.M. A new method for measuring uncertainty and fuzziness in rough set theory. Int. J. Gen. Syst. 2002, 31, 331–342. [Google Scholar] [CrossRef]
  17. Liang, J.Y.; Shi, Z.; Li, D.Y.; Wierman, M.J. Information entropy, rough entropy and knowledge granulation in incomplete information systems. Int. J. Gen. Syst. 2006, 35, 641–654. [Google Scholar] [CrossRef]
  18. Liu, K.Y.; Yang, X.B.; Yu, H.L.; Fujita, H.; Chen, X.J.; Liu, D. Supervised information granulation strategy for attribute reduction. Int. J. Mach. Learn. Cybern. 2020, 11, 2149–2163. [Google Scholar] [CrossRef]
  19. Zhang, X.; Mei, C.L.; Chen, D.G.; Li, J.H. Feature selection in mixed data: A method using a novel fuzzy rough set-based information entropy. Pattern Recognit. 2016, 56, 1–15. [Google Scholar] [CrossRef]
  20. Yao, Y.Y.; Zhang, X.Y. Class-specific attribute reducts in rough set theory. Inf. Sci. 2017, 418, 601–618. [Google Scholar] [CrossRef]
  21. Gao, Y.; Chen, X.J.; Wang, P.X. Attribute reduction over consistent samples. CAAI Trans. Int. Syst. 2019, 14, 1170–1178. [Google Scholar] [CrossRef]
  22. Jiang, Z.H.; Liu, K.Y.; Yang, X.B.; Yu, H.L.; Fujita, H.; Qian, Y.H. Accelerator for supervised neighborhood based attribute reduction. Int. J. Approx. Reason. 2020, 119, 122–150. [Google Scholar] [CrossRef]
  23. Zhang, W.; Miao, D.; Gao, C.; Li, F. Rough Set Attribute Reduction Algorithm for Partially Labeled Data. Comput. Sci. 2017, 44, 25–31. [Google Scholar] [CrossRef]
  24. Yuan, Z.; Chen, H.M.; Li, T.R.; Yu, Z.; Sang, B.B.; Luo, C. Unsupervised attribute reduction for mixed data based on fuzzy rough sets. Inf. Sci. 2021, 572, 67–87. [Google Scholar] [CrossRef]
  25. Xia, S.Y.; Liu, Y.S.; Ding, X.; Wang, G.Y.; Yu, H.; Luo, Y.G. Granular ball computing classifiers for efficient, scalable and robust learning. Inf. Sci. 2019, 483, 136–152. [Google Scholar] [CrossRef]
  26. Zhou, P.; Hu, X.G.; Li, P.P.; Wu, X.D. Online streaming feature selection using adapted Neighborhood Rough Set. Inf. Sci. 2019, 481, 258–279. [Google Scholar] [CrossRef]
  27. Hu, Q.H.; Yu, D.R.; Xie, Z.X. Neighborhood classifiers. Expert Syst. Appl. 2008, 34, 866–876. [Google Scholar] [CrossRef]
  28. Liu, K.Y.; Yang, X.B.; Yu, H.L.; Mi, J.S.; Wang, P.X.; Chen, X.J. Rough set based semi-supervised feature selection via ensemble selector. Knowl. Based Syst. 2019, 165, 282–296. [Google Scholar] [CrossRef]
  29. Qian, Y.H.; Cheng, H.H.; Wang, J.T.; Liang, J.Y.; Pedrycz, W.; Dang, C.Y. Grouping granular structures in human granulation intelligence. Inf. Sci. 2017, 382-383, 150–169. [Google Scholar] [CrossRef]
  30. Yang, X.B.; Qian, Y.H.; Yang, J.Y. On Characterizing Hierarchies of Granulation Structures via Distances. Fundam. Inform. 2013, 123, 365–380. [Google Scholar] [CrossRef]
  31. Xia, S.Y.; Zhang, Z.; Li, W.H.; Wang, G.Y.; Giem, E.; Chen, Z.Z. GBNRS: A Novel Rough Set Algorithm for Fast Adaptive Attribute Reduction in Classification. IEEE Trans. Knowl. Data Eng. 2020, 34, 1231–1242. [Google Scholar] [CrossRef]
  32. Li, J.Z.; Yang, X.B.; Song, X.N.; Li, J.H.; Wang, P.X.; Yu, D.J. Neighborhood attribute reduction: A multi-criterion approach. Int. J. Mach. Learn. Cybern. 2019, 10, 731–742. [Google Scholar] [CrossRef]
  33. Yao, Y.Y.; Zhao, Y.; Wang, J.; Han, S.Q. A Model of User-Oriented Reduct Construction for Machine Learning. Trans. Rough Sets 2008, 8, 332–351. [Google Scholar] [CrossRef]
  34. Ju, H.R.; Yang, X.B.; Yu, H.L.; Li, T.J.; Yu, D.J.; Yang, J.Y. Cost-sensitive rough set approach. Inf. Sci. 2016, 355–356, 282–298. [Google Scholar] [CrossRef]
  35. Dai, J.H.; Xu, Q.; Wang, W.T.; Tian, H.W. Conditional entropy for incomplete decision systems and its application in data mining. Int. J. Gen. Syst. 2012, 41, 713–728. [Google Scholar] [CrossRef]
  36. Xu, J.C.; Yang, J.; Ma, Y.Y.; Qu, K.L.; Kang, Y.H. Feature selection method for color image steganalysis based on fuzzy neighborhood conditional entropy. Appl. Intell. 2022, 52, 9388–9405. [Google Scholar] [CrossRef]
  37. Sang, B.B.; Chen, H.M.; Yang, L.; Li, T.R.; Xu, W.H. Incremental Feature Selection Using a Conditional Entropy Based on Fuzzy Dominance Neighborhood Rough Sets. IEEE Trans. Fuzzy Syst. 2021, 30, 1683–1697. [Google Scholar] [CrossRef]
  38. Américo, A.; Khouzani, M.; Malacaria, P. Conditional Entropy and Data Processing: An Axiomatic Approach Based on Core-Concavity. IEEE Trans. Inform. Theory 2020, 66, 5537–5547. [Google Scholar] [CrossRef]
  39. Gao, C.; Zhou, J.; Miao, D.; Yue, X.D.; Wan, J. Granular conditional entropy-based attribute reduction for partially labeled data with proxy labels. Inf. Sci. 2021, 580, 111–128. [Google Scholar] [CrossRef]
  40. Hu, Q.H.; Liu, J.F.; Yu, D.R. Mixed feature selection based on granulation and approximation. Knowl. Based Syst. 2008, 21, 294–304. [Google Scholar] [CrossRef]
  41. Mohseni, M.; Redies, C.; Gast, V. Approximate Entropy in Canonical and Non-Canonical Fiction. Entropy 2022, 24, 278. [Google Scholar] [CrossRef]
  42. Yang, X.B.; Liang, S.C.; Yu, H.L.; Gao, S.; Qian, Y. Pseudo-label neighborhood rough set: Measures and attribute reductions. Int. J. Approx. Reason. 2019, 105, 112–129. [Google Scholar] [CrossRef]
  43. Li, W.Y.; Chen, H.M.; Li, T.R.; Wan, J.H.; Sang, B.B. Unsupervised feature selection via self-paced learning and low-redundant regularization. Knowl. Based Syst. 2021, 240, 108150. [Google Scholar] [CrossRef]
  44. Wang, P.X.; Shi, H.; Yang, X.B.; Mi, J.S. Three-way k-means: Integrating k-means and three-way decision. Int. J. Mach. Learn. Cybern. 2019, 10, 2767–2777. [Google Scholar] [CrossRef]
  45. Wu, T.F.; Fan, J.C.; Wang, P.X. An Improved Three-Way Clustering Based on Ensemble Strategy. Mathematics 2022, 10, 1457. [Google Scholar] [CrossRef]
  46. Ba, J.; Wang, P.X.; Yang, X.B.; Yu, H.L.; Yu, D.J. Glee: A granularity filter for feature selection. Eng. Appl. Artif. Intell. 2023, 122, 106080. [Google Scholar] [CrossRef]
  47. Yang, X.B.; Yao, Y.Y. Ensemble selector for attribute reduction. Appl. Soft Comput. 2018, 70, 1–11. [Google Scholar] [CrossRef]
  48. Chen, Y.N.; Wang, P.X.; Yang, X.B.; Yu, H.L. Bee: Towards a robust attribute reduction. Int. J. Mach. Learn. Cybern. 2022, 13, 3927–3962. [Google Scholar] [CrossRef]
  49. Wang, P.X.; Yao, Y.Y. CE3: A three-way clustering method based on mathematical morphology. Knowl. Based Syst. 2018, 155, 54–65. [Google Scholar] [CrossRef]
  50. Cover, T.M.; Hart, P.E. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  51. Fukunaga, K.; Narendra, P.M. A Branch and Bound Algorithm for Computing k-Nearest Neighbors. IEEE. Trans. Comput. 1975, C-24, 750–753. [Google Scholar] [CrossRef]
  52. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27:1–27:27. [Google Scholar] [CrossRef]
  53. Krzywinski, M.; Altman, N. Classification and Regression Trees. Nat. Methods 2017, 14, 755–756. [Google Scholar] [CrossRef]
  54. Rao, X.S.; Yang, X.B.; Yang, X.; Chen, X.J.; Liu, D.; Qian, Y. Quickly calculating reduct: An attribute relationship based approach. Knowl. Based Syst. 2020, 200, 106014. [Google Scholar] [CrossRef]
  55. Jin, C.X.; Li, F.C.; Hu, Q.H. Knowledge change rate-based attribute importance measure and its performance analysis. Knowl. Based Syst. 2017, 119, 59–67. [Google Scholar] [CrossRef]
  56. Chen, Y.; Liu, K.Y.; Song, J.J.; Fujita, H.; Yang, X.B.; Qian, Y. Attribute group for attribute reduction. Inf. Sci. 2020, 535, 64–80. [Google Scholar] [CrossRef]
  57. Dong, L.J.; Chen, D.G.; Wang, N.L.; Lu, Z.H. Key energy-consumption feature selection of thermal power systems based on robust attribute reduction with rough sets. Inf. Sci. 2020, 532, 61–71. [Google Scholar] [CrossRef]
Figure 1. The structure of the granular ball part.
Figure 1. The structure of the granular ball part.
Symmetry 15 00996 g001
Figure 2. The structure of the κ -reduct part.
Figure 2. The structure of the κ -reduct part.
Symmetry 15 00996 g002
Figure 3. Classification accuracies of three classifiers.
Figure 3. Classification accuracies of three classifiers.
Symmetry 15 00996 g003
Figure 4. Classification stabilities of three classifiers.
Figure 4. Classification stabilities of three classifiers.
Symmetry 15 00996 g004
Table 1. Dataset descriptions.
Table 1. Dataset descriptions.
IDDatasetsSamplesAttributesLabels
1Cardiotocography21262110
2Climate Model Simulation Crashes540202
3Diabetic Retinopathy Debrecen1151192
4Forest Type Mapping523274
5Ionosphere351342
6Libras Movement3609015
7Parkinson Speech1208262
8Pen-Based Recognition of Handwritten Digits10,9921610
9Statlog (Image Segmentation)2310187
10Statlog (Landsat Satellite)6435367
11Statlog (Vehicle Silhouettes)846184
12Twonorm7400192
13Ultrasonic Flowmeter Diagnostics-Meter D180434
14Urban Land Cover6751479
15Wall-Following Robot Navigation5456243
16Wisconsin Diagnostic Breast Cancer569302
Table 2. The comparisons of the classification accuracies.
Table 2. The comparisons of the classification accuracies.
ClassifierIDGBSAR- κ DBSARBCKCRAGESARMNARRARRNo-Reduct
Classification
KNN10.76540.75260.75760.76330.76020.39950.75320.7318
20.92660.90800.92650.91200.90410.89260.89420.9059
30.61470.60780.61440.61140.61030.51040.61210.6061
40.87290.86700.86040.87280.86490.70860.86840.8476
50.89470.86990.87900.88690.88640.70860.89440.8714
60.69760.67460.65850.66740.65780.65830.53490.6544
70.72590.70630.70080.70930.70310.72530.70950.6984
80.97420.96700.96580.96830.96570.27320.20930.9605
90.94440.94350.94180.94130.93430.52680.15890.9367
100.88970.88970.88900.89340.88900.51760.49290.8883
110.65460.64640.63860.65280.64090.47220.61390.6323
120.91310.89970.89580.89790.89710.60340.60650.8954
130.83910.76580.77670.80140.77000.78890.83890.7681
140.68010.67020.67010.69290.70040.59410.68790.7556
150.85960.86190.86900.86700.86310.73140.57950.8534
160.96040.95490.95110.95330.95500.88250.96020.9499
Average0.82580.81160.81220.81820.81260.62460.65090.8097
↑1.75%↑1.68%↑0.93%↑1.62%↑32.21%↑26.87%↑1.98%
SVM10.77230.74600.74520.75420.75340.45790.75640.7406
20.92080.90890.92000.91040.90740.90740.90770.9092
30.63710.62800.62270.62850.62790.54350.63560.6210
40.87170.86710.85480.86710.86500.64190.86350.8381
50.89390.83690.83660.86440.88260.80000.89370.8014
60.54540.47360.47170.46970.49220.43060.36790.4567
70.66610.65320.65210.65480.65430.66390.65390.6488
80.93530.91280.91450.91920.91860.32330.30150.9041
90.92790.91680.91490.91790.91050.47400.14980.9102
100.86150.85720.86120.86110.85790.57500.57400.8466
110.63420.62300.60460.63130.61410.39880.59340.5872
120.94010.92660.92170.92390.92480.66240.67730.9196
130.63910.58330.56170.59110.59610.58330.63890.6478
140.73870.72690.75370.76130.73990.48890.72100.7267
150.68080.66350.65720.67550.66020.60200.45740.6575
160.98050.96880.96410.96450.96890.87540.97940.9602
Average0.79030.76830.76600.77470.77340.58930.63570.7610
↑2.87%↑3.17%↑2.02%↑2.20%↑34.12%↑24.32%↑3.86%
CART10.80900.79460.80220.80460.80210.46730.79780.7906
20.92800.90710.92750.91390.89360.86110.87680.8982
30.61570.60660.60520.60940.61370.57830.60470.6017
40.81000.80130.79430.80960.80740.71240.80140.8005
50.88330.86510.86690.86640.87930.81140.88270.8613
60.50380.49390.49080.48850.47180.45280.41560.4861
70.64330.64290.64200.64130.64240.63070.64210.6419
80.92700.91860.92440.92610.92580.35390.30450.9218
90.95060.95020.94650.95010.94440.56540.14980.9416
100.85130.84680.84930.85100.84510.58480.59100.8400
110.67940.66030.65150.67480.65720.50770.63910.6641
120.82050.81600.81600.81600.81680.58860.59660.8091
130.87820.84670.83390.85310.81830.80560.87780.8667
140.73110.71990.75670.76700.76860.63410.75830.7282
150.98470.98300.98240.98440.98190.74020.62160.9817
160.92860.92380.92830.91960.92500.87020.90870.9174
Average0.80900.79860.80110.80470.79960.63530.65430.7969
↑1.31%↑0.99%↑0.53%↑1.18%↑27.35%↑23.65%↑1.52%
Table 3. The comparisons of classification stabilities.
Table 3. The comparisons of classification stabilities.
ClassifierIDGBSAR- κ DBSARBCKCRAGESARMNARRARRNo-Reduct
Classification
KNN10.88550.88220.87500.85400.87080.67060.87620.8719
20.97070.93950.93590.93080.94580.97040.96580.9380
30.79430.79330.76830.77260.78140.47130.78900.7660
40.93380.92950.91030.92320.92460.69900.92850.9017
50.92230.86610.86330.87040.89360.60570.93630.8601
60.80320.75960.75560.75380.78260.76110.70190.7517
70.81650.82490.80680.80620.81420.85730.82740.8046
80.97870.96770.96640.95910.96280.69050.86020.9555
90.99250.97700.97380.96880.96610.59480.99090.9678
100.94610.94270.94280.93770.93880.88240.83780.9360
110.83580.82520.81560.81440.82470.63790.83500.8138
120.93000.89340.87470.87440.88090.53000.71390.8772
130.94530.85310.87440.87080.89970.79440.94440.8544
140.79760.76570.74930.75270.78660.65190.81010.7554
150.90610.90660.90230.89040.90010.93510.90430.9010
160.96580.95990.95810.95300.96180.83160.96540.9542
Average0.90150.88040.87330.87080.88340.72400.86800.8642
↑2.40%↑3.23%↑3.53%↑2.05%↑24.52%↑3.87%↑3.70%
SVM11.00001.00001.00001.00001.00001.00001.00000.9722
20.99740.99730.98450.99441.00001.00000.99940.9870
30.92440.93220.92200.92070.92611.00000.92390.9160
40.96690.96450.95350.95650.96380.77520.96540.9586
50.93370.87100.89670.89770.93130.91430.95860.9543
60.79250.75180.76080.74890.77280.68610.73930.7717
70.96180.93610.93810.92590.94200.96100.94540.9396
80.97080.94770.95000.92660.94580.64660.95660.9632
90.98060.97960.97480.96770.96400.72810.74550.9699
100.97380.96970.97320.96700.96740.86780.88760.9680
110.89170.88510.87700.87470.87820.65800.87730.8568
120.96930.93330.91400.91160.92240.55820.81640.9458
130.77810.75750.74750.75420.75750.75560.77780.6333
140.87530.86670.85740.86610.84760.61330.90490.8959
150.92460.93750.93100.90140.92860.99911.00000.9434
160.99680.98940.98610.98170.98570.86320.99540.9865
Average0.93360.92000.91670.91220.92080.81420.90580.9114
↑1.48%↑1.85%↑2.35%↑1.39%↑14.67%↑3.06%↑1.88%
CART10.83660.82680.83580.80810.82820.70400.81500.8055
20.92450.90130.92030.90530.90310.88520.89930.9037
30.64640.64400.64010.63710.63470.53300.63940.6390
40.90160.89050.87260.88790.89470.77710.90130.8776
50.91440.83300.86790.85140.89390.91430.90360.8600
60.64680.63960.66890.64720.66310.65560.65830.6028
70.64190.64110.63820.63380.63920.63150.63750.6259
80.92580.90870.91400.90400.91580.66730.94990.9088
90.94840.94790.94220.94530.93900.62640.77320.9383
100.85820.85370.85790.85650.85570.86620.83680.8457
110.73630.71290.69800.72430.72750.67930.71210.7244
120.77230.76000.75310.74910.75640.52160.68030.7512
130.88360.84310.83690.84720.84060.82780.88330.8356
140.72210.74300.75450.75520.74990.74520.77600.7628
150.98110.98080.97960.98050.97910.93330.89330.9771
160.94170.93120.94110.92610.92670.79470.92460.9298
Average0.83010.81610.82010.81620.82170.73520.80520.8118
↑1.72%↑1.23%↑1.71%↑1.02%↑12.92%↑3.09%↑2.26%
Table 4. The reduced stabilities of deriving reducts.
Table 4. The reduced stabilities of deriving reducts.
IDGBSAR- κ DBSARBCKCRAGESARMNARRARR
10.91450.92800.83270.82580.87070.20000.9916
20.66050.32650.52840.34830.58140.05450.3773
30.92700.93560.78680.80320.89790.40000.9223
40.78160.74370.60170.60370.71640.16010.8677
50.57600.36470.29020.29400.50030.63080.9277
60.65870.20590.15350.12320.60330.12750.4888
70.82450.86870.76590.72810.80071.00000.9045
80.90540.76720.71550.59620.75020.30001.0000
90.91540.89680.86980.67120.79450.30001.0000
100.86090.77230.78830.65490.74420.60000.4000
110.82460.82800.80330.72610.81500.40000.8588
120.90060.71540.59580.58180.63740.14980.6000
130.85050.64660.62120.42540.92100.20510.9506
140.55770.38130.29960.17640.53460.49330.7277
150.89600.87880.79550.71830.82571.00001.0000
160.75880.67160.58240.46750.55060.04540.9157
Average0.80080.68320.62690.54650.72150.37920.8083
↑17.21%↑27.74%↑46.53%↑10.99%↑111.21%↓0.93%
Table 5. The elapsed time of deriving reducts.
Table 5. The elapsed time of deriving reducts.
IDGBSAR- κ DBSARBCKCRAGESARMNARRARR
119.52665.550841.94458.269210.52070.57526.3996
22.45370.22641.09380.24220.34540.55600.6227
37.01721.21537.00781.38582.08731.27954.0856
48.71280.54023.74150.60981.00570.62100.8102
52.53950.19881.00460.20220.38270.61780.9812
6420.37436.150052.15217.229824.57770.88186.9134
715.16361.907611.12122.11583.41220.21575.7393
8175.1957171.29671348.3062240.2581339.098815.61381478.0267
937.33078.846561.976812.082816.62500.655026.8630
10349.0610169.87871376.8752189.1153322.340811.21991895.9858
117.54161.20878.46091.45692.17520.71561.7449
12129.925233.3983168.706733.477151.68006.7983396.9944
1321.02040.21902.12280.24600.62790.33570.8065
14229.366116.997588.532819.066533.186381.656419.6492
1577.324239.9955296.845347.727074.38945.3366528.5469
1610.85220.63363.17150.64301.01782.00332.4340
Average94.587828.6415217.066535.258055.21708.0676273.5377
↑230.25%↓56.42%↑168.27%↑71.30%↑1072.44%↓65.42%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, M.; Chen, J.; Song, J.; Xu, T.; Fan, Y. Forward Greedy Searching to κ-Reduct Based on Granular Ball. Symmetry 2023, 15, 996. https://doi.org/10.3390/sym15050996

AMA Style

Song M, Chen J, Song J, Xu T, Fan Y. Forward Greedy Searching to κ-Reduct Based on Granular Ball. Symmetry. 2023; 15(5):996. https://doi.org/10.3390/sym15050996

Chicago/Turabian Style

Song, Minhui, Jianjun Chen, Jingjing Song, Taihua Xu, and Yan Fan. 2023. "Forward Greedy Searching to κ-Reduct Based on Granular Ball" Symmetry 15, no. 5: 996. https://doi.org/10.3390/sym15050996

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop