Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

Advertisement

A group incremental feature selection based on knowledge granularity under the context of clustering

  • Original Article
  • Published:
International Journal of Machine Learning and Cybernetics Aims and scope Submit manuscript

Abstract

As a widely used data preprocessing method, feature selection with rough sets aims to delete redundant conditional features. However, most of the traditional feature selection methods target to the static data set environment, and the importance of features is used as the base for feature selection. These methods only consider the importance of features themselves and do not consider the impact of features on classification. In order to overcome such shortcomings, we first use the information of knowledge granules to calculate the similarity of samples in the same cluster and samples in different clusters; Secondly, from the perspective of clustering, we stick to the principle that the samples in the same cluster are as close as possible, and the samples in different clusters are as far away as possible, then a feature selection model of knowledge granularity (in short SKG) based on the clustering background is designed; Thirdly, in order to make the SKG model adapt to the reduction of dynamic data sets, we discuss the incremental learning mechanism of sample and feature changes, and two incremental models SKGOA and SKGAA are designed to deal with the dynamic feature reduction when some samples and features are added into the decision system. Finally, some numerical experiments are conducted to assess the performance of the proposed algorithms, and the results shown that our approaches are of a prominent advantage in terms of computational time and classification accuracy.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data availability

The data that support the findings of this study are openly available in UCIrvine Machine Learning Repository at https://archive.ics.uci.edu/datasets, reference [52].

References

  1. Cheruku R, Edla DR, Kuppili V, Dharavath R (2018) RST-Bat-Miner: a fuzzy rule miner integrating rough set feature selection and bat optimization for detection of diabetes disease. Appl Soft Comput 67:764–780

    Google Scholar 

  2. Chan C (1998) A rough set approach to attribute generalization in data mining. Inf Sci 107(1–4):169–176

    MathSciNet  Google Scholar 

  3. Chen DG, Dong LJ, Mi JS (2020) Incremental mechanism of attribute reduction based on discernible relations for dynamically increasing attribute. Soft Comput 24:321–332

    Google Scholar 

  4. Dai J, Tian H (2013) Entropy measures and granularity measures for set-valued information systems. Inf Sci 240(11):72–82

    MathSciNet  Google Scholar 

  5. Ding WP, Lin CT, Cao ZH (2019) Deep neuro-cognitive co-evolution for fuzzy attribute reduction by quantum leaping PSO with nearest-neighbor memeplexes. IEEE Trans Cybern 49(7):2744–2757

    Google Scholar 

  6. Dong LJ, Chen DG (2020) Incremental attribute reduction with rough set for dynamic datasets with simultaneously increasing samples and attributes. Int J Mach Learn Cybern 11:1339–1355

    Google Scholar 

  7. Hao C, Li J, Fan M, Liu W, Tsang ECC (2017) Optimal scale selection in dynamic multi-scale decision tables based on sequential three-way decisions. Inf Sci 415:213–232

    Google Scholar 

  8. Hamouda SKM, Wahed ME, Alez RHA, Riad K (2018) Robust breast cancer prediction system based on rough set theory at National Cancer Institute of Egypt. Comput Methods Programs Biomed 153:259–268

    Google Scholar 

  9. Huang YY, Guo KJ, Yi XW et al (2022) Matrix representation of the conditional entropy for incremental feature selection on multi-source data. Inf Sci 591:263–286

    Google Scholar 

  10. Jia HJ, Ding SF, Ma H, Xing WQ (2014) Spectral clustering with neighborhood attribute reduction based on information entropy. J Comput 9(6):1316–1324 (in Chinese)

    Google Scholar 

  11. Jia XY, Rao YY, Shang L, Li TG (2020) Similarity-based attribute reduction in rough set theory: a clustering perspective. Int J Mach Learn Cybern 11:1047–1060

    Google Scholar 

  12. Jing YG, Li TR, Fujita H et al (2017) An incremental attribute reduction approach based on knowledge granularity with a multi-granulation view. Inf Sci 411:23–38

    MathSciNet  Google Scholar 

  13. Khalil MI, Kim RY, Seo CY (2020) Challenges and opportunities of big data. J Platf Technol 8(2):3–9

    Google Scholar 

  14. Konecny J, Krajca P (2018) On attribute reduction in concept lattices: experimental evaluation shows discernibility matrix based methods inefficient. Inf Sci 467:431–445

    Google Scholar 

  15. Ko YC, Fujita H, Li T (2017) An evidential analysis of Altman Z-score for financial predictions: case study on solar energy companies. Appl Soft Comput 52:748–759

    Google Scholar 

  16. Li JY, Chen JZ, Qi F et al (2022) Two-dimensional unsupervised feature selection via sparse feature filter. IEEE Trans Cybern. https://doi.org/10.1109/TCYB.2022.3162908

    Article  Google Scholar 

  17. Liu GL, Feng YB (2022) Knowledge granularity reduction for decision tables. Int J Mach Learn Cybern 13(3):569–577

    Google Scholar 

  18. Liang D, Xu Z, Liu D (2017) Three-way decisions based on decision-theoretic rough sets with dual hesitant fuzzy information. Inf Sci 396:127–143

    Google Scholar 

  19. Lei L (2018) Wavelet neural network prediction method of stock price trend based on rough set attribute reduction. Appl Soft Comput 62:923–932

    Google Scholar 

  20. Liang J, Wang F, Dang C (2014) A group incremental approach to feature selection applying rough set technique. IEEE Trans Knowl Data Eng 26:294–308

    Google Scholar 

  21. Li S, Li T (2015) Incremental update of approximations in dominance-based rough sets approach under the variation of attribute values. Inf Sci 294:348–361

    MathSciNet  Google Scholar 

  22. Ma FM, Ding MW, Zhang TF et al (2018) Compressed binary discernibility matrix based incremental attribute reduction algorithm form group dynamic data. Neurocomputing 344:20–27

    Google Scholar 

  23. Nie FP, Yang S, Zhang R, Li XL (2019) A general framework for auto-weighted feature selection via global redundancy minimization. IEEE Trans Image Process 28(5):2428–2438

    MathSciNet  Google Scholar 

  24. Nath K, Roy S, Nandi S et al (2020) A fuzzy-rough approach for detecting overlapping communities with intrinsic structures in evolving networks. Appl Soft Comput 89:106096

    Google Scholar 

  25. Ni P, Zhao S, Wang X et al (2019) PARA: a positive-region based attribute reduction accelerator. Inf Sci 503:533–550

    Google Scholar 

  26. Ni P, Zhao SY, Wang XZ et al (2020) Incremental feature selection based on fuzzy rough sets. Inf Sci 536:185–204

    MathSciNet  Google Scholar 

  27. Pawlak Z (1998) Rough set theory and its application to data analysis. Cybern Syst 29:661–668

    Google Scholar 

  28. Singh AK, Baranwal N, Nandi GC (2019) A rough set based reasoning approach for criminal identification. Int J Mach Learn Cybern 10(3):413–431

    Google Scholar 

  29. Shu WH, Qian WB, Xie YH (2020) Incremental feature selection for dynamic hybrid data using neighborhood rough set. Knowl-Based Syst 194:1–15

    Google Scholar 

  30. Shu WH, Shen H (2014) Updating attribute reduct in incomplete decision systems with the variation of attribute set. Int J Approx Reason 55:867–884

    Google Scholar 

  31. Shu WH, Qian WB, Xie YH (2022) Incremental neighborhood entropy-based feature selection for mixed-type data under the variation of feature set. Appl Intell 52:4792–4806

    Google Scholar 

  32. Utkarsh A, Vasudha R, Rahul K (2022) Normalized mutual information-based equilibrium optimizer with chaotic maps for wrapper-filter feature selection. Expert Syst Appl. https://doi.org/10.1016/j.eswa.2022.118107

    Article  Google Scholar 

  33. Wang S, Zhu W (2018) Sparse graph embedding unsupervised feature selection. IEEE Trans Syst Man Cybern Syst 48(3):329–341

    Google Scholar 

  34. Wan JH, Chen HM, Li TR et al (2021) Dynamic interaction feature selection based on fuzzy rough set. Inf Sci 581:891–911

    Google Scholar 

  35. Wan JH, Chen HM, Yuan Z et al (2021) A novel hybrid feature selection method considering feature interaction in neighborhood rough set. Knowl-Based Syst 227(6):107167

    Google Scholar 

  36. Wang C, He Q, Shao M, Hu Q (2018) Feature selection based on maximal neighborhood discernibility. Int J Mach Learn Cybern 9(11):1929–1940

    Google Scholar 

  37. Wang X, Xing H, Li Y et al (2015) A study on relationship between generalization abilities and fuzziness of base classifiers in ensemble learning. IEEE Trans Fuzzy Syst 23(5):1638–1654

    Google Scholar 

  38. Wang R, Wang X, Kwong S, Xu C (2017) Incorporating diversity and informativeness in multiple-instance active learning. IEEE Trans Fuzzy Syst 25(6):1460–1475

    Google Scholar 

  39. Wei W, Wu X, Liang J, Cui J, Sun Y (2018) Discernibility matrix based incremental attribute reduction for dynamic data. Knowl Based Syst 140:142–157

    Google Scholar 

  40. Wang S, Li TR, Luo C, Fujita H (2016) Efficient updating rough approximations with multi-dimensional variation of ordered data. Inf Sci 372:690–708

    Google Scholar 

  41. Xu JC, Shi JC, Sun L (2002) Attribute reduction algorithm based on relative granularity in decision tables. Comput Sci 36(3):205–207 (in Chinese)

    Google Scholar 

  42. Yuan Z, Chen HM, Xie P et al (2021) Attribute reduction methods in fuzzy rough set theory: an overview, comparative experiments, and new directions. Appl Soft Comput 107:107353. https://doi.org/10.1016/j.asoc.2021.107353

    Article  Google Scholar 

  43. Yang CJ, Ge H, Li LS, Ding J (2019) A unified incremental reduction with the variations of the object for decision tables. Soft Comput 23:6407–6427

    Google Scholar 

  44. Yang YY, Chen DG, Zhang X et al (2022) Incremental feature selection by sample selection and feature-based accelerator. Appl Soft Comput 121:108800. https://doi.org/10.1016/j.asoc.2022.108800

    Article  Google Scholar 

  45. Zhang X, Mei CL, Chen DG et al (2016) Feature selection in mixed data: a method using a novel fuzzy rough set-based information entropy. Pattern Recogn 56:1–15

    Google Scholar 

  46. Zhang CC, Dai JH, Chen JL (2020) Knowledge granularity based incremental attribute reduction for incomplete decision systems. Int J Mach Learn Cybern 11(5):1141–1157

    Google Scholar 

  47. Zhao RN, Gu LZ, Zhu XN (2019) Combining fuzzy C-means clustering with fuzzy rough feature selection. Appl Sci 9:679. https://doi.org/10.3390/app9040679

    Article  Google Scholar 

  48. Zeng A, Li T, Hu J (2016) Dynamical updating fuzzy rough approximations for hybrid data under the variation of attribute values. Inf Sci 378:363–388

    MathSciNet  Google Scholar 

  49. Zhang M, Chen DG, Yang YY (2013) A new algorithm of attribute reduction based on fuzzy clustering. In: Proceedings of the 2013 international conference on machine learning and cybernetics, vol 7, pp 14–17

  50. Zhang P, Li T, Yuan Z et al (2022) Heterogeneous feature selection based on neighborhood combination entropy. IEEE Trans Neural Netw Learn Syst. https://doi.org/10.1109/TNNLS.2022.3193929

    Article  Google Scholar 

  51. Zhang X, Li J (2023) Incremental feature selection approach to interval-valued fuzzy decision information systems based on λ-fuzzy similarity self-information. Inf Sci 625:593–619

    Google Scholar 

  52. https://archive.ics.uci.edu/datasets. Accessed 01 Sept 2021

Download references

Acknowledgements

This work is supported by the Natural Science Foundation of China (61836016), the Key Subject of Chaohu University (kj22zdxk01), the Quality Improvement Project of Chaohu university on Discipline Construction(kj21gczx03), the Provincial Natural Science Research Program of Higher Education Institutions of Anhui province (KJ2021A1030).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Baohua Liang.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Liang, B., Liu, Y., Lu, J. et al. A group incremental feature selection based on knowledge granularity under the context of clustering. Int. J. Mach. Learn. & Cyber. 15, 3647–3670 (2024). https://doi.org/10.1007/s13042-024-02113-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s13042-024-02113-7

Keywords