Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Soft estimation by hierarchical classification and regression

Published: 19 April 2017 Publication History
  • Get Citation Alerts
  • Abstract

    Classification and numeric estimation are the two most common types of data mining. The goal of classification is to predict the discrete type of output values whereas estimation is aimed at finding the continuous type of output values. Predictive data mining is generally achieved by using only one specific statistical or machine learning technique to construct a prediction model. Related studies have shown that prediction performance by this kind of single flat model can be improved by the utilization of some hierarchical structures. Hierarchical estimation approaches, usually a combination of multiple estimation models, have been proposed for solving some specific domain problems. However, in the literature, there is no generic hierarchical approach for estimation and no hybrid based solution that combines classification and estimation techniques hierarchically. Therefore, we introduce a generic hierarchical architecture, namely hierarchical classification and regression (HCR), suitable for various estimation problems. Simply speaking, the first level of HCR involves pre-processing a given training set by classifying it into k classes, leading to k subsets. Three approaches are used to perform this task in this study: hard classification (HC); fuzzy c-means (FCM); and genetic algorithms (GA). Then, each training data containing its associated class label is used to train a support vector machine (SVM) classifier for classification. Next, for the second level of HCR, k regression (or estimation) models are trained based on their corresponding subsets for final prediction. The experiments based on 8 different UCI datasets show that most hierarchical prediction models developed with the HCR architecture significantly outperform three well-known single flat prediction models, i.e., linear regression (LR), multilayer perceptron (MLP) neural networks, and support vector regression (SVR) in terms of mean absolute percentage error (MAPE) and root mean squared error (RMSE) rates. In addition, it is found that using the GA-based data pre-processing approach to classify the training set into 4 subsets is the best threshold (i.e., k=4) and the 4-class SVM+MLP outperforms three baseline hierarchical regression models.

    References

    [1]
    C.H. Achen, Two-step hierarchical estimation: beyond regression analysis, Political Anal., 13 (2005) 447-456.
    [2]
    B. Ayerdi, M. Grana, Hyperspectral image nonlinear unmixing and reconstruction by ELM regression ensemble, Neurocomputing, 174 (2016) 299-309.
    [3]
    F. Bellocchio, S. Ferrari, V. Piuri, N.A. Borghese, Hierarchical approach for multiscale support vector regression, IEEE Trans. Neural Netw. Learn. Syst., 23 (2012) 1448-1460.
    [4]
    J.C. Bezdek, Pattern Recognition With Fuzzy Objective Function Algorithms, Plenum Press, New York, USA, 1981.
    [5]
    H. Byun, S.-W. Lee, A survey on pattern recognition applications of support vector machines, Int. J. Pattern Recognit. Artif. Intell., 17 (2003) 459-486.
    [6]
    J.-S. Chou, C.-F. Tsai, Preliminary cost estimates for thin-film transistor liquid-crystal display inspection and repair equipment: a hybrid hierarchical approach, Comput. Ind. Eng., 62 (2012) 661-669.
    [7]
    J.-S. Chou, C.-F. Tsai, Concrete compressive strength analysis using a combined classification and regression technique, Autom. Constr., 24 (2012) 52-60.
    [8]
    S. Danesh, R. Famoosh, T. Razzaghnia, Fuzzy nonparametric regression based on an adaptive neuro-fuzzy inference system, Neurocomputing, 173 (2016) 1450-1460.
    [9]
    J. Demsar, Statistical comparisons of classifiers over multiple data sets, J. Mach. Learn. Res., 7 (2006) 1-30.
    [10]
    Z. Deng, L. Cao, Y. Jiang, S. Wang, Minimax probability TSK fuzzy system classifier: a more transparent and highly interpretable classification model, IEEE Trans. Fuzzy Syst., 23 (2015) 813-826.
    [11]
    Z. Deng, K.-S. Choi, F.L. Chung, S. Wang, Enhanced soft subspace clustering integrating within-cluster and between-cluster information, Pattern Recognit., 43 (2010) 767-781.
    [12]
    Z. Deng, K.-S. Choi, Y. Jiang, S. Wang, Generalized hidden-mapping ridge regression, knowledge-leveraged inductive transfer learning for neural networks, fuzzy systems and kernel methods, IEEE Trans. Cybern., 44 (2014) 2585-2599.
    [13]
    N.R. Draper, H. Smith, Applied regression analysis, Wiley, New York, USA, 1998.
    [14]
    R.O. Duda, P.E. Hart, D.G. Stork, John Wiley, New York, 2001.
    [15]
    S. Dumais, H. Chen, Hierarchical classification of web content, in: Proceedings of the International ACM SIGIR Conference on Research and Development in Information Retrieval, 2000. pp. 256263.
    [16]
    G.J. Gan, J.H. Wu, A convergence theorem for the fuzzy subspace clustering (FSC) algorithm, Pattern Recognit., 41 (2008) 1939-1947.
    [17]
    B.L.M. Happel, J.M.J. Murre, The design and evolution of modular neural network architectures, Neural Netw., 7 (1994) 985-1004.
    [18]
    J.A. Hartigan, M.A. Wong, A k-means clustering algorithm, Appl. Stat., 28 (1979) 100-108.
    [19]
    S. Haykin, Prentice Hall, New Jersey, 1999.
    [20]
    P. Hellier, C. Barillort, E. Memin, P. Perez, Hierarchical estimation of a dense deformation field for 3-D robust registration, IEEE Trans. Med. Imaging, 20 (2001) 388-402.
    [21]
    T.K. Ho, J.J. Hull, S.N. Srihari, Decision combination in multiple classifier systems, IEEE Trans. Pattern Anal. Mach. Intell., 16 (1994) 66-75.
    [22]
    A.K. Jain, M.N. Murty, P.J. Flynn, Data clustering: a review, ACM Comput. Surv., 31 (1999) 264-323.
    [23]
    Y. Jiang, F.-L. Chung, H. Ishibuchi, Z. Deng, S. Wang, Multitask TSK fuzzy system modeling by mining intertask common hiddenstructure, IEEE Trans. Cybern., 45 (2015) 548-561.
    [24]
    R. Kohavi, A study of cross-validation and bootstrap for accuracy estimation and model selection, in: Proceedings of the International Joint Conference on Artificial Intelligence, 1995, pp. 11371143.
    [25]
    E.L. Lehmann, G. Casella, Theory of Point Estimation, Springer, New York, USA, 1998.
    [26]
    J. Luo, C.-M. Vong, P.-K. Wong, Sparse bayesian extreme learning machine for multi-classification, IEEE Trans. Neural Netw. Learn. Syst., 25 (2014) 836-843.
    [27]
    E. Memin, P. Perez, Hierarchical estimation and segmentation of dense motion fields, Int. J. Comput. Vis., 46 (2002) 129-155.
    [28]
    M. Mitchell, An Introduction to Genetic Algorithms, MIT Press, Massachusetts, USA, 1996.
    [29]
    T. Mitchell, Machine Learning, McGraw Hill, New York, USA, 1997.
    [30]
    R. Nock, F. Nielsen, On weighting clustering, IEEE Trans. Pattern Anal. Mach. Intell., 28 (2006) 1-13.
    [31]
    C. Ozturk, E. Hancer, D. Karaboga, A novel binary artificial bee colony algorithm based on genetic operators, Inf. Sci., 297 (2015) 154-170.
    [32]
    A. Pwasong, S. Sathasivam, A new hybrid quadratic regression and cascade forward backpropagation neural network, Neurcomputing, 182 (2016) 197-209.
    [33]
    J.R. Quinlan, C4.5: Programs for Machine Learning, Morgan Kaufmann, San Francisco, CA, USA, 1993.
    [34]
    S.W. Raudenbush, A.S. Bryk, SAGE Publications, Inc, 2002.
    [35]
    J. Rousu, C. Saunders, S. Szedmak, Learning hierarchical multi-category text classification models, Int. Conf. Mach. Learn. (2005) 744-751.
    [36]
    A. Smola, B. Schlkopf, A Tutorial on Support Vector Regression. NeuroCOLT Technical Report NC-TR-98-030, University of London, 1998.
    [37]
    L.W.G. Strijbosch, J.J.A. Moors, Calculating the accuracy of hierarchical estimation, IMA J. Manag. Math., 21 (2010) 303-315.
    [38]
    A. Sun, and E.-P. Lim, Hierarchical text classification and evaluation, in: Proceedings of the IEEE International Conference on Data Mining, 2001, pp. 521528.
    [39]
    V. Vapnik, Statistical Learning Theory, John Wiley, New York, USA, 1998.
    [40]
    J. Wang, H. He, Y. Cao, J. Xu, D. Zhao, A hierarchical neural network architecture for classification, in: Proceedings of the International Symposium on Neural Networks, 2012, pp. 3746.
    [41]
    H. You, T. Ryu, Development of a hierarchical estimation method for anthropometric variables, Int. J. Ind. Ergon., 35 (2005) 331-343.
    [42]
    S. Yun, J. Na,W.-S. Kang, J. Choi, Hierarchical estimation for adaptive visual tracking. International Conference on Pattern Recognition, 2008, pp. 14.
    [43]
    X. Zuo, C. Chen, W. Tan, M.C. Zhou, Vehicle scheduling of an urban bus line via an improved multiobjective genetic algorithm, IEEE Trans. Intell. Transp. Syst., 16 (2015) 1030-1041.

    Cited By

    View all
    • (2021)Hierarchical classification with multi-path selection based on granular computingArtificial Intelligence Review10.1007/s10462-020-09899-254:3(2067-2089)Online publication date: 1-Mar-2021
    • (2020)Real-time evaluation of driver cognitive loads based on multivariate biosignal analysis2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC)10.1109/SMC42975.2020.9282983(623-628)Online publication date: 11-Oct-2020
    • (2019)A Dustbin Category Based Feedback Incremental Learning Strategy for Hierarchical Image ClassificationPattern Recognition and Computer Vision10.1007/978-3-030-31654-9_41(480-491)Online publication date: 8-Nov-2019

    Index Terms

    1. Soft estimation by hierarchical classification and regression
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Information & Contributors

          Information

          Published In

          cover image Neurocomputing
          Neurocomputing  Volume 234, Issue C
          April 2017
          214 pages

          Publisher

          Elsevier Science Publishers B. V.

          Netherlands

          Publication History

          Published: 19 April 2017

          Author Tags

          1. Classification
          2. Data mining
          3. Hierarchical estimation
          4. Prediction
          5. Regression

          Qualifiers

          • Research-article

          Contributors

          Other Metrics

          Bibliometrics & Citations

          Bibliometrics

          Article Metrics

          • Downloads (Last 12 months)0
          • Downloads (Last 6 weeks)0
          Reflects downloads up to 10 Aug 2024

          Other Metrics

          Citations

          Cited By

          View all
          • (2021)Hierarchical classification with multi-path selection based on granular computingArtificial Intelligence Review10.1007/s10462-020-09899-254:3(2067-2089)Online publication date: 1-Mar-2021
          • (2020)Real-time evaluation of driver cognitive loads based on multivariate biosignal analysis2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC)10.1109/SMC42975.2020.9282983(623-628)Online publication date: 11-Oct-2020
          • (2019)A Dustbin Category Based Feedback Incremental Learning Strategy for Hierarchical Image ClassificationPattern Recognition and Computer Vision10.1007/978-3-030-31654-9_41(480-491)Online publication date: 8-Nov-2019

          View Options

          View options

          Get Access

          Login options

          Media

          Figures

          Other

          Tables

          Share

          Share

          Share this Publication link

          Share on social media