Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

Ensembles of Least Squares Classifiers with Randomized Kernels

  • Chapter
Data Mining: Foundations and Practice

Part of the book series: Studies in Computational Intelligence ((SCI,volume 118))

  • 1213 Accesses

Summary

For the recent NIPS-2003 feature selection challenge we studied ensembles of regularized least squares classifiers (RLSC). We showed that stochastic ensembles of simple least squares kernel classifiers give the same level of accuracy as the best single RLSC. Results achieved were ranked among the best at the challenge. We also showed that performance of a single RLSC is much more sensitive to the choice of kernel width than that of an ensemble. As a continuation of this work we demonstrate that stochastic ensembles of least squares classifiers with randomized kernel widths and OOB-post-processing often outperform the best single RLSC, and require practically no parameter tuning. We used the same set of very high dimensional classification problems presented at the NIPS challenge. Fast exploratory Random Forests were applied for variable filtering first.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 219.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. A. Borisov, V. Eruhimov, and E. Tuv. Dynamic soft feature selection for tree-based ensembles. In I. Guyon, S. Gunn, M. Nikravesh, and L. Zadeh, editors, Feature Extraction, Foundations and Applications. Springer, Berlin Heidelberg New York, 2005

    Google Scholar 

  2. O. Bousquet and A. Elisseeff. Algorithmic stability and generalization performance. In NIPS, pages 196–202, 2000

    Google Scholar 

  3. L. Breiman. Bagging predictors. Machine Learning, 24(2):123–140, 1996

    MATH  MathSciNet  Google Scholar 

  4. L. Breiman. Random forests. Machine Learning, 45(1):5–32, 2001

    Article  MATH  Google Scholar 

  5. L. Breiman, J.H. Friedman, R.A. Olshen, and C.J. Stone. Classification and Regression Trees. CRC, Boca Raton, FL, 1984

    MATH  Google Scholar 

  6. F. Cucker and S. Smale. On the mathematial foundations of learning. Bulletin of the American Mathematical Society, 89(1):1–49, 2001

    Article  MathSciNet  Google Scholar 

  7. F. Cucker and S. Smale. Best choices for regularization parameters in learning theory: on the bias-variance problem. Foundations of Computational Mathematics, 2(4):413–428, 2003

    Article  MathSciNet  Google Scholar 

  8. T. Evgeniou. Learning with Kernel Machine Architectures. PhD thesis, Massachusetts Institute of Technology, EECS, July 2000

    Google Scholar 

  9. Y. Freund and R. E. Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. In European Conference on Computational Learning Theory, pages 23–37, 1995

    Google Scholar 

  10. J.H. Friedman. Greedy function approximation: a gradient boosting machine. Technical report, Department of Statistics, Stanford University, 1999

    Google Scholar 

  11. J.H. Friedman. Stochastic gradient boosting. Technical report, Department of Statistics, Stanford University, 1999

    Google Scholar 

  12. A. Hoerl and R. Kennard. Ridge regression; biased estimation for nonorthogonal problems. Technometrics, 12(3):55–67, 1970

    Article  MATH  Google Scholar 

  13. T. Poggio, R. Rifkin, S. Mukherjee, and A. Rakhlin. Bagging regularizes. CBCL Paper 214, Massachusetts Institute of Technology, Cambridge, MA, February 2002. AI Memo #2002–2003

    Google Scholar 

  14. T. Poggio and S. Smale. The mathematics of learning: Dealing with data. Notices of the American Mathematical Society (AMS), 50(5):537–544, 2003

    MATH  MathSciNet  Google Scholar 

  15. R. Rifkin. Everything Old is New Again: A Fresh Look at Historical Approaches in Machine Learning. PhD thesis, MIT, 2002

    Google Scholar 

  16. J.A.K. Suykens and J. Vandervalle. Least squares support vector machines. Neural Processing Letters, 9(3):293–300, June 1999

    Article  Google Scholar 

  17. A.N. Tikhonov and V.Y. Arsenin. Solutions of Ill-posed Problems. W.H. Wingston, Washington DC, 1977

    MATH  Google Scholar 

  18. K. Torkkola. Feature extraction by non-parametric mutual information maximization. Journal of Machine Learning Research, 3:1415–1438, March 2003

    Article  MATH  MathSciNet  Google Scholar 

  19. K. Torkkola and E. Tuv. Ensembles of regularized least squares classifiers for high-dimensional problems. In I. Guyon, S. Gunn, M. Nikravesh, and L. Zadeh, editors, Feature Extraction, Foundations and Applications. Springer, Berlin Heidelberg New York, 2005

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2008 Springer-Verlag Berlin Heidelberg

About this chapter

Cite this chapter

Torkkola, K., Tuv, E. (2008). Ensembles of Least Squares Classifiers with Randomized Kernels. In: Lin, T.Y., Xie, Y., Wasilewska, A., Liau, CJ. (eds) Data Mining: Foundations and Practice. Studies in Computational Intelligence, vol 118. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-78488-3_22

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-78488-3_22

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-78487-6

  • Online ISBN: 978-3-540-78488-3

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics