Abstract
In this paper, we address the problem of one-class classification for medical image classification. Indeed, in some situations, pathological samples may be difficult to acquire. In this case, one class classification (OCC) is a natural learning paradigm to be used. It consists in learning from only one class of objects, while two or more classes may be presented in prediction. We propose an original OCC method called One-Class Random Forest (OCRF), that combines ensemble learning principles from traditional Random Forest algorithm with an original outlier generation method. These two key processes complement each other for responding to OCC issues, and are shown to perform well on medical datasets in comparison to few other state-of-the-art OCC methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Tarassenko, L., Hayton, P., Cerneaz, N., Brady, M.: Novelty detection for the identification of masses in mammograms. In: Fourth International Conference on Artificial Neural Networks, pp. 442–447 (1995)
Mourão-Miranda, J., Hardoon, D., Hahn, T., Marquand, A., Williams, S., Shawe-Taylor, J., Brammer, M.: Patient classification as an outlier detection problem: An application of the one-class support vector machine. NeuroImage (2011)
Khan, S.S., Madden, M.G.: A Survey of Recent Trends in One Class Classification. In: Coyle, L., Freyne, J. (eds.) AICS 2009. LNCS, vol. 6206, pp. 188–197. Springer, Heidelberg (2010)
Tax, D.M.J., Duin, R.P.W.: Combining One-Class Classifiers. In: Kittler, J., Roli, F. (eds.) MCS 2001. LNCS, vol. 2096, pp. 299–308. Springer, Heidelberg (2001)
Scholkopf, B., Platt, J., Shawe-Taylor, J., Smola, A., Williamson, R.: Estimating the support of a high-dimensional distribution. Neural Computation 13(7), 1443–1471 (2001)
Hempstalk, K., Frank, E., Witten, I.H.: One-Class Classification by Combining Density and Class Probability Estimation. In: Daelemans, W., Goethals, B., Morik, K. (eds.) ECML PKDD 2008, Part I. LNCS (LNAI), vol. 5211, pp. 505–519. Springer, Heidelberg (2008)
Dietterich, T.G.: Ensemble Methods in Machine Learning. In: Kittler, J., Roli, F. (eds.) MCS 2000. LNCS, vol. 1857, pp. 1–15. Springer, Heidelberg (2000)
Breiman, L.: Random forests. Machine Learning 45(1), 5–32 (2001)
Gray, K.R., Aljabar, P., Heckemann, R.A., Hammers, A., Rueckert, D.: Random Forest-Based Manifold Learning for Classification of Imaging Data in Dementia. In: Suzuki, K., Wang, F., Shen, D., Yan, P. (eds.) MLMI 2011. LNCS, vol. 7009, pp. 159–166. Springer, Heidelberg (2011)
Bernard, S., Heutte, L., Adam, S.: Forest-RK: A New Random Forest Induction Method. In: Huang, D.-S., Wunsch II, D.C., Levine, D.S., Jo, K.-H. (eds.) ICIC 2008. LNCS (LNAI), vol. 5227, pp. 430–437. Springer, Heidelberg (2008)
Geurts, P., Ernst, D., Wehenkel, L.: Extremely randomized trees. Machine Learning 63(1), 3–42 (2006)
Ho, T.: The random subspace method for constructing decision forests. IEEE Transactions on Pattern Analysis and Machine Intelligence 20(8), 832–844 (1998)
Thiberville, L., Salaün, M., Lachkar, S., Dominique, S., Moreno-Swirc, S., Vever-Bizet, C., Bourg-Heckly, G.: Human in vivo fluorescence microimaging of the alveolar ducts and sacs during bronchoscopy. Eur. Respir. J. 33(5), 974–985 (2009)
Blake, C., Merz, C.: Uci repository of machine learning databases. Department of Information and Computer Science, University of California, Irvine, CA, vol. 55 (1998), http://www.ics.uci.edu/~mlearn/mlrepository.html
Ojala, T., Pietikäinen, M., Mäenpää, T.: Gray Scale and Rotation Invariant Texture Classification with Local Binary Patterns. In: Vernon, D. (ed.) ECCV 2000, Part I. LNCS, vol. 1842, pp. 404–420. Springer, Heidelberg (2000)
Baldi, P., Brunak, S., Chauvin, Y., Andersen, C., Nielsen, H.: Assessing the accuracy of prediction algorithms for classification: an overview. Bioinformatics 16(5), 412 (2000)
Duin, R.: PRTools version 3.0: A Matlab toolbox for pattern recognition. In: Proc. of SPIE (2000)
Bernard, S., Heutte, L., Adam, S.: Influence of Hyperparameters on Random Forest Accuracy. In: Benediktsson, J.A., Kittler, J., Roli, F. (eds.) MCS 2009. LNCS, vol. 5519, pp. 171–180. Springer, Heidelberg (2009)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Désir, C., Bernard, S., Petitjean, C., Heutte, L. (2012). A Random Forest Based Approach for One Class Classification in Medical Imaging. In: Wang, F., Shen, D., Yan, P., Suzuki, K. (eds) Machine Learning in Medical Imaging. MLMI 2012. Lecture Notes in Computer Science, vol 7588. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-35428-1_31
Download citation
DOI: https://doi.org/10.1007/978-3-642-35428-1_31
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-35427-4
Online ISBN: 978-3-642-35428-1
eBook Packages: Computer ScienceComputer Science (R0)