Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Learning Distributions by Their Density Levels

Published: 01 August 1997 Publication History

Abstract

We propose a mathematical model for learning the high-density areas of an unknown distribution from (unlabeled) random points drawn according to this distribution. While this type of a learning task has not been previously addressed in the computational learnability literature, we believe that this it a rather basic problem that appears in many practical learning scenarios. From a statistical theory standpoint, our model may be viewed as a restricted instance of the fundamental issue of inferring information about a probability distribution from the random samples it generates. From a computational learning angle, what we propose is a few framework of unsupervised concept learning. The examples provided to the learner in our model are not labeled (and are not necessarily all positive or all negative). The only information about their membership is indirectly disclosed to the student through the sampling distribution. We investigate the basic features of the proposed model and provide lower and upper bounds on the sample complexity of such learning tasks. We prove that classes whose VC-dimension is finite are learnable in a very strong sense, while on the other hand, -covering numbers of a concept class impose lower bounds on the sample size needed for learning in our models. One direction of the proof involves a reduction of the density-level learnability to PAC learning with respect to fixed distributions (as well as some fundamental statistical lower bounds), while the sufficiency condition is proved through the introduction of a generic learning algorithm.

References

[1]
D. Angluin, P. Laird, Learning from noisy examples, Machine Learning, 2 (1988) 343-370.
[2]
G.M. Benedek, A. Itai, Learnability with respect to fixed distributions, Theoret. Comput. Sci., 86 (1991) 377-389.
[3]
S. Ben-David, M. Lindebaum, Localization vs identification of semi-algebraic sets, Proc. 6th COLT, 1993, 327, 336
[4]
A. Blumer, A. Ehrenfeucht, D. Haussler, M.K. Warmuth, Learnability and the Vapnik¿Chervonenkis dimension, Assoc. Comput. Machine, 36 (1989) 929-965.
[5]
R. Canetti, G. Even, O. Goldreich, 1993, Lower bounds for sampling algorithms for estimating the average, Technion¿Israel Institute of Technology
[6]
E. Dichterman
[7]
R.O. Duda, P.E. Hart, Wiley, New York, 1973.
[8]
R.M. Dudley, Springer-Verlag, New York/Berlin, 1984.
[9]
A. Ehrenfeucht, D. Haussler, M. Kearns, L. Valiant, A general lower bound on the number of examples needed for learning, Inform. and Comput., 82 (1989) 247-261.
[10]
P. Goldberg, M. Jerrum, Bounding the Vapnik¿Chervonenkis dimension of concept classes parameterized by real numbers, Machine Learning, 18 (1995) 131-148.
[11]
J. Illingworth, J. Kittler, A survey of Hough transform, Comput. Vision, Graphics, and Image Process., 44 (1988) 87-116.
[12]
M. Karpinski, A. MacIntyre, Polynomial bounds for VC-dimension of sigmoidal neural networks, Proceedings of 27th STOC, 1995, 200, 208
[13]
A.G. Khovanskii, Fewnomials, Transl. Math. Mono- graphs, 88 (1981).
[14]
M. Kearns, Y. Mansour, D. Ron, R. Rubinfeld, R. E. Shapire, L. Sellie, On the learnability of discrete distributions, Proceedings of 26th ACM STOC, 1994, 273, 282
[15]
M. Kearns, R. E. Shapire, Efficient distribution-free learning of probabilistic concepts, Proc. of 31st FOCS, 1994, 382, 392
[16]
M. W. Kim, Learning by smoothing: A morphological approach, Proc. of 4th COLT, 1991, 43, 57
[17]
B.K. Natarajan, Probably approximate learning of sets and functions, SIAM J. Comput., 20 (1991) 328-351.
[18]
A. Papoulis, MacGraw¿Hill, New York, 1984.
[19]
H.U. Simon, General bounds on the number of examples needed for learning probabilistic concepts, J. Comput. System Sci., 52 (1996) 239-254.
[20]
V.N. Vapnik, A.Y. Chervonenkis, On the uniform convergence of relative frequencies of events to their probabilities, Theory Probab. and Its Appl., 16 (1971) 264-280.

Cited By

View all
  • (2019)Multivariate triangular quantile maps for novelty detectionProceedings of the 33rd International Conference on Neural Information Processing Systems10.5555/3454287.3454742(5060-5071)Online publication date: 8-Dec-2019
  • (2012)Learning high-density regions for a generalized Kolmogorov-Smirnov test in high-dimensional dataProceedings of the 25th International Conference on Neural Information Processing Systems - Volume 110.5555/2999134.2999216(728-736)Online publication date: 3-Dec-2012
  • (2012)Learning rates of support vector machine classifier for density level detectionNeurocomputing10.1016/j.neucom.2011.10.03282(84-90)Online publication date: 1-Apr-2012
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

Publisher

Academic Press, Inc.

United States

Publication History

Published: 01 August 1997

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 04 Oct 2024

Other Metrics

Citations

Cited By

View all
  • (2019)Multivariate triangular quantile maps for novelty detectionProceedings of the 33rd International Conference on Neural Information Processing Systems10.5555/3454287.3454742(5060-5071)Online publication date: 8-Dec-2019
  • (2012)Learning high-density regions for a generalized Kolmogorov-Smirnov test in high-dimensional dataProceedings of the 25th International Conference on Neural Information Processing Systems - Volume 110.5555/2999134.2999216(728-736)Online publication date: 3-Dec-2012
  • (2012)Learning rates of support vector machine classifier for density level detectionNeurocomputing10.1016/j.neucom.2011.10.03282(84-90)Online publication date: 1-Apr-2012
  • (2011)The use of hybrid manifold learning and support vector machines in the prediction of business failureKnowledge-Based Systems10.1016/j.knosys.2010.07.00924:1(95-101)Online publication date: 1-Feb-2011
  • (2010)A lower bound for learning distributions generated by probabilistic automataProceedings of the 21st international conference on Algorithmic learning theory10.5555/1893193.1893216(179-193)Online publication date: 6-Oct-2010
  • (2006)Learning Minimum Volume SetsThe Journal of Machine Learning Research10.5555/1248547.12485717(665-704)Online publication date: 1-Dec-2006
  • (2006)Outlier detection by sampling with accuracy guaranteesProceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining10.1145/1150402.1150501(767-772)Online publication date: 20-Aug-2006
  • (2006)Outlier detection by active learningProceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining10.1145/1150402.1150459(504-509)Online publication date: 20-Aug-2006
  • (2005)Learning minimum volume setsProceedings of the 18th International Conference on Neural Information Processing Systems10.5555/2976248.2976400(1209-1216)Online publication date: 5-Dec-2005
  • (2005)An application of support vector machines for customer churn analysisProceedings of the First international conference on Advances in Natural Computation - Volume Part II10.1007/11539117_91(636-647)Online publication date: 27-Aug-2005
  • Show More Cited By

View Options

View options

Get Access

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media