Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
Skip header Section
Computational learning theory: an introductionMay 1992
Publisher:
  • Cambridge University Press
  • 40 W. 20 St. New York, NY
  • United States
ISBN:978-0-521-41603-0
Published:01 May 1992
Pages:
157
Skip Bibliometrics Section
Reflects downloads up to 25 Jan 2025Bibliometrics
Abstract

No abstract available.

Cited By

  1. Vu K, Poirion P and Liberti L (2019). Random Projections for Linear Programming, Mathematics of Operations Research, 43:4, (1051-1071), Online publication date: 1-Nov-2018.
  2. Anthony M and Ratsaby J (2017). Classification based on prototypes with spheres of influence, Information and Computation, 256:C, (372-380), Online publication date: 1-Oct-2017.
  3. Kim S, Yu Z, Kil R and Lee M (2018). Deep learning of support vector machines with class probability output networks, Neural Networks, 64:C, (19-28), Online publication date: 1-Apr-2015.
  4. ACM
    Cate B, Dalmau V and Kolaitis P (2013). Learning schema mappings, ACM Transactions on Database Systems, 38:4, (1-31), Online publication date: 1-Nov-2013.
  5. Anthony M and Ratsaby J (2012). Analysis of a multi-category classifier, Discrete Applied Mathematics, 160:16-17, (2329-2338), Online publication date: 1-Nov-2012.
  6. ACM
    ten Cate B, Dalmau V and Kolaitis P Learning schema mappings Proceedings of the 15th International Conference on Database Theory, (182-195)
  7. Anthony M and Ratsaby J (2012). Robust cutpoints in the logical analysis of numerical data, Discrete Applied Mathematics, 160:4, (355-364), Online publication date: 1-Mar-2012.
  8. Teng F, Chen Y and Dang X (2012). Multiclass classification with potential function rules, Pattern Recognition, 45:1, (540-551), Online publication date: 1-Jan-2012.
  9. Goldman S Computational learning theory Algorithms and theory of computation handbook, (26-26)
  10. Anthony M and Ratsaby J (2019). Maximal width learning of binary functions, Theoretical Computer Science, 411:1, (138-147), Online publication date: 1-Jan-2010.
  11. Li D and Liu C (2018). A neural network weight determination model designed uniquely for small data set learning, Expert Systems with Applications: An International Journal, 36:6, (9853-9858), Online publication date: 1-Aug-2009.
  12. ACM
    Pyrga E and Ray S New existence proofs ε-nets Proceedings of the twenty-fourth annual symposium on Computational geometry, (199-207)
  13. Elbassioni K (2008). On the complexity of monotone dualization and generating minimal hypergraph transversals, Discrete Applied Mathematics, 156:11, (2109-2123), Online publication date: 1-Jun-2008.
  14. Zeugmann T and Zilles S (2008). Learning recursive functions, Theoretical Computer Science, 397:1-3, (4-56), Online publication date: 10-May-2008.
  15. ACM
    Heikinheimo H, Hinkkanen E, Mannila H, Mielikäinen T and Seppänen J Finding low-entropy sets and trees from binary data Proceedings of the 13th ACM SIGKDD international conference on Knowledge discovery and data mining, (350-359)
  16. ACM
    Benedikt M, Libkin L and Neven F (2007). Logical definability and query languages over ranked and unranked trees, ACM Transactions on Computational Logic, 8:2, (11-es), Online publication date: 1-Apr-2007.
  17. Khachiyan L, Boros E, Elbassioni K and Gurvich V (2006). An efficient implementation of a quasi-polynomial algorithm for generating hypergraph transversals and its application in joint generation, Discrete Applied Mathematics, 154:16, (2350-2372), Online publication date: 1-Nov-2006.
  18. Anthony M and Hammer P (2006). A Boolean measure of similarity, Discrete Applied Mathematics, 154:16, (2242-2246), Online publication date: 1-Nov-2006.
  19. Chang F Using the small data set learning for the monitor indicator forecast Proceedings of the 10th WSEAS international conference on Systems, (778-784)
  20. Erdoğan E and Iyengar G (2006). Ambiguous chance constrained problems and robust optimization, Mathematical Programming: Series A and B, 107:1-2, (37-61), Online publication date: 1-Jun-2006.
  21. Tao D, Li X, Hu W, Maybank S and Wu X Supervised Tensor Learning Proceedings of the Fifth IEEE International Conference on Data Mining, (450-457)
  22. Mielikäinen T Transaction databases, frequent itemsets, and their condensed representations Proceedings of the 4th international conference on Knowledge Discovery in Inductive Databases, (139-164)
  23. Bshouty N, Jackson J and Tamon C (2005). Exploring learnability between exact and PAC, Journal of Computer and System Sciences, 70:4, (471-484), Online publication date: 1-Jun-2005.
  24. Anthony M (2004). Generalization Error Bounds for Threshold Decision Lists, The Journal of Machine Learning Research, 5, (189-217), Online publication date: 1-Dec-2004.
  25. Anthony M (2004). Some connections between learning and optimization, Discrete Applied Mathematics, 144:1-2, (17-26), Online publication date: 1-Nov-2004.
  26. Anthony M (2004). On data classification by iterative linear partitioning, Discrete Applied Mathematics, 144:1-2, (2-16), Online publication date: 1-Nov-2004.
  27. Case J, Jain S, Stephan F and Wiehagen R (2004). Robust learning, Journal of Computer and System Sciences, 69:2, (123-165), Online publication date: 1-Sep-2004.
  28. Boros E, Horiyama T, Ibaraki T, Makino K and Yagiura M (2019). Finding Essential Attributes from Binary Data, Annals of Mathematics and Artificial Intelligence, 39:3, (223-257), Online publication date: 1-Nov-2003.
  29. Makino K, Ono H and Ibaraki T (2003). Interior and exterior functions of positive Boolean functions, Discrete Applied Mathematics, 130:3, (417-436), Online publication date: 23-Aug-2003.
  30. Mărginean F Computational science and data mining Proceedings of the 2003 international conference on Computational science: PartIII, (644-651)
  31. Mărginean F Computational aspects of data mining Proceedings of the 2003 international conference on Computational science and its applications: PartI, (614-622)
  32. Boros E, Ibaraki T and Makino K (2003). Variations on extending partially defined boolean functions with missing bits, Information and Computation, 180:1, (53-70), Online publication date: 10-Jan-2003.
  33. Ono H, Makino K and Ibaraki T (2002). Logical analysis of data with decomposable structures, Theoretical Computer Science, 289:2, (977-995), Online publication date: 30-Oct-2002.
  34. Eiter T, Ibaraki T and Makino K (2019). Decision lists and related Boolean functions, Theoretical Computer Science, 270:1-2, (493-524), Online publication date: 6-Jan-2002.
  35. Polkowski L and Skowron A Logic prespective on data and knowledge Handbook of data mining and knowledge discovery, (99-115)
  36. Guijarro D, Lavin V and Raghavan V (2001). Monotone term decision lists, Theoretical Computer Science, 259:1, (549-575), Online publication date: 28-May-2001.
  37. Giraud-Carrier C (2018). A note on the utility of incremental learning, AI Communications, 13:4, (215-223), Online publication date: 1-Dec-2000.
  38. Jiang W (2000). The VC Dimension for Mixtures of Binary Classifiers, Neural Computation, 12:6, (1293-1301), Online publication date: 1-Jun-2000.
  39. Velardi P and Cucchiarelli A Dependency of context-based word sense disambiguation from representation and domain complexity Proceedings of the 2000 NAACL-ANLP Workshop on Syntactic and semantic complexity in natural language processing systems - Volume 1, (28-34)
  40. Schmitt M (2019). On the Sample Complexity for Nonoverlapping Neural Networks, Machine Language, 37:2, (131-141), Online publication date: 1-Nov-1999.
  41. Goldman S and Scott S (1999). A Theoretical and Empirical Study of a Noise-TolerantAlgorithm to Learn Geometric Patterns, Machine Language, 37:1, (5-49), Online publication date: 1-Oct-1999.
  42. ACM
    Gamarnik D Extension of the PAC framework to finite and countable Markov chains Proceedings of the twelfth annual conference on Computational learning theory, (308-317)
  43. ACM
    Goldberg P Learning fixed-dimension linear thresholds from fragmented data Proceedings of the twelfth annual conference on Computational learning theory, (88-99)
  44. ACM
    Benedikt M and Libkin L Exact and approximate aggregation in constraint query languages Proceedings of the eighteenth ACM SIGMOD-SIGACT-SIGART symposium on Principles of database systems, (102-113)
  45. ACM
    Anthony M and Holden S Cross-validation for binary classification by real-valued functions Proceedings of the eleventh annual conference on Computational learning theory, (218-229)
  46. Akutsu T, Kuhara S, Maruyama O and Miyano S Identification of gene regulatory networks by strategic gene disruptions and gene overexpressions Proceedings of the ninth annual ACM-SIAM symposium on Discrete algorithms, (695-702)
  47. Anthony M (1998). Probabilistic ’generalization‘ of functions and dimension-based uniform convergence results, Statistics and Computing, 8:1, (5-14), Online publication date: 1-Jan-1998.
  48. ACM
    Kowalczyk A Dense shattering and teaching dimensions for differentiable families (extended abstract) Proceedings of the tenth annual conference on Computational learning theory, (143-151)
  49. ACM
    Maass W and Schmitt M On the complexity of learning for a spiking neuron (extended abstract) Proceedings of the tenth annual conference on Computational learning theory, (54-61)
  50. ACM
    Angluin D (1996). A 1996 snapshot of computational learning theory, ACM Computing Surveys (CSUR), 28:4es, (216-es), Online publication date: 1-Dec-1996.
  51. Lovell B and Bradley A (1996). The Multiscale Classifier, IEEE Transactions on Pattern Analysis and Machine Intelligence, 18:2, (124-137), Online publication date: 1-Feb-1996.
  52. ACM
    Lee W, Bartlett P and Williamson R The importance of convexity in learning with squared loss Proceedings of the ninth annual conference on Computational learning theory, (140-146)
  53. ACM
    Holden S PAC-like upper bounds for the sample complexity of leave-one-out cross-validation Proceedings of the ninth annual conference on Computational learning theory, (41-50)
  54. Kowalczyk A, Szymanski J, Bartlett P and Williamson R Examples of learning curves from a modified VC-formalism Proceedings of the 9th International Conference on Neural Information Processing Systems, (344-350)
  55. Koiran P and Sontag E Neural networks with quadratic VC dimension Proceedings of the 9th International Conference on Neural Information Processing Systems, (197-203)
  56. ACM
    Lee W, Bartlett P and Williamson R On efficient agnostic learning of linear combinations of basis functions Proceedings of the eighth annual conference on Computational learning theory, (369-376)
  57. ACM
    Karpinski M and Macintyre A Polynomial bounds for VC dimension of sigmoidal neural networks Proceedings of the twenty-seventh annual ACM symposium on Theory of computing, (200-208)
  58. Gärtner B and Welzl E (1994). Vapnik-Chervonenkis dimension and (pseudo-)hyperplane arrangements, Discrete & Computational Geometry, 12:4, (399-432), Online publication date: 1-Dec-1994.
  59. Spoerri A InfoCrystal Intelligent Multimedia Information Retrieval Systems and Management - Volume 1, (687-696)
  60. Turmon M and Fine T Sample size requirements for feedforward neural networks Proceedings of the 8th International Conference on Neural Information Processing Systems, (327-334)
  61. Kowalczyk A and Ferra H Generalisation in feedforward networks Proceedings of the 8th International Conference on Neural Information Processing Systems, (215-222)
  62. ACM
    Kietz J and Džeroski S (1994). Inductive logic programming and learnability, ACM SIGART Bulletin, 5:1, (22-32), Online publication date: 1-Jan-1994.
  63. ACM
    Natarajan B Occam's razor for functions Proceedings of the sixth annual conference on Computational learning theory, (370-376)
  64. ACM
    Goldberg P and Jerrum M Bounding the Vapnik-Chervonenkis dimension of concept classes parameterized by real numbers Proceedings of the sixth annual conference on Computational learning theory, (361-369)
  65. ACM
    Anthony M and Holden S On the power of polynomial discriminators and radial basis function networks Proceedings of the sixth annual conference on Computational learning theory, (158-164)
  66. ACM
    Lange S and Zeugmann T Language learning in dependence on the space of hypotheses Proceedings of the sixth annual conference on Computational learning theory, (127-136)
  67. ACM
    Anthony M, Brightwell G, Cohen D and Shawe-Taylor J On exact specification by examples Proceedings of the fifth annual workshop on Computational learning theory, (311-318)
Contributors
  • London School of Economics and Political Science
  • London School of Economics and Political Science

Reviews

Robert Stewart Roos

The authors of this brief, accessible introduction to the main ideas of probably approximately correct (PAC) learnability emphasize the main theoretical results and provide several illustrative examples. The presentation is leisurely, building from intuitive concepts and simple examples (such as learning monomials) to the definition, in chapter 3, of PAC learning. Considerations of computational complexity (as a function of accuracy, confidence, sample size, and representation size) are deferred until chapters 5 and 6. Chapters 7 and 8 introduce the Vapnik-Chervonenkis (VC) dimension and contain the central theorem of the book—the proof that classes with finite VC dimension are precisely those that are potentially learnable, that is, for which the existence of a consistent learning algorithm implies PAC-learnability. (The authors do not neglect to mention the necessary measure-theoretic qualifications to this statement, but the book does not contain a detailed discussion of such matters.) The next chapter revisits and extends results from previous chapters in light of the VC dimension material. The book concludes with applications to perceptrons and neural nets. The preface states that the book is intended for graduate students and researchers unfamiliar with the basic concepts of PAC-learning. In fact, since the text makes few assumptions about the reader's background in complexity theory (notions such as O -notation, NP-completeness, and polynomial-time reductions are introduced at the appropriate places), topics from the first six or seven chapters could easily be used to enrich an undergraduate course in the theory of algorithms or complexity. The book's brevity makes it unsuitable as the sole textbook for a semester course in learning theory, but it could easily be used in conjunction with a collection of papers dealing with topics not covered in the book. In particular, the book omits detailed explorations of weak learnability, learning with queries, learning under fixed distributions, learning from noisy data, and mistake-bounded learning (although these subjects are mentioned in remarks at the ends of the chapters, with bibliographic references). Only the classic examples are used; readers should not expect a catalog of VC dimension results or learning algorithms for specific concept classes. The authors do not deal with learning in the limit, another area that is often classified under the heading of computational learning theory. Each of the ten chapters ends with a set of exercises; they range in difficulty from extremely simple calculus or plug-in-the-numbers problems to moderately difficult proofs. The bibliography contains all the basic pape rs on PAC learnability, but is by no means a comprehensive list of PAC learning papers. The index is good. The text appears to have few typographical errors. Some might quibble with choices of notation, such as the use of D n,k rather than the more familiar k -DNF. A comparison to Natarajan's recent book on computational learning theory [1] seems appropriate. The Anthony and Biggs book moves at a slower pace than the Natarajan text. Natarajan's book includes a number of additional topics, such as a chapter devoted to the learnability of non-Boolean concepts, one on learning finite automata, and sections devoted to weak learnability and cryptography-based hardness results. In short, the Anthony and Biggs work has been aptly classified by the publisher as a tract rather than a textbook; it is notable for its clean, readable, self-contained treatment of the foundations of PAC learnability theory.

Access critical reviews of Computing literature here

Become a reviewer for Computing Reviews.

Recommendations