Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.3115/980845.980875dlproceedingsArticle/Chapter ViewAbstractPublication PagesaclConference Proceedingsconference-collections
Article
Free access

Beyond n-grams: can linguistic sophistication improve language modeling?

Published: 10 August 1998 Publication History

Abstract

It seems obvious that a successful model of natural language would incorporate a great deal of both linguistic and world knowledge. Interestingly, state of the art language models for speech recognition are based on a very crude linguistic model, namely conditioning the probability of a word on a small fixed number of preceding words. Despite many attempts to incorporate more sophisticated information into the models, the n-gram model remains the state of the art, used in virtually all speech recognition systems. In this paper we address the question of whether there is hope in improving language modeling by incorporating more sophisticated linguistic and world knowledge, or whether the n-grams are already capturing the majority of the information that can be employed.

References

[1]
Brill E., Harris D., Lowe S., Luo X., Rao P., Ristad E and Roukos S. (1996). A hidden tag model for language. In "Research Notes", Center for Language and speech processing. The Johns Hopkins University. Chapter 2.
[2]
Chelba C., Eagle D., Jelinek F., Jimenez V., Khudanpur S., Mangu L., Printz H., Ristad E., Rosenfeld R., Stolcke A. and Wu D. (1997) Structure and Performance of a Dependency Language Model. In Eurospeech '97. Rhodes, Greece.
[3]
Chomsky N. (1956) Three models for the description of language. IRE Trans. On Inform. Theory. IT-2, 113-124.
[4]
Della Pietra S., Della Pietra V., Gillett J., Lafferty J., Printz H. and tires L. (1994) Inference and Estimation of a Long-Range Trigram Model. In Proceedings of the Second International Colloquium on Grammatical Inference. Alicante, Spain.
[5]
Fong E. and Wu D. (1995) Learning restricted probabilistic link grammars. IJCAI Workshop on New Approaches to Learning for Natural Language Processing, Montreal.
[6]
Golding A and Roth D. (1996) Applying Winnow to Context-Sensitive Spelling Correction. In Proceedings of ICML '96.
[7]
Jelinek F., Lafferty J. D. and Mercer R. L. (1992) Basic Methods of Probabilistic Context-Free Grammars. In "Speech Recognition and Understanding. Recent Advances, Trends, and Applications", Volume F75, 345--360. Berlin:Springer Verlag.
[8]
Jurafsky D., Wooters C., Segal J., Stolcke A., Fosler E., Tajchman G. and Morgan N. (1995) Using a stochastic context-free grammar as a language model for speech recognition. In ICASSP '95.
[9]
Knight K and Chandler I. (1994). Automated Postediting of Documents. Proceedings, Twelfth National Conference on Artificial Intelligence.
[10]
Seymore K. and Rosenfeld R. (1997) Using Story Topics for Language Model Adaptation. In Eurospeech '97. Rhodes, Greece.
[11]
Stolcke A. (1997) Linguistic Knowledge and Empirical Methods in Speech Recognition. In AI Magazine, Volume 18, 25--31, No.4.

Cited By

View all
  • (2010)Spoken language understanding via supervised learning and linguistically motivated featuresProceedings of the Natural language processing and information systems, and 15th international conference on Applications of natural language to information systems10.5555/1894525.1894541(117-128)Online publication date: 23-Jun-2010
  • (2010)Third-party error detection support mechanisms for dictation speech recognitionInteracting with Computers10.1016/j.intcom.2010.02.00222:5(375-388)Online publication date: 1-Sep-2010
  • (2006)Discovering Cues to Error Detection in Speech Recognition OutputJournal of Management Information Systems10.2753/MIS0742-122222040922:4(237-270)Online publication date: 1-Apr-2006
  • Show More Cited By
  1. Beyond n-grams: can linguistic sophistication improve language modeling?

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image DL Hosted proceedings
      ACL '98/COLING '98: Proceedings of the 36th Annual Meeting of the Association for Computational Linguistics and 17th International Conference on Computational Linguistics - Volume 1
      August 1998
      768 pages

      Sponsors

      • Government of Canada
      • Université de Montréal

      Publisher

      Association for Computational Linguistics

      United States

      Publication History

      Published: 10 August 1998

      Qualifiers

      • Article

      Acceptance Rates

      Overall Acceptance Rate 85 of 443 submissions, 19%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)37
      • Downloads (Last 6 weeks)5
      Reflects downloads up to 10 Oct 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2010)Spoken language understanding via supervised learning and linguistically motivated featuresProceedings of the Natural language processing and information systems, and 15th international conference on Applications of natural language to information systems10.5555/1894525.1894541(117-128)Online publication date: 23-Jun-2010
      • (2010)Third-party error detection support mechanisms for dictation speech recognitionInteracting with Computers10.1016/j.intcom.2010.02.00222:5(375-388)Online publication date: 1-Sep-2010
      • (2006)Discovering Cues to Error Detection in Speech Recognition OutputJournal of Management Information Systems10.2753/MIS0742-122222040922:4(237-270)Online publication date: 1-Apr-2006
      • (2005)Error detection using linguistic featuresProceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing10.3115/1220575.1220581(41-48)Online publication date: 6-Oct-2005
      • (2004)Incremental parsing with reference interactionProceedings of the Workshop on Incremental Parsing: Bringing Engineering and Cognition Together10.5555/1613148.1613152(18-25)Online publication date: 25-Jul-2004

      View Options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Get Access

      Login options

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media