Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/1647314.1647325acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
research-article

Building multimodal applications with EMMA

Published: 02 November 2009 Publication History

Abstract

Multimodal interfaces combining natural modalities such as speech and touch with dynamic graphical user interfaces can make it easier and more effective for users to interact with applications and services on mobile devices. However, building these interfaces remains a complex and high specialized task. The W3C EMMA standard provides a representation language for inputs to multimodal systems facilitating plug-and-play of system components and rapid prototyping of interactive multimodal systems. We illustrate the capabilities of the EMMA standard through examination of its use in a series of mobile multimodal applications for the iPhone.

References

[1]
J. Axelsson, C. Cross, J. Ferrans, G. McCobb, T. V. Raman, and L. Wilson. Mobile X+V 1.2, 2005. http://openstandardswork.net/specs/multimodal/x+v/mobile/12/.
[2]
S. Bangalore and M. Johnston. Robust understanding in multimodal interfaces. Computational Linguistics (Forthcoming).
[3]
A. Cheyer and L. Julia. Multimodal Maps: An Agent-Based Approach. Lecture Notes in Computer Science, 1374:103--113, 1998.
[4]
P. Cohen, M. Johnston, D. McGee, S. L. Oviatt, J. Pittman, I. Smith, L. Chen, and J. Clow. Multimodal interaction for distributed interactive simulation. In M. Maybury and W. Wahlster, editors, Readings in Intelligent Interfaces. Morgan Kaufmann Publishers, 1998.
[5]
G. D. Fabbrizio, J. Wilpon, and T. Okken. A speech mashup framework for multimodal mobile services. In Proceedings of ICMI, Boston, USA, 2009.
[6]
V. Goffin, C. Allauzen, E. Bocchieri, D. Hakkani-Tur, A. Ljolje, S. Parthasarathy, M. Rahim, G. Riccardi, and M. Saraclar. The AT&T WATSON Speech Recognizer. In Proceedings of ICASSP, Philadelphia, PA, 2005.
[7]
A. Gruenstein, I. McGraw, and I. Badr. The WAMI toolkit for developing, deploying, and evaluating web-accessible multimodal interfaces. In Proceedings of ICMI, pages 141--148, 2008.
[8]
A. Gruenstein, J. Orszulak, S. Liu, S. Roberts, J. Zabel, B. Reimer, B. Mehler, S. Seneff, J. Glass, and J. Coughlin. City browser: developing a conversational automotive HMI. In Proceedings CHI '09 Extended Abstracts, pages 4291--4296. ACM, 2009.
[9]
J. Gustafson, L. Bell, J. Beskow, J. Boye, R. Carlson, J. Edlund, B. Granstrom, D. House, and M.Wiren. AdApt - a multimodal conversational dialogue system in an apartment domain. In Proceedings of ICSLP, pages 134--137, Beijing, China, 2000.
[10]
A. Hunt and S. McGlashan. Speech Recognition Grammar Specification Version 1.0, March 2004. http://www.w3.org/TR/2004/REC-speech-grammar-20040316/.
[11]
M. Johnston, P. Baggia, D. C. Burnett, J. Carter, D. A. Dahl, G. McCobb, and D. Raggett. EMMA: Extensible MultiModal Annotation markup language, February 2009. http://www.w3.org/TR/2009/REC-emma-20090210.
[12]
M. Johnston and S. Bangalore. Finite-state multimodal integration and understanding. Journal of Natural Language Engineering, 11(2):159--187, 2005.
[13]
M. Johnston, S. Bangalore, G. Vasireddy, A. Stent, P. Ehlen, M. Walker, S. Whittaker, and P. Maloor. MATCH: An architecture for multimodal dialog systems. In Proceedings of ACL, pages 376--383, Philadelphia, 2002.
[14]
M. Johnston, L.-F. D'Haro, M. Levine, and B. Renger. A multimodal interface for access to content in the home. In Proceedings of ACL, pages 376--383, 2007.
[15]
N. Reithinger, S. Bergweiler, R. Engel, G. Herzog, N. Pfleger, M. Romanelli, and D. Sonntag. A look under the hood: design and development of the first SmartWeb system demonstrator. In Proceedings of ICMI, pages 159--166, Trento, Italy, 2005.
[16]
L. V. Tichelen and D. Burke. Semantic Interpretation for Speech Recognition (SISR) Version 1.0, April 2007. http://www.w3.org/TR/2007/REC-semanticinterpretation-20070405/.
[17]
W. Wahlster. SmartKom: Fusion and fission of speech, gestures, and facial expressions. In Proceedings of the 1st International Workshop on Man-Machine Symbiotic Systems, pages 213--225, Kyoto, Japan, 2002.
[18]
K. Wang. SALT: A spoken language interface for web-based multimodal dialog systems. In Proceedings of ICASSP, pages 2241--2244, 2002.
[19]
K. Wittenburg, T. Lanning, D. Schwenke, H. Shubin, and A. Vetro. The prospects for unrestricted speech input for TV content search. In AVI '06: Proceedings of the working conference on Advanced visual interfaces, pages 352--359, New York, NY, USA, 2006. ACM.

Cited By

View all
  • (2022)Affordance embeddings for situated language understandingFrontiers in Artificial Intelligence10.3389/frai.2022.7747525Online publication date: 23-Sep-2022
  • (2021)The Role of Embodiment and Simulation in Evaluating HCI: Experiments and EvaluationDigital Human Modeling and Applications in Health, Safety, Ergonomics and Risk Management. Human Body, Motion and Behavior10.1007/978-3-030-77817-0_17(220-232)Online publication date: 3-Jul-2021
  • (2020)Jarvis: A Multimodal Visualization Tool for Bioinformatic DataHCI International 2020 – Late Breaking Papers: Interaction, Knowledge and Social Media10.1007/978-3-030-60152-2_9(104-116)Online publication date: 27-Sep-2020
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
ICMI-MLMI '09: Proceedings of the 2009 international conference on Multimodal interfaces
November 2009
374 pages
ISBN:9781605587721
DOI:10.1145/1647314
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 02 November 2009

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. gesture
  2. multimodal
  3. prototyping
  4. speech
  5. standards

Qualifiers

  • Research-article

Conference

ICMI-MLMI '09
Sponsor:

Acceptance Rates

Overall Acceptance Rate 453 of 1,080 submissions, 42%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)7
  • Downloads (Last 6 weeks)1
Reflects downloads up to 27 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2022)Affordance embeddings for situated language understandingFrontiers in Artificial Intelligence10.3389/frai.2022.7747525Online publication date: 23-Sep-2022
  • (2021)The Role of Embodiment and Simulation in Evaluating HCI: Experiments and EvaluationDigital Human Modeling and Applications in Health, Safety, Ergonomics and Risk Management. Human Body, Motion and Behavior10.1007/978-3-030-77817-0_17(220-232)Online publication date: 3-Jul-2021
  • (2020)Jarvis: A Multimodal Visualization Tool for Bioinformatic DataHCI International 2020 – Late Breaking Papers: Interaction, Knowledge and Social Media10.1007/978-3-030-60152-2_9(104-116)Online publication date: 27-Sep-2020
  • (2019)Commercialization of multimodal systemsThe Handbook of Multimodal-Multisensor Interfaces10.1145/3233795.3233812(621-658)Online publication date: 1-Jul-2019
  • (2019)Standardized representations and markup languages for multimodal interactionThe Handbook of Multimodal-Multisensor Interfaces10.1145/3233795.3233806(347-392)Online publication date: 1-Jul-2019
  • (2019)Smart Emergency Alert System Using Internet of Things and Linked Open Data for Chronic Disease PatientsThe Sundarbans: A Disaster-Prone Eco-Region10.1007/978-3-030-11884-6_17(174-184)Online publication date: 6-Feb-2019
  • (2017)Using programming by demonstration for multimodality in mobile-human interactionsProceedings of the 29th Conference on l'Interaction Homme-Machine10.1145/3132129.3132154(243-251)Online publication date: 29-Aug-2017
  • (2016)An IDE for multimodal controls in smart buildingsProceedings of the 18th ACM International Conference on Multimodal Interaction10.1145/2993148.2993162(61-65)Online publication date: 31-Oct-2016
  • (2016)Extensible Multimodal Annotation for Intelligent Interactive SystemsMultimodal Interaction with W3C Standards10.1007/978-3-319-42816-1_3(37-64)Online publication date: 18-Nov-2016
  • (2016)Applications of the Multimodal Interaction Architecture in Ambient Assisted LivingMultimodal Interaction with W3C Standards10.1007/978-3-319-42816-1_12(271-291)Online publication date: 18-Nov-2016
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media