Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3660515.3664244acmconferencesArticle/Chapter ViewAbstractPublication PageseicsConference Proceedingsconference-collections
short-paper

Exploiting Semantic Search and Object-Oriented Programming to Ease Multimodal Interface Development

Published: 24 June 2024 Publication History
  • Get Citation Alerts
  • Abstract

    Multimodal interaction has been adopted across various platforms and devices, with supporting tools enhancing the developer experience in developing (). While traditionally, these tools faced challenges balancing expressiveness and usability, recent progress in tends to mitigate this rule. However, adding multimodal interaction still remains challenging, especially when integrating the voice modality, and s remain to be better integrated into today’s applications. To address these challenges, we introduce a Unity tool-based system named. allows developers to use their knowledge in Object-Oriented Programming to handle the expert knowledge required to create a while allowing end users great flexibility in the natural language they can use to interact. Our contributions are: 1) the presentation of ’s architecture and its inherent concepts along with its open-source implementation, and 2) a successful evaluation of its usability for describing s through the questionnaire with twelve participants.

    References

    [1]
    Aaron Bangor, Philip Kortum, and James Miller. 2009. Determining what individual SUS scores mean: Adding an adjective rating scale. Journal of usability studies 4, 3 (2009), 114–123.
    [2]
    Richard A. Bolt. 1980. “Put-that-there”: Voice and gesture at the graphics interface. SIGGRAPH Comput. Graph. 14, 3, 262–270. https://doi.org/10.1145/965105.807503
    [3]
    Jullien Bouchet, Laurence Nigay, and Thierry Ganille. 2004. ICARE software components for rapidly developing multimodal interfaces. In Proceedings of the 6th international conference on Multimodal interfaces. 251–258.
    [4]
    Marie-Luce Bourguet. 2002. A toolkit for creating and testing multimodal interface designs. companion proceedings of UIST 2 (2002), 29–30.
    [5]
    John Brooke. 1996. Sus: a “quick and dirty’usability. Usability evaluation in industry 189, 3 (1996), 189–194.
    [6]
    Paul Chandler and John Sweller. 1992. The Split-Attention Effect as a Factor in the Design of Instruction. British Journal of Educational Psychology 62, 2 (1992), 233–246. https://doi.org/10.1111/j.2044-8279.1992.tb01017.x
    [7]
    G. Convertino, J. Chen, Y. Ryu, C. North, and B. Yost. 2003. Exploring Context Switching and Cognition in Dual-View Coordinated Visualizations. In International Conference on Coordinated and Multiple Views in Exploratory Visualization. IEEE, 55. https://doi.org/10.1109/CMV.2003.1215003
    [8]
    Joëlle Coutaz, Laurence Nigay, Daniel Salber, Ann Blandford, Jon May, and Richard M Young. 1995. Four easy pieces for assessing the usability of multimodal interaction: the CARE properties. Human—Computer Interaction: Interact’95 (1995), 115–120.
    [9]
    Fredy Cuenca, Jan Van den Bergh, Kris Luyten, and Karin Coninx. 2015. Hasselt uims: a tool for describing multimodal interactions with composite events. In Proceedings of the 7th ACM SIGCHI Symposium on Engineering Interactive Computing Systems. 226–229.
    [10]
    Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).
    [11]
    Bruno Dumas, Denis Lalanne, and Rolf Ingold. 2008. Prototyping multimodal interfaces with the SMUIML modeling language. In CHI 2008 Workshop on User Interface Description Languages for Next Generation User Interfaces, CHI.
    [12]
    Bruno Dumas, Denis Lalanne, and Rolf Ingold. 2010. Description languages for multimodal interaction: a set of guidelines and its illustration with SMUIML. Journal on multimodal user interfaces 3 (2010), 237–247.
    [13]
    Bruno Dumas, Denis Lalanne, and Sharon Oviatt. 2009. Multimodal Interfaces: A Survey of Principles, Models and Frameworks. Springer Berlin Heidelberg, Berlin, Heidelberg, 3–26. https://doi.org/10.1007/978-3-642-00437-7_1
    [14]
    Bruno Dumas, Beat Signer, and Denis Lalanne. 2011. A graphical uidl editor for multimodal interaction design based on smuiml. (2011).
    [15]
    Jerome Fink, Pierre Poitier, Maxime André, Loup Meurice, Benoît Frénay, Anthony Cleve, Bruno Dumas, and Laurence Meurant. 2023. Sign Language to Text Dictionary with Lightweight Transformer Models. In Proceedings of the 32nd International Joint Conference on Artificial Intelligence (IJCAI 2023): AI for Social Good track.
    [16]
    Charles L Forgy. 1989. Rete: A fast algorithm for the many pattern/many object pattern match problem. In Readings in Artificial Intelligence and Databases. Elsevier, 547–559.
    [17]
    Lode Hoste, Bruno Dumas, and Beat Signer. 2011. Mudra: a unified multimodal interaction framework. In Proceedings of the 13th international conference on multimodal interfaces. 97–104.
    [18]
    Karina Li, Daniel Wan Rosli, Shuning Zhang, Yuhan Zhang, Monica S Lam, James A Landay, 2023. ReactGenie: An Object-Oriented State Abstraction for Complex Multimodal Interactions Using Large Language Models. arXiv preprint arXiv:2306.09649 (2023).
    [19]
    Vivian Genaro Motti. 2020. Wearable Interaction. Springer International Publishing. https://doi.org/10.1007/978-3-030-27111-4
    [20]
    Sharon Oviatt. 1999. Ten myths of multimodal interaction. Commun. ACM 42, 11 (1999), 74–81.
    [21]
    Sharon Oviatt. 2022. Multimodal Interaction, Interfaces, and Analytics. Springer International Publishing, Cham, 1–29. https://doi.org/10.1007/978-3-319-27648-9_22-1
    [22]
    Nils Reimers and Iryna Gurevych. 2019. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, China, 3982–3992. https://doi.org/10.18653/v1/D19-1410
    [23]
    Ritam Jyoti Sarmah, Yunpeng Ding, Di Wang, Cheuk Yin Phipson Lee, Toby Jia-Jun Li, and Xiang’Anthony’ Chen. 2020. Geno: A Developer Tool for Authoring Multimodal Interaction on Existing Web Applications. In Proceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology. 1169–1181.
    [24]
    Jeff Sauro. 2011. A practical guide to the system usability scale: Background, benchmarks & best practices. Measuring Usability LLC.
    [25]
    Marcos Serrano, Laurence Nigay, Jean-Yves L Lawson, Andrew Ramsay, Roderick Murray-Smith, and Sebastian Denef. 2008. The openinterface framework: A tool for multimodal interaction. In CHI’08 Extended abstracts on human factors in computing systems. 3501–3506.
    [26]
    Jacqueline N Stetson and Thomas S Tullis. 2004. A comparison of questionnaires for assessing website usability. UPA Presentation (2004).
    [27]
    Matthew Turk. 2014. Multimodal interaction: A review. Pattern recognition letters 36 (2014), 189–195.
    [28]
    Minh Tue Vo and Cindy Wood. 1996. Building an application framework for speech and pen input integration in multimodal learning interfaces. In 1996 IEEE International Conference on Acoustics, Speech, and Signal Processing Conference Proceedings, Vol. 6. IEEE, 3545–3548.

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    EICS '24 Companion: Companion Proceedings of the 16th ACM SIGCHI Symposium on Engineering Interactive Computing Systems
    June 2024
    129 pages
    ISBN:9798400706516
    DOI:10.1145/3660515
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 24 June 2024

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Fusion Engine Architecture
    2. Multimodal Interaction
    3. Multimodal Interfaces

    Qualifiers

    • Short-paper
    • Research
    • Refereed limited

    Conference

    EICS '24
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 73 of 299 submissions, 24%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 18
      Total Downloads
    • Downloads (Last 12 months)18
    • Downloads (Last 6 weeks)18
    Reflects downloads up to 26 Jul 2024

    Other Metrics

    Citations

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media