Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/2502081.2502184acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
poster

A multimodal probabilistic model for gesture--based control of sound synthesis

Published: 21 October 2013 Publication History
  • Get Citation Alerts
  • Abstract

    In this paper, we propose a multimodal approach to create the mapping between gesture and sound in interactive music systems. Specifically, we propose to use a multimodal HMM to conjointly model the gesture and sound parameters. Our approach is compatible with a learning method that allows users to define the gesture--sound relationships interactively. We describe an implementation of this method for the control of physical modeling sound synthesis. Our model is promising to capture expressive gesture variations while guaranteeing a consistent relationship between gesture and sound.

    References

    [1]
    F. Bevilacqua, N. Schnell, N. Rasamimanana, B. Zamborlin, and F. Guédy. Online Gesture Analysis and Control of Audio Processing. In Musical Robots and Interactive Multimodal Systems, pages 127--142. Springer, 2011.
    [2]
    M. Brand. Voice puppetry. In Proceedings of the 26th annual conference on Computer graphics and interactive techniques, pages 21--28. 1999.
    [3]
    K. Choi, Y. Luo, and J.-n. Hwang. Hidden Markov model inversion for audio-to-visual conversion in an MPEG-4 facial animation system. The Journal of VLSI Signal Processing, 29(1):51--61, 2001.
    [4]
    R. Fiebrink, P. R. Cook, and D. Trueman. Play-along mapping of musical controllers. In Proceedings of the International Computer Music Conference, 2009.
    [5]
    R. Fiebrink, P. R. Cook, and D. Trueman. Human model evaluation in interactive supervised learning. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pages 147--156. 2011.
    [6]
    S. Fu, R. Gutierrez-Osuna, A. Esposito, P. Kakumanu, and O. Garcia. Audio/visual mapping with cross-modal hidden Markov models. Multimedia, IEEE Transactions on, 7(2):243--252, Apr. 2005.
    [7]
    N. Gillian and R. Knapp. A Machine Learning Toolbox For Musician Computer Interaction. In proceedings ot the International Conference on New Interfaces for Musical Expression, Oslo, Norway, 2011.
    [8]
    G. Hofer. Speech-driven animation using multi-modal hidden Markov models. Phd dissertation, University of Edimburgh, 2009.
    [9]
    L. R. Rabiner. A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition. Proceedings of the IEEE, 77(2):257--286, 1989.
    [10]
    N. Rasamimanana, F. Bevilacqua, J. Bloit, N. Schnell, E. Fléty, A. Cera, U. Petrevski, and J.-L. Frechin. The urban musical game: using sport balls as musical interfaces. In CHI '12 Extended Abstracts on Human Factors in Computing Systems, pages 1027--1030. 2012.
    [11]
    D. Rocchesso, S. Serafin, F. Behrendt, N. Bernardini, R. Bresin, G. Eckel, K. Franinovic, T. Hermann, S. Pauletto, P. Susini, and Y. Visell. Sonic interaction design: sound, information and experience. In CHI '08 Extended Abstracts on Human Factors in Computing Systems, pages 3969--3972. 2008.
    [12]
    D. Wessel and M. Wright. Problems and Prospects for Intimate Musical Control of Computers. Computer Music Journal, 26(3):11--22, Sept. 2002.
    [13]
    E. Yamamoto, S. Nakamura, and K. Shikano. Speech-to-lip movement synthesis based on the EM algorithm using audio-visual HMMs. In Multimedia Signal Processing, 1998 IEEE Second Workshop on, pages 2--5, 1998.
    [14]
    H. Zen, Y. Nankaku, and K. Tokuda. Continuous Stochastic Feature Mapping Based on Trajectory HMMs. Audio, Speech, and Language Processing, IEEE Transactions on, 19(2):417--430, 2011.

    Cited By

    View all
    • (2024)Embodied exploration of deep latent spaces in interactive dance-music performanceProceedings of the 9th International Conference on Movement and Computing10.1145/3658852.3659072(1-9)Online publication date: 30-May-2024
    • (2020)Designing and Evaluating the Usability of a Machine Learning API for Rapid Prototyping Music TechnologyFrontiers in Artificial Intelligence10.3389/frai.2020.000133Online publication date: 3-Apr-2020
    • (2019)Understanding the Role of Interactive Machine Learning in Movement Interaction DesignACM Transactions on Computer-Human Interaction10.1145/328730726:1(1-34)Online publication date: 6-Feb-2019
    • Show More Cited By

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    MM '13: Proceedings of the 21st ACM international conference on Multimedia
    October 2013
    1166 pages
    ISBN:9781450324045
    DOI:10.1145/2502081
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 21 October 2013

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. HMM
    2. gesture
    3. multimodal
    4. music
    5. music performance
    6. sound synthesis

    Qualifiers

    • Poster

    Conference

    MM '13
    Sponsor:
    MM '13: ACM Multimedia Conference
    October 21 - 25, 2013
    Barcelona, Spain

    Acceptance Rates

    MM '13 Paper Acceptance Rate 47 of 235 submissions, 20%;
    Overall Acceptance Rate 995 of 4,171 submissions, 24%

    Upcoming Conference

    MM '24
    The 32nd ACM International Conference on Multimedia
    October 28 - November 1, 2024
    Melbourne , VIC , Australia

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)7
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 09 Aug 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Embodied exploration of deep latent spaces in interactive dance-music performanceProceedings of the 9th International Conference on Movement and Computing10.1145/3658852.3659072(1-9)Online publication date: 30-May-2024
    • (2020)Designing and Evaluating the Usability of a Machine Learning API for Rapid Prototyping Music TechnologyFrontiers in Artificial Intelligence10.3389/frai.2020.000133Online publication date: 3-Apr-2020
    • (2019)Understanding the Role of Interactive Machine Learning in Movement Interaction DesignACM Transactions on Computer-Human Interaction10.1145/328730726:1(1-34)Online publication date: 6-Feb-2019
    • (2018)User-Centred Design Actions for Lightweight Evaluation of an Interactive Machine Learning ToolkitJournal of Science and Technology of the Arts10.7559/citarj.v10i2.50910:2(2)Online publication date: 11-Jul-2018
    • (2018)SERKET: An Architecture for Connecting Stochastic Models to Realize a Large-Scale Cognitive ModelFrontiers in Neurorobotics10.3389/fnbot.2018.0002512Online publication date: 26-Jun-2018
    • (2018)Motion-Sound Mapping through InteractionACM Transactions on Interactive Intelligent Systems10.1145/32118268:2(1-30)Online publication date: 13-Jun-2018
    • (2016)Innovative Tools for Sound Sketching Combining Vocalizations and GesturesProceedings of the Audio Mostly 201610.1145/2986416.2986442(12-19)Online publication date: 4-Oct-2016
    • (2016)Multilayer and Multimodal Fusion of Deep Neural Networks for Video ClassificationProceedings of the 24th ACM international conference on Multimedia10.1145/2964284.2964297(978-987)Online publication date: 1-Oct-2016
    • (2016)SoundGuidesProceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems10.1145/2851581.2892420(2829-2836)Online publication date: 7-May-2016
    • (2016)Promoting Stretching Activity with Smartwatch - A Pilot StudyEntertainment Computing - ICEC 201610.1007/978-3-319-46100-7_19(211-216)Online publication date: 20-Sep-2016
    • Show More Cited By

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media