Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Audio Musical Dice Game: A User-Preference-Aware Medley Generating System

Published: 02 June 2015 Publication History

Abstract

This article proposes a framework for creating user-preference-aware music medleys from users' music collections. We treat the medley generation process as an audio version of a musical dice game. Once the user's collection has been analyzed, the system is able to generate various pleasing medleys. This flexibility allows users to create medleys according to the specified conditions, such as the medley structure or the must-use clips. Even users without musical knowledge can compose medley songs from their favorite tracks. The effectiveness of the system has been evaluated through both objective and subjective experiments on individual components in the system.

References

[1]
Claudio Baccigalupo and Enric Plaza. 2006. Case-based sequential ordering of songs for playlist recommendation. In Advances in Case-Based Reasoning, Lecture Notes in Computer Science, vol. 4106. 286--300.
[2]
Luke Barrington, Antoni B. Chan, and Gert Lanckriet. 2010. Modeling music as a dynamic texture. IEEE Trans. Audio Speech Lang. Process. 18, 3, 602--612.
[3]
Sumit Basu. 2004. Mixing with Mozart. In Proceedings of the International Computer Music Conference.
[4]
Gilberto Bernardes, Carlos Guedes, and Bruce Pennycook. 2012. EarGram: An application for interactive exploration of large databases of audio snippets for creative purposes. In Proceedings of the Symposium on Computer Music Multidisciplinary Research. 19--22.
[5]
Charles Burkhart. 2005. The phrase rhythm of Chopin's A-Flat Major Mazurka, Op. 59, No. 2. In Engaging Music: Essays in Music Analysis, Deborah Jane Stein (Ed.), Oxford University Press, 3--12.
[6]
Luca Chiarandini, Massimiliano Zanoni, and Augusto Sarti. 2011. A system for dynamic playlist generation driven by multimodal control signals and descriptors. In Proceedings of the International Workshop on Multimedia Signal Processing.
[7]
Dave Cliff. 2000. Hang the DJ: Automatic sequencing and seamless mixing of dance-music tracks. Tech. Rep. HP Labs.
[8]
Richard Cole and Ed Schwartz. 2012. Virginia Tech Multimedia Music Dictionary. http://www.music.vt.edu/musicdictionary/.
[9]
David Cope. 1987. Experiments in music intelligence. In Proceedings of the International Computer Music Conference.
[10]
David Cope. 1996. Experiments in Musical Intelligence. A-R Editions, Madison, WI.
[11]
Roger B. Dannenberg. 2006. Concatenative synthesis using score-aligned transcriptions music analysis and segmentation. In Proceedings of the International Computer Music Conference. 352--355.
[12]
Matthew E. P. Davies, Philippe Hamel, Kazuyoshi Yoshii, and Masataka Goto. 2013. AutoMashUpper: An automatic multi-song mashup system. In Proceedings of the International Society for Music Information Retrieval.
[13]
Steven B. Davis and Paul Mermelstein. 1980. Comparison of parametric representations for monosyllabic word recognition in continuously spoken sentences. IEEE Trans. Audio Speech Lang. Process. 28, 4, 357--366.
[14]
Simon Dixon. 2007. Evaluation of the audio beat tracking system BeatRoot. J. New Music Res. 36, 1, 39--50.
[15]
Mark Dolson. 1986. The phase vocoder: A tutorial. Computer Music J. 10, 4, 14--27.
[16]
Arthur Flexer, Dominik Schnitzer, Martin Gasser, and Gerhard Widmer. 2008. Playlist generation using start and end songs. In Proceedings of the International Society for Music Information Retrieval. 173--178.
[17]
Jonathan Foote. 1999. Visualizing music and audio using self-similarity. In Proceedings of the. ACM Multimedia Conference. 77--80.
[18]
G. David Forney Jr. 1973. The Viterbi algorithm. Proc. IEEE 61, 3, 302--309.
[19]
Garth Griffin, Y. E. Kim, and Douglas Turnbull. 2010. Beat-Sync-Mash-Coder: A web application for real-time creation of beat-synchronous music mashups. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing.
[20]
Pierre Hanna, Pascal Ferraro, and Matthias Robine. 2007. On optimizing the editing algorithms for evaluating similarity between monophonic musical sequences. J. New Music Res. 36, 4, 267--279.
[21]
Pierre Hanna, Matthias Robine, and Thomas Rocher. 2009. An alignment based system for chord sequence retrieval. In Proceedings of the Joint Conference on Digital Libraries. 101.
[22]
Hiromi Ishizaki, Keiichiro Hoashi, and Yasuhiro Takishima. 2009. Full-automatic DJ mixing system with optimal tempo adjustment based on measurement function of user discomfort. In Proceedings of the International Society for Music Information Retrieval. 135--140.
[23]
Tristan Jehan. 2005. Creating music by listening. PhD Dissertation, Massachusetts Institute of Technology.
[24]
Jesper Højvang Jensen, Mads Græsbø ll Christensen, Daniel P. W. Ellis, and Søren Holdt Jensen. 2009. Quantitative analysis of a common audio similarity measure. IEEE Trans. Audio Speech Lang. Process. 17, 4, 693--703.
[25]
Mohsen Kamalzadeh, Dominikus Baur, and Torsten Möller. 2012. A survey on music listening and management behaviors. In Proceedings of the International Society for Music Information Retrieval. 373--378.
[26]
Ryoho Kobayashi. 2003. Sound clustering synthesis using spectral data. In Proceedings of the International Computer Music Conference.
[27]
Alison Latham. 2011. The Oxford Companion to Music. http://www.oxfordmusiconline.com/subscriber/book/omo t114.
[28]
Ioannidis Leonidas and Jean-Luc Rouas. 2012. Exploiting semantic content for singing voice detection. In Proceedings of the IEEE International Conference on Semantic Computing (ICSC'12). 134--137.
[29]
Yipeng Li and DeLiang Wang. 2007. Separation of singing voice from music accompaniment for monaural recordings. IEEE Trans. Audio Speech Lang. Process. 15, 4, 1475--1487.
[30]
Rensis Likert. 1932. A technique for the measurement of attitudes. Archives Psychol. 22, 140, 1--55.
[31]
Heng-Yi Lin, Yin-Tzu Lin, Ming-Chun Tien, and Ja-Ling Wu. 2009. Music paste: Concatenating music clips based on chroma and rhythm features. In Proceedings of the International Society for Music Information Retrieval. Kobe.
[32]
Qian Lin, Lie Lu, ChristopherWeare, and Frank Seide. 2010. Music rhythm characterization with application to workout-mix generation. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing. 69--72.
[33]
I-Ting Liu, Yin-Tzu Lin, and Ja-Ling Wu. 2013. Music cut and paste: A personalized musical medley generating system. In Proceedings of the International Society for Music Information Retrieval.
[34]
Zhang Liu, Chaokun Wang, Jianmin Wang, Hao Wang, and Yiyuan Bai. 2012. Adaptive music resizing with stretching, cropping and insertion. Multimedia Syst. 19, 4, 359--380.
[35]
Beth Logan. 2002. Content-based playlist generation: Exploratory experiments. In Proceedings of the International Society for Music Information Retrieval. 2--3.
[36]
Beth Logan and Stephen Chu. 2000. Music summarization using key phrases. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing. Vol. 2. 749--752.
[37]
Gareth Loy. 2006. Musimathics. Vol. 1, MIT Press, 295--296,347--350.
[38]
Namunu C. Maddage, Changsheng Xu, Mohan S. Kankanhalli, and Xi Shao. 2004. Content-based music structure analysis with applications to music semantics understanding. In Proceedings of the. ACM Multimedia Conference. 112--119.
[39]
Yizhao Ni, Matt McVicar, Paul Santos-Rodriguez, and Tijl De Bie. 2012. An end-to-end machine learning system for harmonic analysis of music. IEEE Trans. Audio Speech Lang. Process. 20, 6, 1771--1783.
[40]
Oriol Nieto, M. M. Farbood, Tristan Jehan, and J. P. Bello. 2014. Perceptual analysis of the f-measure for evaluating section boundaries in music. In Proceedings of the International Society for Music Information Retrieval. 265--270.
[41]
Tin Lay Nwe, Arun Shenoy, and Ye Wang. 2004. Singing voice detection in popular music. In Proceedings of the. ACM Multimedia Conference. 324--327.
[42]
Jouni Paulus, Meinard Müller, and Anssi Klapuri. 2010. Audio-based music structure analysis. In Proceedings of the International Society for Music Information Retrieval. 625--636.
[43]
Johan Pauwels, Florian Kaiser, and Geoffroy Peeters. 2013. Combining harmony-based and novelty-based approaches for structural segmentation. In Proceedings of the International Society for Music Information Retrieval.
[44]
Don Michael Randel. 2003. The Harvard Dictionary of Music. Belknap Press.
[45]
Lise Regnier and Geoffroy Peeters. 2009. Singing voice detection in music tracks using direct voice vibrato detection. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing. 1685--1688.
[46]
Diemo Schwarz. 2005. Current research in concatenative sound synthesis. In Proceedings of the International Computer Music Conference.
[47]
Diemo Schwarz. 2007. Corpus-based concatenative synthesis. IEEE Signal Process Mag. 24, 2, 92--104.
[48]
Diemo Schwarz, Roland Cahen, and Sam Britton. 2008. Principles and applications of interactive corpus based concatenative synthesis. Journées d'Informatique Musicale.
[49]
Man-Kwan Shan and Shih-Chuan Chiu. 2010. Algorithmic compositions based on discovered musical patterns. Multimedia Tools Appl. 46, 1, 1--23.
[50]
Mohammad Soleymani, Micheal N. Caro, Erik M. Schmidt, Cheng-Ya Sha, and Yi-Hsuan Yang. 2013. 1000 songs for emotional analysis of music. In Proceedings of the International ACM Workshop on Crowdsourcing for Multimedia. 1--6.
[51]
Min-Yian Su, Yi-Hsuan Yang, Yu-Ching Lin, and Homer H. Chen. 2009. An integrated approach to music boundary detection. In Proceedings of the International Society for Music Information Retrieval. 705--710.
[52]
Douglas Turnbull and Gert Lanckriet. 2007. A supervised approach for detecting boundaries in music using difference features and boosting. In Proceedings of the International Society for Music Information Retrieval. 42--49.
[53]
Stephen Webber. 2007. DJ Skills: The Essential Guide to Mixing and Scratching. Focal Press. 212--239.
[54]
Ron J. Weiss and Juan Pablo Bello. 2010. Identifying repeated patterns in music using sparse convolutive non-negative matrix factorization. In Proceedings of the International Society for Music Information Retrieval.
[55]
Stephan Wenger and Marcus Magnor. 2011. Constrained example-based audio synthesis. In Proceedings of the International Conference on Multimedia and Expo.
[56]
Robert B. Zajonc. 1968. Attitudinal effects of mere exposure. J. Personality Social Psych. 9, 2, Part 2, 1--27.

Cited By

View all
  • (2022)Automatic Detection of Cue Points for the Emulation of DJ MixingComputer Music Journal10.1162/comj_a_0065246:3(67-82)Online publication date: 1-Sep-2022
  • (2021)Embodiment in 18th Century Depictions of Human-Machine Co-CreativityFrontiers in Robotics and AI10.3389/frobt.2021.6620368Online publication date: 28-Jun-2021
  • (2019)A Dual-Camera Surveillance Video Summarization Generating Strategy for Multi-Target CapturingProceedings of the 3rd International Conference on Video and Image Processing10.1145/3376067.3376071(121-125)Online publication date: 20-Dec-2019
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Transactions on Multimedia Computing, Communications, and Applications
ACM Transactions on Multimedia Computing, Communications, and Applications  Volume 11, Issue 4
April 2015
231 pages
ISSN:1551-6857
EISSN:1551-6865
DOI:10.1145/2788342
Issue’s Table of Contents
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 02 June 2015
Accepted: 01 December 2014
Revised: 01 July 2014
Received: 01 February 2014
Published in TOMM Volume 11, Issue 4

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Music editing
  2. concatenating music
  3. musical medley

Qualifiers

  • Research-article
  • Research
  • Refereed

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)9
  • Downloads (Last 6 weeks)1
Reflects downloads up to 22 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2022)Automatic Detection of Cue Points for the Emulation of DJ MixingComputer Music Journal10.1162/comj_a_0065246:3(67-82)Online publication date: 1-Sep-2022
  • (2021)Embodiment in 18th Century Depictions of Human-Machine Co-CreativityFrontiers in Robotics and AI10.3389/frobt.2021.6620368Online publication date: 28-Jun-2021
  • (2019)A Dual-Camera Surveillance Video Summarization Generating Strategy for Multi-Target CapturingProceedings of the 3rd International Conference on Video and Image Processing10.1145/3376067.3376071(121-125)Online publication date: 20-Dec-2019
  • (2019)Sequencing the musical sections with deep learning2019 International Joint Conference on Neural Networks (IJCNN)10.1109/IJCNN.2019.8851935(1-7)Online publication date: Jul-2019
  • (2018)Generating music medleys via playing music puzzle gamesProceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence10.5555/3504035.3504313(2281-2288)Online publication date: 2-Feb-2018
  • (2018)Data MusicalizationACM Transactions on Multimedia Computing, Communications, and Applications10.1145/318474214:2(1-27)Online publication date: 25-Apr-2018
  • (2017)The CrossSong Puzzle: Developing a Logic Puzzle for Musical ThinkingJournal of New Music Research10.1080/09298215.2017.130351946:3(213-228)Online publication date: 21-Mar-2017
  • (2015)Bridging Music Using Sound-Effect InsertionIEEE MultiMedia10.1109/MMUL.2015.7122:4(30-38)Online publication date: Oct-2015

View Options

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media