Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/2504335.2504382acmotherconferencesArticle/Chapter ViewAbstractPublication PagesConference Proceedings
research-article

Genre classification of symbolic music with SMBGT

Published: 29 May 2013 Publication History
  • Get Citation Alerts
  • Abstract

    Automatic music genre classification is a task that has attracted the interest of the music community for more than two decades. Music can be of high importance within the area of assistive technologies as it can be seen as an assistive technology with high therapeutic and educational functionality for children and adults with disabilities. Several similarity methods and machine learning techniques have been applied in the literature to deal with music genre classification, and as a result data mining and Music Information Retrieval (MIR) are strongly interconnected. In this paper, we deal with music genre classification for symbolic music, and specifically MIDI, by combining the recently proposed novel similarity measure for sequences, SMBGT, with the k-Nearest Neighbor (k-NN) classifier. For all MIDI songs we first extract all of their channels and then transform each channel into a sequence of 2D points, providing information for pitch and duration of their music notes. The similarity between two songs is found by computing the SMBGT for all pairs of the songs' channels and getting the maximum pairwise channel score as their similarity. Each song is treated as a query to which k-NN is applied, and the returned genre of the classifier is the one with the majority of votes in the k neighbors. Classification accuracy results indicate that there is room for improvement, especially due to the ambiguous definitions of music genres that make it hard to clearly discriminate them. Using this framework can also help us analyze and understand potential disadvantages of SMBGT, and thus identify how it can be improved when used for classification of real-time sequences.

    References

    [1]
    N. Adams, M. Bartsch, J. Shifrin, and G. Wakefield. Time series alignment for music information retrieval. In Proceedings of ISMIR, pages 303--311, 2004.
    [2]
    J. Aucouturier and F. Pachet. Representing musical genre: A state of the art. Journal of New Music Research, 32(1):83--93, 2003.
    [3]
    A. Berenzweig, D. Ellis, and S. Lawrence. Anchor space for classification and similarity measurement of music. In Proceedings of ICME, volume 1, pages 1--29, 2003.
    [4]
    L. Bergroth, H. Hakonen, and T. Raita. A survey of longest common subsequence algorithms. In Proceedings of SPIRE, pages 39--48, 2000.
    [5]
    J. Bergstra, N. Casagrande, D. Erhan, D. Eck, and B. Kégl. Aggregate features and a da b oost for music classification. Machine Learning, 65(2):473--484, 2006.
    [6]
    Z. Cataltepe, Y. Yaslan, and A. Sonmez. Music genre classification using midi and audio features. EURASIP Journal on Advances in Signal Processing, 2007(1):275--279, 2007.
    [7]
    R. Cilibrasi, P. Vitányi, and R. Wolf. Algorithmic clustering of music based on string compression. Computer Music Journal, 28(4):49--67, 2004.
    [8]
    R. Dannenberg, B. Thom, and D. Watson. A machine learning approach to musical style recognition. In Proceedings of ICMC, pages 344--347, 1997.
    [9]
    F. Fabbri. A theory of musical genres: Two applications. Popular Music Perspectives, 1:52--81, 1982.
    [10]
    Z. Fu, G. Lu, K. Ting, and D. Zhang. A survey of audio-based music classification and annotation. IEEE Transactions on Multimedia, 13(2):303--319, 2011.
    [11]
    N. Hu, R. Dannenberg, and A. Lewis. A probabilistic model of melodic similarity. In Proceedings of ICMC, pages 509--515, 2002.
    [12]
    C. Iliopoulos and M. Kurokawa. String matching with gaps for musical melodic recognition. In Proceedings of PSC, pages 55--64, 2002.
    [13]
    J. Jang and M. Gao. A query-by-singing system based on dynamic programming. In International Workshop on Intelligent Systems Resolutions, pages 85--89, 2000.
    [14]
    A. Kotsifakos, P. Papapetrou, J. Hollmén, and D. Gunopulos. A subsequence matching with gaps-range-tolerances framework: a query-by-humming application. Proceedings of VLDB, 4(11):761--771, 2011.
    [15]
    A. Kotsifakos, P. Papapetrou, J. Hollmén, D. Gunopulos, and V. Athitsos. A survey of query-by-humming similarity methods. In Proceedings of PETRA, 2012.
    [16]
    A. Kotsifakos, P. Papapetrou, J. Hollmén, D. Gunopulos, V. Athitsos, and G. Kollios. Hum-a-song: a subsequence matching with gaps-range-tolerances query-by-humming system. Proceedings of the VLDB Endowment, 5(12):1930--1933, 2012.
    [17]
    J. B. Kruskall and M. Liberman. The symmetric time warping algorithm: From continuous to discrete. In Time Warps. Addison-Wesley, 1983.
    [18]
    J. Lee and J. Downie. Survey of music information needs, uses, and seeking behaviours: Preliminary findings. In Proceedings of ICMIR, pages 441--446, 2004.
    [19]
    K. Lemström and E. Ukkonen. Including interval encoding into edit distance based music comparison and retrieval. In Proceedings of AISB, pages 53--60, 2000.
    [20]
    T. Lidy and A. Rauber. Evaluation of feature extractors and psycho-acoustic transformations for music genre classification. In Proceedings of ISMIR, pages 34--41, 2005.
    [21]
    S. Lippens, J. Martens, and T. De Mulder. A comparison of human and automatic musical genre classification. In Proceedings of ICASSP, volume 4, pages 233--236, 2004.
    [22]
    M. Mandel and D. Ellis. Song-level features and support vector machines for music classification. In Proceedings of ISMIR, pages 594--599, 2005.
    [23]
    D. Mazzoni and R. Dannenberg. Melody matching directly from audio. In Proceedings of ISMIR, pages 17--18, 2001.
    [24]
    C. McKay and I. Fujinaga. Automatic music classification and the importance of instrument identification. In Proceedings of CIM, 2005.
    [25]
    C. McKay and I. Fujinaga. Musical genre classification: Is it worth pursuing and how can it be improved. In Proceedings of ISMIR, pages 101--107, 2006.
    [26]
    R. McNab, L. Smith, I. Witten, C. Henderson, and S. Cunningham. Towards the digital music library: Tune retrieval from acoustic input. In Proceedings of ICDL, pages 11--18, 1996.
    [27]
    C. Meek and W. Birmingham. A comprehensive trainable error model for sung music queries. Journal of Artificial Intelligence Research, 22(1):57--91, 2004.
    [28]
    M. Mongeau and D. Sankoff. Comparison of musical sequences. Computers and the Humanities, 24(3):161--175, 1990.
    [29]
    F. Pachet, D. Cazaly, et al. A taxonomy of musical genres. In Proceedings of Content-Based Multimedia Information Access (RIAO), pages 1238--1245, 2000.
    [30]
    B. Pardo and W. Birmingham. Encoding timing information for musical query matching. In Proceedings of ISMIR, pages 267--268, 2002.
    [31]
    S. Pauws. Cubyhum: A fully operational query by humming system. In Proceedings of ISMIR, pages 187--196, 2002.
    [32]
    D. Perrot and R. Gjerdigen. Scanning the dial: An exploration of factors in the identification of musical style. In Proceedings of SMPC, page 88, 1999.
    [33]
    Y. Sakurai, C. Faloutsos, and M. Yamamuro. Stream monitoring under the time warping distance. In Proceedings of ICDE, pages 1046--1055, 2007.
    [34]
    N. Scaringella, G. Zoia, and D. Mlynek. Automatic genre classification of music content: a survey. Signal Processing Magazine, 23(2):133--141, 2006.
    [35]
    J. Shifrin, B. Pardo, C. Meek, and W. Birmingham. HMM-based musical query retrieval. In Proceedings of JCDL, pages 295--300, 2002.
    [36]
    H. Soltau, T. Schultz, M. Westphal, and A. Waibel. Recognition of music types. In Proceedings of ICASSP, volume 2, pages 1137--1140, 1998.
    [37]
    G. Tzanetakis and P. Cook. Musical genre classification of audio signals. IEEE Transactions on Speech and Audio Processing, 10(5):293--302, 2002.
    [38]
    A. Uitdenbogerd and J. Zobel. Melodic matching techniques for large music databases. In Proceedings of ACM Multimedia (Part 1), pages 57--66, 1999.
    [39]
    E. Unal, E. Chew, P. Georgiou, and S. Narayanan. Challenging uncertainty in query by humming systems: a fingerprinting approach. Transactions on Audio Speech and Language Processing, 16(2):359--371, 2008.
    [40]
    Y. Zhu and D. Shasha. Warping indexes with envelope transforms for query by humming. In Proceedings of SIGMOD, pages 181--192, 2003.

    Cited By

    View all
    • (2022)A Content Analysis of the Research Approaches in Music Genre Recognition2022 International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA)10.1109/HORA55278.2022.9799935(1-13)Online publication date: 9-Jun-2022
    • (2022)Genre Recognition from Symbolic Music with CNNs: Performance and ExplainabilitySN Computer Science10.1007/s42979-022-01490-64:2Online publication date: 17-Dec-2022
    • (2021)Efficient Retrieval of Music Recordings Using Graph-Based Index StructuresSignals10.3390/signals20200212:2(336-352)Online publication date: 17-May-2021
    • Show More Cited By

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    PETRA '13: Proceedings of the 6th International Conference on PErvasive Technologies Related to Assistive Environments
    May 2013
    413 pages
    ISBN:9781450319737
    DOI:10.1145/2504335
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    • NSF: National Science Foundation
    • FORTH: Foundation for Research and Technology - Hellas
    • HERACLEIA: HERACLEIA Human-Centered Computing Laboratory at UTA
    • U of Tex at Arlington: U of Tex at Arlington
    • TEI: Technological Educational Institution of Athens
    • UCG: University of Central Greece
    • NCRS: Demokritos National Center for Scientific Research
    • Fulbrigh, Greece: Fulbright Foundation, Greece
    • Ionian: Ionian University, GREECE

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 29 May 2013

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. MIDI
    2. SMBGT
    3. classification
    4. genre
    5. sequence

    Qualifiers

    • Research-article

    Funding Sources

    Conference

    PETRA '13
    Sponsor:
    • NSF
    • FORTH
    • HERACLEIA
    • U of Tex at Arlington
    • TEI
    • UCG
    • NCRS
    • Fulbrigh, Greece
    • Ionian

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)10
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 09 Aug 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2022)A Content Analysis of the Research Approaches in Music Genre Recognition2022 International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA)10.1109/HORA55278.2022.9799935(1-13)Online publication date: 9-Jun-2022
    • (2022)Genre Recognition from Symbolic Music with CNNs: Performance and ExplainabilitySN Computer Science10.1007/s42979-022-01490-64:2Online publication date: 17-Dec-2022
    • (2021)Efficient Retrieval of Music Recordings Using Graph-Based Index StructuresSignals10.3390/signals20200212:2(336-352)Online publication date: 17-May-2021
    • (2021)Genre Recognition from Symbolic Music with CNNsArtificial Intelligence in Music, Sound, Art and Design10.1007/978-3-030-72914-1_7(98-114)Online publication date: 2-Apr-2021
    • (2020)A Novel Approach to Music Genre Classification using Natural Language Processing and Spark2020 14th International Conference on Ubiquitous Information Management and Communication (IMCOM)10.1109/IMCOM48794.2020.9001675(1-8)Online publication date: Jan-2020
    • (2018)Towards the use of similarity distances to music genre classification: A comparative studyPLOS ONE10.1371/journal.pone.019141713:2(e0191417)Online publication date: 14-Feb-2018
    • (2017)Music Genre Classification: A N-Gram Based Musicological Approach2017 IEEE 7th International Advance Computing Conference (IACC)10.1109/IACC.2017.0141(671-677)Online publication date: Jan-2017
    • (2017)Genre classification of symbolic pieces of musicJournal of Intelligent Information Systems10.1007/s10844-016-0430-748:3(579-599)Online publication date: 1-Jun-2017

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media