Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Is it Violin or Viola? Classifying the Instruments’ Music Pieces using Descriptive Statistics

Published: 16 March 2023 Publication History
  • Get Citation Alerts
  • Abstract

    Classifying music pieces based on their instrument sounds is pivotal for analysis and application purposes. Given its importance, techniques using machine learning have been proposed to classify violin and viola music pieces. The violin and viola are two different instruments with three overlapping strings of the same notes, and it is challenging for ordinary people or even musicians to distinguish the sound produced by these instruments. However, the classification of musical instrument pieces was barely performed by prior research. To solve this problem, we propose a technique using descriptive statistics to reliably distinguish between violin and viola music pieces. Likewise, a similar technique on the basis of histogram is introduced alongside the main descriptive statistics approach. These approaches are derived based on the nature of the instruments’ strings and the range of their pieces. We also solve the problem in the current literature which divide the audio into segments for processing instead of managing the whole song. Thereby, we compile a dataset of recordings that comprises of violin and viola solo pieces from the Baroque, Classical, Romantic, and Modern eras. Experiment results suggest that our approach achieves high accuracy on solo pieces as compared to other methods with 0.97 accuracy on Baroque pieces.

    References

    [1]
    Giulio Agostini, Maurizio Longari, and Emanuele Pollastri. 2003. Musical instrument timbres classification with spectral features. EURASIP J. Adv. Signal Process 2003 (Jan.2003), 5–14. DOI:
    [2]
    Jean-Julien Aucouturier and François Pachet. 2002. Scaling up music playlist generation. In Proceedings of the 2002 IEEE International Conference on Multimedia and Expo, ICME 2002, Lausanne, Switzerland. August 26–29, 2002. Volume I. IEEE Computer Society, Lausanne, Switzerland, 105–108. DOI:
    [3]
    Kemal Avci, Murat Arican, and Kemal Polat. 2018. Machine learning based classification of violin and viola instrument sounds for the same notes. In 26th Signal Processing and Communications Applications Conference, SIU 2018, Izmir, Turkey, May 2–5, 2018. IEEE, Izmir, Turkey, 1–4. DOI:
    [4]
    Jan Van Balen, John Ashley Burgoyne, Frans Wiering, and Remco C. Veltkamp. 2013. An analysis of chorus features in popular song. In Proceedings of the 14th International Society for Music Information Retrieval Conference, ISMIR 2013, Curitiba, Brazil, November 4–8, 2013, Alceu de Souza Britto Jr., Fabien Gouyon, and Simon Dixon (Eds.). ISMIR, Curitiba, Brazil, 107–112. http://www.ppgia.pucpr.br/ismir2013/wp-content/uploads/2013/09/180_Paper.pdf.
    [5]
    M. A. Bartsch and G. H. Wakefield. 2005. Audio thumbnailing of popular music using chroma-based representations. IEEE Transactions on Multimedia 7, 1 (2005), 96–104. DOI:
    [6]
    Rachel M. Bittner, Justin Salamon, Mike Tierney, Matthias Mauch, Chris Cannam, and Juan Pablo Bello. 2014. MedleyDB: A multitrack dataset for annotation-intensive MIR research. In Proceedings of the 15th International Society for Music Information Retrieval Conference, ISMIR 2014, Taipei, Taiwan, October 27–31, 2014, Hsin-Min Wang, Yi-Hsuan Yang, and Jin Ha Lee (Eds.). ISMIR, Taipei, Taiwan, 155–160. http://www.terasoft.com.tw/conf/ismir2014/proceedings/T028_322_Paper.pdf.
    [7]
    Juan J. Bosch, Jordi Janer, Ferdinand Fuhrmann, and Perfecto Herrera. 2012. A comparison of sound segregation techniques for predominant instrument recognition in musical audio signals. In Proceedings of the 13th International Society for Music Information Retrieval Conference, ISMIR 2012, Mosteiro S.Bento Da Vitória, Porto, Portugal, October 8-12, 2012, Fabien Gouyon, Perfecto Herrera, Luis Gustavo Martins, and Meinard Müller (Eds.). FEUP Edições, Porto, Portugal, 559–564. http://ismir2012.ismir.net/event/papers/559-ismir-2012.pdf.
    [8]
    J. Brown, O. Houix, and S. McAdams. 2001. Feature dependence in the automatic identification of musical woodwind instruments. The Journal of the Acoustical Society of America 109 3 (2001), 1064–72.
    [9]
    Szu-Yu Chou, Jyh-Shing Roger Jang, and Yi-Hsuan Yang. 2018. Learning to recognize transient sound events using attentional supervision. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI 2018, July 13-19, 2018, Stockholm, Sweden, Jérôme Lang (Ed.). ijcai.org, Stockholm, Sweden, 3336–3342. DOI:
    [10]
    Aleksandr Diment, Padmanabhan Rajan, Toni Heittola, and Tuomas Virtanen. 2013. Modified group delay feature for musical instrument recognition. In 10th International Symposium on Computer Music Multidisciplinary Research, 15- 18 October 2013, Marseille, France (International Symposium on Computer Music Multidisciplinary Research). LMA, Marseille, France, 431–438.
    [11]
    A. Eronen. 2001. Comparison of features for musical instrument recognition. In Proceedings of the 2001 IEEE Workshop on the Applications of Signal Processing to Audio and Acoustics. IEEE, New Platz, NY, USA, 19–22.
    [12]
    Alan George. 1995. Classical and romantic chamber music for strings. Early Music XXIII, 2 (051995), 341–348. DOI:
    [13]
    Masataka Goto, Hiroki Hashiguchi, Takuichi Nishimura, and Ryuichi Oka. 2002. RWC music database: Popular, classical and jazz music databases. In ISMIR 2002, 3rd International Conference on Music Information Retrieval, Paris, France, October 13-17, 2002, Proceedings. ISMIR, Paris, France, 287–288. http://ismir2002.ismir.net/proceedings/03-SP04-1.pdf.
    [14]
    Y. Han, J. Kim, and K. Lee. 2017. Deep convolutional neural networks for predominant instrument recognition in polyphonic music. IEEE/ACM Transactions on Audio, Speech, and Language Processing 25, 1 (2017), 208–221.
    [15]
    Yun-Ning Hung and Yi-Hsuan Yang. 2018. Frame-level instrument recognition by timbre and pitch. In Proceedings of the 19th International Society for Music Information Retrieval Conference, ISMIR 2018, Paris, France, September 23-27, 2018, Emilia Gómez, Xiao Hu, Eric Humphrey, and Emmanouil Benetos (Eds.). ISMIR, Paris, France, 135–142. http://ismir2018.ircam.fr/doc/pdfs/55_Paper.pdf.
    [16]
    Ian Jolliffe. 2011. Principal Component Analysis. Springer, Berlin, Berlin, 1094–1096. DOI:
    [17]
    Ian Ihor Kaminskyj and Tadeusz Czaszejko. 2005. Automatic recognition of isolated monophonic musical instrument sounds using kNNC. Journal of Intelligent Information Systems 24, 2/3 (2005), 199–221. DOI:
    [18]
    B. Kostek. 2004. Musical instrument classification and duet analysis employing music information retrieval techniques. Proc. IEEE 92, 4 (2004), 712–729.
    [19]
    Frank Kurth and Meinard Müller. 2008. Efficient index-based audio matching. Audio, Speech, and Language Processing, IEEE Transactions on 16 (032008), 382–395.
    [20]
    Yann LeCun, Y. Bengio, and Geoffrey Hinton. 2015. Deep learning. Nature 521 (052015), 436–44. DOI:
    [21]
    Peter Li, Jiyuan Qian, and Tian Wang. 2015. Automatic Instrument Recognition in Polyphonic Music Using Convolutional Neural Networks. (2015). arxiv:cs.SD/1511.05520
    [22]
    Thomas Lidy and Andreas Rauber. 2005. Evaluation of feature extractors and psycho-acoustic transformations for music genre classification. In ISMIR 2005, 6th International Conference on Music Information Retrieval, London, UK, 11-15 September 2005, Proceedings. ISMIR, London, UK, 34–41.
    [23]
    Eric Durian, Matt Hallaron, and Scott Adamson. 1997. University of Iowa musical instrument samples. http://theremin.music.uiowa.edu/MIS.html.
    [24]
    Brian McFee, Colin Raffel, Dawen Liang, Daniel Ellis, Matt McVicar, Eric Battenberg, and Oriol Nieto. 2015. Librosa: Audio and music signal analysis in Python. In Proceedings of the 14th Python in Science Conference. SciPy, Austin, Texas, 18–24. DOI:
    [25]
    Cory McKay and Ichiro Fujinaga. 2006. Musical genre classification: Is it worth pursuing and how can it be improved? In ISMIR 2006, 7th International Conference on Music Information Retrieval, Victoria, Canada, 8–12 October 2006, Proceedings. ISMIR, Victoria, Canada, 101–106.
    [26]
    Meinard Müller, Frank Kurth, and Michael Clausen. 2005. Audio matching via chroma-based statistical features. In ISMIR 2005, 6th International Conference on Music Information Retrieval, London, UK, 11-15 September 2005, Proceedings. ISMIR, London, UK, 288–295.
    [27]
    Taejin Park and Taejin Lee. 2015. Musical instrument sound classification with deep convolutional neural network using feature fusion approach. (2015). arxiv:cs.SD/1512.07370
    [28]
    Freya Parr. 2018. What’s the difference between a violin and a viola? (2018). https://www.classical-music.com/features/articles/what-difference-between-violin-and-viola/.
    [29]
    Geoffroy Peeters, Stephen McAdams, and Perfecto Herrera. 2000. Instrument sound description in the context of MPEG-7. In ICMC: International Computer Music Conference, Berlin, Germany, 166–169. https://hal.archives-ouvertes.fr/hal-01161319.cote interne IRCAM: Peeters00a.
    [30]
    Elizabeth Roche. 2019. Of myths and baroque music. Early Music 47, 1 (032019), 132–135. DOI:
    [31]
    Eva E. Subotnik. 2015. Copyright and the living dead?: Succession law and the postmortem term. Estate Planning eJournal 29, 1 (2015), 82–93.
    [32]
    StringOvation Team. 2018. The Romantic Period of Music. (2018). https://www.connollymusic.com/stringovation/the-romantic-period-of-music.
    [33]
    [34]
    Yuen-Hsien Tseng. 1999. Content-based retrieval for music collections. In Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. Association for Computing Machinery, New York, NY, USA, 176–182. DOI:
    [35]
    Qi Wang and Changchun Bao. 2020. Individual violin recognition method combining tonal and nontonal features. Electronics 9, 6 (Jun.2020), 950. DOI:
    [36]
    Yi Yu, Suhua Tang, Francisco Raposo, and Lei Chen. 2019. Deep cross-modal correlation learning for audio and lyrics in music retrieval. ACM Trans. Multimedia Comput. Commun. Appl. 15, 1, Article 20 (Feb.2019), 16 pages. DOI:

    Index Terms

    1. Is it Violin or Viola? Classifying the Instruments’ Music Pieces using Descriptive Statistics

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Transactions on Multimedia Computing, Communications, and Applications
      ACM Transactions on Multimedia Computing, Communications, and Applications  Volume 19, Issue 2s
      April 2023
      545 pages
      ISSN:1551-6857
      EISSN:1551-6865
      DOI:10.1145/3572861
      • Editor:
      • Abdulmotaleb El Saddik
      Issue’s Table of Contents

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 16 March 2023
      Online AM: 14 September 2022
      Accepted: 24 August 2022
      Revised: 23 May 2022
      Received: 01 December 2021
      Published in TOMM Volume 19, Issue 2s

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. Musical information retrieval
      2. musical pieces classification

      Qualifiers

      • Research-article

      Funding Sources

      • Advanced Engineering Platform’s Cluster Funding

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • 0
        Total Citations
      • 288
        Total Downloads
      • Downloads (Last 12 months)126
      • Downloads (Last 6 weeks)8
      Reflects downloads up to 10 Aug 2024

      Other Metrics

      Citations

      View Options

      Get Access

      Login options

      Full Access

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Full Text

      View this article in Full Text.

      Full Text

      HTML Format

      View this article in HTML Format.

      HTML Format

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media