Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3660570.3660574acmotherconferencesArticle/Chapter ViewAbstractPublication PagesdlfmConference Proceedingsconference-collections
research-article

Acoustic Classification of Guitar Tunings with Deep Learning

Published: 27 June 2024 Publication History
  • Get Citation Alerts
  • Abstract

    A guitar tuning is the allocation of pitches to the open strings of the guitar. A wide variety of guitar tunings are featured in genres such as blues, classical, folk, and rock. Standard tuning provides a convenient placing of intervals and a manageable selection of fingerings. However, numerous other tunings are frequently used as they offer different harmonic possibilities and playing methods.
    A robust method for the acoustic classification of guitar tunings would provide the following benefits for digital libraries for musicology: (i) guitar tuning tags could be assigned to music recordings; these tags could be used to better organise, retrieve, and analyse music in digital libraries, (ii) tuning classification could be integrated into an automatic music transcription system, thus facilitating the production of more accurate and fine-grained symbolic representations of guitar recordings, (iii) insights acquired through guitar tunings research, would be helpful when designing systems for indexing, analysing, and transcribing other string instruments.
    Neural networks offer a promising approach for the automated identification of guitar tunings as they can learn useful features for complex discriminative tasks. Furthermore, they can learn directly from unstructured data, thereby reducing the need for elaborate feature extraction techniques.
    Thus, we evaluate the potential of neural networks for the acoustic classification of guitar tunings. A dataset of authentic song recordings, which featured polyphonic acoustic guitar performances in various tunings, was compiled and annotated. Additionally, a dataset of synthetic polyphonic guitar audio in 5 different tunings was generated with sample-based audio software and tablatures. Using audio converted into log mel spectrograms and chromagrams as input, convolutional neural networks were trained to classify guitar tunings. The resulting models were tested using unseen data from disparate recording conditions. The best performing systems attained a classification accuracy of 97.5% (2 tuning classes) and 73.9% (5 tuning classes).
    This research provides evidence that neural networks can classify guitar tunings from music audio recordings; produces novel annotated datasets that contain authentic and synthetic guitar audio, which can serve as a benchmark for future guitar tuning research; proposes new methods for the collection, annotation, processing, and synthetic generation of guitar data.

    References

    [1]
    Sviatoslav Abakumov. 2023. PyGuitarPro. Retrieved 2024-02-21 from https://github.com/Perlence/PyGuitarPro?tab=readme-ov-file
    [2]
    Jakob Abeßer. 2013. Automatic String Detection for Bass Guitar and Electric Guitar. In From Sounds to Music and Emotions, Mitsuko Aramaki, Mathieu Barthet, Richard Kronland-Martinet, and Sølvi Ystad (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 333–352. https://doi.org/10.1007/978-3-642-41248-6_18
    [3]
    Ana M. Barbancho, Anssi Klapuri, Lorenzo J. Tardon, and Isabel Barbancho. 2012. Automatic Transcription of Guitar Chords and Fingering From Audio. IEEE Transactions on Audio, Speech, and Language Processing 20, 3 (2012), 915–921. https://doi.org/10.1109/TASL.2011.2174227
    [4]
    Emmanouil Benetos, Simon Dixon, Zhiyao Duan, and Sebastian Ewert. 2019. Automatic Music Transcription: An Overview. IEEE Signal Processing Magazine 36, 1 (2019), 20–30. https://doi.org/10.1109/MSP.2018.2869928
    [5]
    Joel Bernstein and Daniel Libertino. 1996. Joni Mitchell Complete Guitar Songbook Edition. Alfred Publishing Co., Inc., Los Angeles, CA, USA.
    [6]
    David Braun. 2021. DawDreamer: Bridging the Gap Between Digital Audio Workstations and Python Interfaces. https://doi.org/10.48550/arXiv.2111.09931 arxiv:2111.09931 [cs.SD]
    [7]
    Gregory Burlet and Abram Hindle. 2017. Isolated guitar transcription using a deep belief network. PeerJ Computer Science 3 (2017), e109–e109. https://doi.org/10.7717/peerj-cs.109
    [8]
    Keunwoo Choi, György Fazekas, Kyunghyun Cho, and Mark Sandler. 2017. A Tutorial on Deep Learning for Music Information Retrieval. (2017). https://doi.org/10.48550/arXiv.1709.04396 arXiv:1709.04396
    [9]
    Frank Cwitkowitz, Toni Hirvonen, and Anssi Klapuri. 2023. Fretnet: Continuous-Valued Pitch Contour Streaming For Polyphonic Guitar Tablature Transcription. In ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 1–5. https://doi.org/10.1109/ICASSP49357.2023.10094825
    [10]
    Nollene Davies. 1994. The Guitar in Zulu “maskanda” Tradition. The World of Music 36, 2 (1994), 118–137. http://www.jstor.org/stable/43561390
    [11]
    Christian Dittmar, Andreas Männchen, and Jakob Abeber. 2013. Real-time guitar string detection for music education software. In 2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS). 1–4. https://doi.org/10.1109/WIAMIS.2013.6616120
    [12]
    Jesse Engel, Lamtharn Hantrakul, Chenjie Gu, and Adam Roberts. 2020. DDSP: Differentiable Digital Signal Processing. arXiv e-prints (Jan. 2020). https://doi.org/10.48550/arXiv.2001.04643
    [13]
    Xander Fiss and Andres Kwasinski. 2011. Automatic real-time electric guitar audio transcription. In 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 373–376. https://doi.org/10.1109/ICASSP.2011.5946418
    [14]
    Jort F. Gemmeke, Daniel P. W. Ellis, Dylan Freedman, Aren Jansen, Wade Lawrence, R. Channing Moore, Manoj Plakal, and Marvin Ritter. 2017. Audio Set: An ontology and human-labeled dataset for audio events. In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2017). New Orleans, LA.
    [15]
    Ian Goodfellow, Yoshua Bengio, and Aaron Courville. 2016. Deep Learning. MIT Press. http://www.deeplearningbook.org.
    [16]
    GoogleDevlopers. 2022. Classification: ROC Curve and AUC. Retrieved 2023-05-02 from https://developers.google.com/machine-learning/crash-course/classification/roc-and-auc
    [17]
    Thomas Grill and Jan Schlüter. 2017. Two convolutional neural networks for bird detection in audio signals. In 2017 25th European Signal Processing Conference (EUSIPCO). 1764–1768. https://doi.org/10.23919/EUSIPCO.2017.8081512
    [18]
    Mark Hanson. 1995. The complete book of alternate tunings. Accent on Music, West Linn, OR, USA.
    [19]
    Curtis Glenn-Macway Hawthorne, Ian Simon, Rigel Jacob Swavely, Ethan Manilow, and Jesse Engel. 2021. Sequence-to-Sequence Piano Transcription with Transformers. https://doi.org/10.48550/arXiv.2107.09142 arXiv:2107.09142
    [20]
    Romain Hennequin, Anis Khlif, Felix Voituret, and Manuel Moussallam. 2020. Spleeter: a fast and efficient music source separation tool with pre-trained models. The Journal of Open Source Software 5, 50 (June 2020), 2154. https://doi.org/10.21105/joss.02154
    [21]
    Eric J. Humphrey and Juan P. Bello. 2014. From music audio to chord tablature: Teaching deep convolutional networks toplay guitar. In 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 6974–6978. https://doi.org/10.1109/ICASSP.2014.6854952
    [22]
    Yogesh Jadhav, Ashish Patel, Rutvij H. Jhaveri, Roshani Raut, and Saqib Hakak. 2022. Transfer Learning for Audio Waveform to Guitar Chord Spectrograms Using the Convolution Neural Network. Mob. Inf. Syst. 2022 (jan 2022), 11 pages. https://doi.org/10.1155/2022/8544765
    [23]
    David Regnier Jules Cournut, Nicolas Martin. 2021. What are the most used guitar positions?. In 8th International Conference on Digital Libraries for Musicology (DLfM2021). 84–92. https://doi.org/10.1145/3469013.3469024
    [24]
    Christian Kehling, Jakob Abeßer, Christian Dittmar, and Gerald Schuller. 2014. Automatic Tablature Transcription of Electric Guitar Recordings by Estimation of Score-and Instrument-Related Parameters. In DAFx. 219–226.
    [25]
    Varun Khatri and Lukas Dillingham. 2020. Guitar Tuning Identification. Technical Report. University of Rochester, Department of Electrical and Computer Engineering.
    [26]
    Sidrah Liaqat, Narjes Bozorg, Neenu Jose, Patrick Conrey, Antony Tamasi, and Michael T Johnson. 2018. Domain tuning methods for bird audio detection.DCASE.
    [27]
    MasterClass. 2021. How to Use Chord Voicing in Music. Retrieved 2023-05-11 from https://www.masterclass.com/articles/how-to-use-chord-voicing-in-music
    [28]
    Joni Mitchell. 1968. Club 47. Troubador Records - CD 5 060446 070178.
    [29]
    Joni Mitchell. 1968. Philadelphia Folk Festival. Retrieved June 29, 2023 from https://www.youtube.com/watch?v=9futFx1iEU4
    [30]
    Joni Mitchell. 1969. Clouds. Reprise Records - CD 6341.
    [31]
    Joni Mitchell. 1970. Ladies of the Canyon. Reprise Records - CD 6376.
    [32]
    Joni Mitchell. 1971. Blue. Reprise Records - CD 2038.
    [33]
    Joni Mitchell. 1972. For the Roses. Elektra/Asylum Records - CD 7559-60624-2.
    [34]
    Joni Mitchell and David Crosby. 1968. Song to a Seagull. Reprise Records - CD 6293.
    [35]
    Himadri Mukherjee, Ankita Dhar, Sk. Md. Obaidullah, K. C. Santosh, Santanu Phadikar, and Kaushik Roy. 2019. Segregating Musical Chords for Automatic Music Transcription: A LSTM-RNN Approach. In Pattern Recognition and Machine Intelligence. Springer International Publishing, Cham, 427–435. https://doi.org/10.1007/978-3-030-34872-4_47
    [36]
    Meinard Müller. 2015. Fundamentals of Music Processing. Springer, Switzerland. https://doi.org/10.1007/978-3-319-21945-5
    [37]
    Meinard Müller and Tech Anssi Klapuri. 2014. Automatic transcription of bass guitar tracks applied for music genre classification and sound synthesis. Ph. D. Dissertation. Gustav-Kirchhoff-Straße 1, 98693 Ilmenau, Germany.
    [38]
    Chris J Murray and Scott B Whitfield. 2022. Inharmonicity in plucked guitar strings. American Journal of Physics 90, 7 (2022), 487–493. https://doi.org/10.1119/5.0064373
    [39]
    Michael A Nielsen. 2015. Neural networks and deep learning. Vol. 25. Determination press San Francisco, CA, USA. http://neuralnetworksanddeeplearning.com/.
    [40]
    Julien Osmalsky, Jean-Jacques Embrechts, Marc Van Droogenbroeck, and Sébastien Pierard. May 2012. Neural networks for musical chords recognition. In Journées d’informatique musicale (Mons, Belgium).
    [41]
    Guitar Pro. 2023. Tabs. Retrieved 2024-02-17 from https://www.guitar-pro.com/tabs/artists
    [42]
    John William Strutt Baron Rayleigh. 1878. The theory of sound. Vol. 2. Macmillan, London, England.
    [43]
    Rikky Rooksby. 2010. How to Write Songs in Altered Guitar Tunings. Backbeat Books, London, England.
    [44]
    Simon Rouard, Francisco Massa, and Alexandre Défossez. 2023. Hybrid Transformers for Music Source Separation. In ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 1–5. https://doi.org/10.1109/ICASSP49357.2023.10096956
    [45]
    Ample Sound. 2024. Ample Guitar T. Retrieved 2024-02-21 from https://www.amplesound.net/en/pro-pd.asp?id=6
    [46]
    Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. 2014. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. Journal of Machine Learning Research 15, 56 (2014), 1929–1958. http://jmlr.org/papers/v15/srivastava14a.html
    [47]
    Dan Stowell, Michael D. Wood, Hanna Pamuła, Yannis Stylianou, and Hervé Glotin. 2019. Automatic acoustic detection of birds through deep learning: The first Bird Audio Detection challenge. Methods in Ecology and Evolution 10, 3 (2019), 368–380. https://doi.org/10.1111/2041-210X.13103 arXiv:https://besjournals.onlinelibrary.wiley.com/doi/pdf/10.1111/2041-210X.13103
    [48]
    European Broadcasting Union. 2000. Alignment level in digital audio production equipment and recorders. EBU Technical Recommendation R68-2000. https://tech.ebu.ch/docs/r/r068.pdf
    [49]
    Valerio Velardo. 2020. Deep Learning Audio Application from Design to Deployment. Retrieved 2024-02-16 from https://github.com/musikalkemist/Deep-Learning-Audio-Application-From-Design-to-Deployment/tree/master
    [50]
    Andrew Wiggins and Youngmoo E Kim. 2019. Guitar Tablature Estimation with a Convolutional Neural Network. In Proceedings of the 20th International Society for Music Information Retrieval Conference. 284–291. https://doi.org/10.5281/zenodo.3527800
    [51]
    Xiaowei Qin Ximin Li, Xiaodong Wei. 2020. Small-Footprint Keyword Spotting with Multi-Scale Temporal Convolution. In Interspeech 2020. 1987–1991. https://doi.org/10.21437/Interspeech.2020-3177
    [52]
    Yongyi Zang, Yi Zhong, Frank Cwitkowitz, and Zhiyao Duan. 2024. SynthTab: Leveraging Synthesized Data for Guitar Tablature Transcription. https://doi.org/10.48550/arXiv.2309.09085 arxiv:2309.09085 [cs.SD]

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    DLfM '24: Proceedings of the 11th International Conference on Digital Libraries for Musicology
    June 2024
    83 pages
    ISBN:9798400717208
    DOI:10.1145/3660570
    • Editor:
    • David M. Weigl
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 27 June 2024

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. audio datasets
    2. guitar tunings
    3. metadata
    4. music indexing
    5. neural networks
    6. transcription

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    DLfM 2024

    Acceptance Rates

    Overall Acceptance Rate 27 of 48 submissions, 56%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 6
      Total Downloads
    • Downloads (Last 12 months)6
    • Downloads (Last 6 weeks)6

    Other Metrics

    Citations

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media