Note: This bibliographic page is archived and will no longer be updated. For an up-to-date list of publications from the Music Technology Group see the Publications list .

Phrase-based Rāga Recognition Using Vector Space Modeling

Title Phrase-based Rāga Recognition Using Vector Space Modeling
Publication Type Conference Paper
Year of Publication 2016
Conference Name 41st IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2016)
Authors Gulati, S. , Serrà J. , Ishwar V. , Şentürk S. , & Serra X.
Pagination 66-70
Conference Start Date 20/3/2016
Publisher IEEE
Conference Location Shanghai, China
Abstract Automatic rāga recognition is one of the fundamental computational tasks in Indian art music. Motivated by the way seasoned listeners identify rāgas, we propose a rāga recognition approach based on melodic phrases. Firstly, we extract melodic patterns from a collection of audio recordings in an unsupervised way. Next, we group similar patterns by exploiting complex networks concepts and techniques. Drawing an analogy to topic modeling in text classification, we then represent audio recordings using a vector space model. Finally, we employ a number of classification strategies to build a predictive model for rāga recognition. To evaluate our approach, we compile a music collection of over 124 hours, comprising 480 recordings and 40 rāgas. We obtain 70% accuracy with the full 40-rāga collection, and up to 92% accuracy with its 10-rāga subset. We show that phrase-based rāga recognition is a successful strategy, on par with the state of the art, and sometimes outperforms it. A by-product of our approach, which arguably is as important as the task of rāga recognition, is the identification of rāga-phrases. These phrases can be used as a dictionary of semantically-meaningful melodic units for several computational tasks in Indian art music.
preprint/postprint document http://hdl.handle.net/10230/32879
Final publication https://doi.org/10.1109/ICASSP.2016.7471638
Additional material:
To access shared resources for this article visit its companion webpage at: http://compmusic.upf.edu/node/278