Audio to Score Matching by Combining Phonetic and Duration Information

TitleAudio to Score Matching by Combining Phonetic and Duration Information
Publication TypeConference Paper
Year of Publication2017
Conference NameThe 18th International Society for Music Information Retrieval Conference
AuthorsGong, R., Pons J., & Serra X.
Conference Start Date23/10/2017
Conference LocationSuzhou, China
AbstractWe approach the singing phrase audio to score matching problem by using phonetic and duration information – with a focus on studying the jingju a cappella singing case. We argue that, due to the existence of a basic melodic contour for each mode in jingju music, only using melodic information (such as pitch contour) will result in an ambiguous matching. This leads us to propose a matching approach based on the use of phonetic and duration information. Phonetic information is extracted with an acoustic model shaped with our data, and duration information is considered with the Hidden Markov Models (HMMs) variants we investigate. We build a model for each lyric path in our scores and we achieve the matching by ranking the posterior probabilities of the decoded most likely state sequences. Three acoustic models are investigated: (i) convolutional neural networks (CNNs), (ii) deep neural networks (DNNs) and (iii) Gaussian mixture models (GMMs). Also, two duration models are compared: (i) hidden semi-Markov model (HSMM) and (ii) post-processor duration model. Results show that CNNs perform better in our (small) audio dataset and also that HSMM outperforms the post-processor duration model.
preprint/postprint documenthttps://arxiv.org/pdf/1707.03547.pdf
intranet