Have a personal or library account? Click to login
Structural Segmentation of Alap in Dhrupad Vocal Concerts Cover

Abstract

Dhrupad vocal concerts exhibit a temporal evolution through a sequence of homogeneous sections marked by shared rhythmic characteristics. In this work, we address the segmentation of a concert audio’s unmetered improvisatory section into musically meaningful segments at the highest time scale. Motivated by the distinct musical properties of the sections and their corresponding acoustic correlates, we compute a number of features for the segment boundary detection task. Both supervised and unsupervised approaches are tested using a dataset of commercial performance recordings that is manually annotated. The dataset is augmented suitably for training and testing of the models to obtain new insights about the relevance of the different rhythmic, melodic and timbral cues in the automatic boundary detection task. We also explore the use of a convolutional neural network trained on mel-scale magnitude spectrograms for the boundary detection task to observe that while the implicit musical cues are largely learned by the network, it is less robust to deviations from training data characteristics. We conclude that it can be rewarding to investigate knowledge driven features on new genres and tasks, both to achieve reasonable performance outcomes given limited datasets and for drawing a deeper understanding of genre characteristics from the acoustical analyses.
DOI: https://doi.org/10.5334/tismir.64 | Journal eISSN: 2514-3298
Language: English
Submitted on: May 5, 2020
Accepted on: Jul 7, 2020
Published on: Sep 16, 2020
Published by: Ubiquity Press
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2020 Preeti Rao, Thallam Prasad Vinutha, Mattur Ananthanarayana Rohit, published by Ubiquity Press
This work is licensed under the Creative Commons Attribution 4.0 License.