Have a personal or library account? Click to login
Selective Annotation of Few Data for Beat Tracking of Latin American Music Using Rhythmic Features Cover

Selective Annotation of Few Data for Beat Tracking of Latin American Music Using Rhythmic Features

Open Access
|May 2024

References

  1. 1Araujo Junior, S. M. (1992). Acoustic labor in the timing of everyday life: A critical contribution to the history of samba in Rio de Janeiro. PhD thesis, University of Illinois, Urbana, USA.
  2. 2Böck, S., and Davies, M. E. P. (2020). Deconstruct, analyse, reconstruct: How to improve tempo, beat, and downbeat estimation. In Proceedings of the 21st International Society for Music Information Retrieval Conference, pages 574582, Montreal, Canada.
  3. 3Böck, S., Korzeniowski, F., Schlüter, J., Krebs, F., and Widmer, G. (2016). madmom: A new Python audio and music signal processing library. In Proceedings of the 24th ACM International Conference on Multimedia, pages 11741178, Amsterdam, The Netherlands. DOI: 10.1145/2964284.2973795
  4. 4Böck, S., Krebs, F., and Widmer, G. (2014). A multi-model approach to beat tracking considering heterogeneous music styles. In Proceedings of the 15th International Society for Music Information Retrieval Conference, pages 603608, Taipei, Taiwan.
  5. 5Böck, S., and Schedl, M. (2011). Enhanced beat tracking with context-aware neural networks. In Proceedings of the 14th Conference on Digital Audio Effects, pages 135139, Paris, France.
  6. 6Cano, E., Mora-Ángel, F., Gil, G. A. L., Zapata, J. R., Escamilla, A., Alzate, J. F., and Betancur, M. (2021). Sesquialtera in the Colombian bambuco: Perception and estimation of beat and meter – extended version. Transactions of the International Society for Music Information Retrieval, 4(1):248262. DOI: 10.5334/tismir.118
  7. 7Davies, M. E. P., Bock, S., and Fuentes, M. (2021). Tempo, beat and downbeat estimation. In Proceedings 22nd International Society for Music Information Retrieval Conference, Online. Available: https://tempobeatdownbeat.github.io/tutorial/intro.html.
  8. 8Dixon, S. (2007). Evaluation of the audio beat tracking system BeatRoot. Journal of New Music Research, 36(1):3950. DOI: 10.1080/09298210701653310
  9. 9Fiocchi, D., Buccoli, M., Zanoni, M., Antonacci, F., and Sarti, A. (2018). Beat tracking using recurrent neural network: A transfer learning approach. In Proceedings of the 26th European Signal Processing Conference, pages 19291933, Rome, Italy. DOI: 10.23919/EUSIPCO.2018.8553059
  10. 10Fuentes, M., Maia, L. S., Rocamora, M., Biscainho, L. W. P., Crayencour, H.-C., Essid, S., and Bello, J. P. (2019). Tracking beats and microtiming in Afro-Latin American music using conditional random fields and deep learning. In Proceedings of the 20th International Society for Music Information Retrieval Conference, pages 251258, Delft, The Netherlands.
  11. 11Gonçalves, G., and Costa, O. (2000). The Carioca Groove: The Rio de Janeiro’s Samba Schools Drum Sections. Groove, Rio de Janeiro, Brazil.
  12. 12Gouyon, F., Klapuri, A., Dixon, S., Alonso, M., Tzanetakis, G., Uhle, C., and Cano, P. (2006). An experimental comparison of audio tempo induction algorithms. IEEE Transactions on Audio, Speech, and Language Processing, 14(5):18321844. DOI: 10.1109/TSA.2005.858509
  13. 13Hainsworth, S. W., and Macleod, M. D. (2004). Particle filtering applied to musical tempo tracking. EURASIP Journal on Advances in Signal Processing, 15:23852395. DOI: 10.1155/S1110865704408099
  14. 14Heydari, M., Cwitkowitz, F., and Duan, Z. (2021). BeatNet: CRNN and particle filtering for online joint beat downbeat and meter tracking. In Proceedings of the 22nd International Society for Music Information Retrieval Conference, pages 270277, Online.
  15. 15Holzapfel, A., Davies, M. E. P., Zapata, J. R., Oliveira, J. L., and Gouyon, F. (2012). Selective sampling for beat tracking evaluation. IEEE Transactions on Audio, Speech, and Language Processing, 20(9):25392548. DOI: 10.1109/TASL.2012.2205244
  16. 16Holzapfel, A., Flexer, A., and Widmer, G. (2011). Improving tempo-sensitive and tempo-robust descriptors for rhythmic similarity. In Proceedings of the 8th Sound and Music Computing Conference, pages 247252, Padua, Italy.
  17. 17Holzapfel, A., and Stylianou, Y. (2009). A scale transform based method for rhythmic similarity of music. In Proceedings of the 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, pages 317320, Taipei, Taiwan. DOI: 10.1109/ICASSP.2009.4959584
  18. 18Holzapfel, A., and Stylianou, Y. (2011). Scale transform in rhythmic similarity of music. IEEE Transactions on Audio, Speech, and Language Processing, 19(1):176185. DOI: 10.1109/TASL.2010.2045782
  19. 19Jia, B., Lv, J., and Liu, D. (2019). Deep learning-based automatic downbeat tracking: A brief review. Multimedia Systems, 25(6):617638. DOI: 10.1007/s00530-019-00607-x
  20. 20Kim, B., and Pardo, B. (2018). A human-in-the-loop system for sound event detection and annotation. ACM Transactions on Interactive Intelligent Systems, 8(2):123. DOI: 10.1145/3214366
  21. 21Koch, G., Zemel, R., and Salakhutdinov, R. (2015). Siamese neural networks for one-shot image recognition. In 32nd International Conference on Machine Learning Deep Learning Workshop, Lille, France.
  22. 22Krebs, F., Böck, S., and Widmer, G. (2013). Rhythmic pattern modeling for beat and downbeat tracking in musical audio. In Proceedings of the 14th International Society for Music Information Retrieval Conference, pages 227232, Curitiba, Brazil.
  23. 23Lidy, T., and Rauber, A. (2005). Evaluation of feature extractors and psycho-acoustic transformations for music genre classification. In Proceedings of the 6th International Conference on Music Information Retrieval, pages 3441, London, United Kingdom.
  24. 24Lin, H., and Bilmes, J. A. (2009). How to select a good training-data subset for transcription: Submodular active selection for sequences. In Proceedings of the 10th Annual Conference of the International Speech Communication Association, pages 28592862, Brighton, United Kingdom. DOI: 10.21437/Interspeech.2009-730
  25. 25Maia, L. S., Rocamora, M., Biscainho, L. W. P., and Fuentes, M. (2022). Adapting meter tracking models to Latin American music. In Proceedings of the 23rd International Society for Music Information Retrieval Conference, pages 361368, Bengaluru, India.
  26. 26Maia, L. S., Tomaz, P. D. Jr., Fuentes, M., Rocamora, M., Biscainho, L. W. P., Costa, M. V. M., and Cohen, S. (2018). A novel dataset of Brazilian rhythmic instruments and some experiments in computational rhythm analysis. In Proceedings of the 2018 AES Latin American Congress of Audio Engineering, pages 5360, Montevideo, Uruguay.
  27. 27McInnes, L., Healy, J., and Melville, J. (2020). UMAP: Uniform manifold approximation and projection for dimension reduction. arXiv preprint, arXiv:1802.03426v3.
  28. 28McInnes, L., Healy, J., Saul, N., and Großberger, L. (2018). UMAP: Uniform manifold approximation and projection. The Journal of Open Source Software, 3(29):861. DOI: 10.21105/joss.00861
  29. 29Misgeld, O., Gulz, T., Miniotaitė, J., and Holzapfel, A. (2021). A case study of deep enculturation and sensorimotor synchronization to real music. In Proceedings of the 22nd International Society for Music Information Retrieval Conference, pages 460467, Online.
  30. 30Nunes, L., Rocamora, M., Jure, L., and Biscainho, L. W. P. (2015). Beat and downbeat tracking based on rhythmic patterns applied to the Uruguayan Candombe drumming. In Proceedings of the 16th International Society for Music Information Retrieval Conference, pages 264270, Málaga, Spain.
  31. 31Panteli, M., and Dixon, S. (2016). On the evaluation of rhythmic and melodic descriptors for music similarity. In Proceedings of the 17th International Society for Music Information Retrieval Conference, pages 468474, New York, USA.
  32. 32Pinto, A. S., Böck, S., Cardoso, J. S., and Davies, M. E. P. (2021). User-driven fine-tuning for beat tracking. Electronics, 10(13):1518. DOI: 10.3390/electronics10131518
  33. 33Pinto, A. S., and Davies, M. E. P. (2021). Tapping along to the difficult ones: Leveraging user-input for beat tracking in highly expressive musical content. Kronland-Martinet, R., Ystad, S., and Aramaki, M., editors, Perception, Representations, Image, Sound, Music. CMMR 2019, volume 12631 of Lecture Notes in Computer Science, pages 7590. Springer, Cham, Switzerland. DOI: 10.1007/978-3-030-70210-6_5
  34. 34Pohle, T., Schnitzer, D., Schedl, M., Knees, P., and Widmer, G. (2009). On rhythm and general music similarity. In Proceedings of the 10th International Society for Music Information Retrieval Conference, pages 525530, Kobe, Japan.
  35. 35Rocamora, M. (2018). Computational Methods for Percussion Music Analysis: The Afro-Uruguayan Candombe Drumming as a Case Study. PhD thesis, Universidad de la República, Montevideo, Uruguay.
  36. 36Rocamora, M., Jure, L., Marenco, B., Fuentes, M., Lanzaro, F., and Gómez, A. (2015). An audio-visual database of Candombe performances for computational musicological studies. In Memorias del II Congreso Internacional de Ciencia y Tecnología Musical, pages 1724, Buenos Aires, Argentina.
  37. 37Sarasúa, Á., Laurier, C., and Herrera, P. (2012). Support vector machine active learning for music mood tagging. In Proceedings of the 9th International Symposium on Computer Music Modelling and Retrieval, pages 518525, London, UK.
  38. 38Settles, B. (2009). Active learning literature survey. Computer Sciences Technical Report 1648, University of Wisconsin–Madison, Madison, USA.
  39. 39Seyerlehner, K., Widmer, G., and Pohle, T. (2010). Fusing block-level features for music similarity estimation. In Proceedings of the 13th Conference on Digital Audio Effects, pages 225232, Graz, Austria.
  40. 40Shuyang, Z., Heittola, T., and Virtanen, T. (2017). Active learning for sound event classification by clustering unlabeled data. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing, pages 751755, New Orleans, USA. DOI: 10.1109/ICASSP.2017.7952256
  41. 41Snell, J., Swersky, K., and Zemel, R. (2017). Prototypical networks for few-shot learning. In Advances in Neural Information Processing Systems, volume 30, pages 40774087.
  42. 42Su, H., Kasai, J., Wu, C. H., Shi, W., Wang, T., Xin, J., Zhang, R., Ostendorf, M., Zettlemoyer, L., Smith, N. A., and Yu, T. (2023). Selective annotation makes language models better few-shot learners. In Proceedings of the 11th International Conference on Learning Representations, Kigali, Rwanda.
  43. 43Vinyals, O., Blundell, C., Lillicrap, T., Kavukcuoglu, K., and Wierstra, D. (2016). Matching networks for one shot learning. In Advances in Neural Information Processing Systems, volume 29, pages 36303638.
  44. 44Wang, Y., Cartwright, M., and Bello, J. P. (2022a). Active few-shot learning for sound event detection. In Proceedings of the 23rd Annual Conference of the International Speech Communication Association, pages 15511555, Incheon, Korea. DOI: 10.21437/Interspeech.2022-10907
  45. 45Wang, Y., Salamon, J., Cartwright, M., Bryan, N. J., and Bello, J. P. (2020). Few-shot drum transcription in polyphonic music. In Proceedings of the 21st International Society for Music Information Retrieval Conference, pages 117124, Montreal, Canada.
  46. 46Wang, Y., Stoller, D., Bittner, R. M., and Bello, J. P. (2022b). Few-shot musical source separation. In Proceedings of the 2022 IEEE International Conference on Acoustics, Speech and Signal Processing, pages 121125, Singapore. DOI: 10.1109/ICASSP43922.2022.9747536
  47. 47Williams, W. J., and Zalubas, E. J. (2000). Helicopter transmission fault detection via time-frequency, scale and spectral methods. Mechanical Systems and Signal Processing, 14(4):545559. DOI: 10.1006/mssp.2000.1296
  48. 48Yamamoto, K. (2021). Human-in-the-loop adaptation for interactive musical beat tracking. In Proceedings of the 22nd International Society for Music Information Retrieval Conference, pages 794801, Online.
DOI: https://doi.org/10.5334/tismir.170 | Journal eISSN: 2514-3298
Language: English
Submitted on: Aug 11, 2023
Accepted on: Mar 30, 2024
Published on: May 14, 2024
Published by: Ubiquity Press
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2024 Lucas S. Maia, Martín Rocamora, Luiz W. P. Biscainho, Magdalena Fuentes, published by Ubiquity Press
This work is licensed under the Creative Commons Attribution 4.0 License.