Skip to main content
Article
Learning Hierarchical Metrical Structure Beyond Measures
arXiv
  • Junyan Jiang, Music X Lab, NYU Shanghai, China & Mohamed bin Zayed University of Artificial Intelligence
  • Daniel Chin, Music X Lab, NYU Shanghai, China & Mohamed bin Zayed University of Artificial Intelligence
  • Yixiao Zhang, Music X Lab, NYU Shanghai, China & Centre for Digital Music, QMUL, United Kingdom
  • Gus Xia, Music X Lab, NYU Shanghai, China & Mohamed bin Zayed University of Artificial Intelligence
Document Type
Article
Abstract

Music contains hierarchical structures beyond beats and measures. While hierarchical structure annotations are helpful for music information retrieval and computer musicology, such annotations are scarce in current digital music databases. In this paper, we explore a data-driven approach to automatically extract hierarchical metrical structures from scores. We propose a new model with a Temporal Convolutional Network-Conditional Random Field (TCN-CRF) architecture. Given a symbolic music score, our model takes in an arbitrary number of voices in a beat-quantized form, and predicts a 4-level hierarchical metrical structure from downbeat-level to section-level. We also annotate a dataset using RWC-POP MIDI files to facilitate training and evaluation. We show by experiments that the proposed method performs better than the rule-based approach under different orchestration settings. We also perform some simple musicological analysis on the model predictions. All demos, datasets and pre-trained models are publicly available on Github 1 © 2022, CC BY.

DOI
10.48550/arXiv.2209.10259
Publication Date
9-21-2022
Keywords
  • Audio and Speech Processing (eess.AS),
  • Machine Learning (cs.LG),
  • Sound (cs.SD)
Comments

Preprint: arXiv

Archived with thanks to arXiv

Preprint License: CC by 4.0

Uploaded 31 October 2022

Citation Information
J. Jiang, D. Chin, Y. Zhang, and G. Xia, "Learning Hierarchical Metrical Structure Beyond Measures", 2022, doi: 10.48550/arXiv.2209.10259