Polyphonic Music Sequence Transduction With Meter-Constrained Lstm Networks

2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP)(2018)

Cited 21|Views11
No score
Abstract
Automatic transcription of polyphonic music remains a challenging task in the field of Music Information Retrieval. In this paper, we propose a new method to post-process the output of a multi-pitch detection model using recurrent neural networks. In particular, we compare the use of a fixed sample rate against a meter-constrained time step on a piano performance audio dataset. The metric ground truth is estimated using automatic symbolic alignment, which we make available for further study. We show that using musically-relevant time steps improves system performance despite the choice of a basic representation, although mostly because it quantises the output durations. This is an encouraging result for further investigation of musically-motivated neural network designs.
More
Translated text
Key words
Multi-pitch detection, automatic music transcription, music language models, long short term memory networks
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined