Optical Music Recognition with Convolutional Sequence-to-Sequence Models
| Authors |
|
|---|---|
| Publication date | 10-2017 |
| Host editors |
|
| Book title | ISMIR 2017 |
| Book subtitle | Proceedings of the 18th International Society for Music Information Retrieval Conference : October 23-27, 2017, Suzhou, China |
| ISBN (electronic) |
|
| Event | International Society for Music Information Retrieval Conference |
| Pages (from-to) | 731-737 |
| Publisher | ISMIR |
| Organisations |
|
| Abstract |
Optical Music Recognition (OMR) is an important technology within Music Information Retrieval. Deep learning models show promising results on OMR tasks, but symbol-level annotated data sets of sufficient size to train such models are not available and difficult to develop. We present a deep learning architecture called a Convolutional Sequence-to-Sequence model to both move towards an end-to-end trainable OMR pipeline, and apply a learning process that trains on full sentences of sheet music instead of individually labeled symbols. The model is trained and evaluated on a human generated data set, with various image augmentations based on real-world scenarios. This data set is the first publicly available set in OMR research with sufficient size to train and evaluate deep learning models. With the introduced augmentations a pitch recognition accuracy of 81% and a duration accuracy of 94% is achieved, resulting in a note level accuracy of 80%.
Finally, the model is compared to commercially available methods, showing a large improvements over these applications. |
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.5281/zenodo.1415664 |
| Other links | https://www.ismir.net/conferences/ismir2017.html |
| Downloads |
WelU17
(Final published version)
|
| Permalink to this page | |