| --- |
| library_name: symupe |
| license: cc-by-nc-sa-4.0 |
| datasets: |
| - loubb/aria-midi |
| pipeline_tag: feature-extraction |
| tags: |
| - music |
| - piano |
| - midi |
| - pre-training |
| - transformer |
| - MLM |
| --- |
| |
| # SyMuPe: Aria-MIDI MLM Backbone |
|
|
| **Aria-MIDI-MLM** is a 12-layer Transformer encoder designed for symbolic piano music feature extraction. |
| It was pre-trained using a **Multi-Mask Language Modeling (mMLM)** objective on 371,053 diverse piano MIDI files from the deduped subset of [Aria-MIDI](https://huggingface.co/datasets/loubb/aria-midi) dataset. |
|
|
| This model serves as the foundation for the [MIDI Quality Classifier](https://huggingface.co/SyMuPe/MIDI-Quality-Classifier), presented in the article: [**PianoCoRe: Combined and Refined Piano MIDI Dataset**](https://doi.org/10.5334/tismir.333). |
|
|
| - **TISMIR:** https://doi.org/10.5334/tismir.333 |
| - **arXiv:** https://arxiv.org/abs/2605.06627 |
| - **SyMuPe:** https://github.com/ilya16/SyMuPe |
| - **PianoCoRe:** https://github.com/ilya16/PianoCoRe |
| - **Dataset:** https://huggingface.co/datasets/loubb/aria-midi |
|
|
| ## Architecture |
|
|
| - **Type:** Transformer Encoder |
| - **Configuration:** 12 layers, 768 hidden dimensions, 12 attention heads. |
| - **Objective:** Multi-Mask Language Modeling (mMLM). |
| - **Inputs (score-agnostic):** `Pitch`, `Velocity`, `TimeShift`, `Duration`, absolute `TimePosition` |
| - **Training:** Pre-trained for 600,000 steps on 512-note sequences sampled from the deduplicated [Aria-MIDI](https://huggingface.co/datasets/loubb/aria-midi) corpus. |
|
|
| ## Quick Start |
|
|
| Before using this model, ensure you have the `symupe` library installed: |
| ```shell |
| pip install -U symupe |
| ``` |
|
|
| Use the following code to embed MIDI files: |
| ```python |
| import torch |
| from symupe import AutoEmbedder |
| |
| device = torch.device("cuda" if torch.cuda.is_available() else "cpu") |
| |
| # Build Embedder by loading the model and tokenizer directly from the Hub |
| embedder = AutoEmbedder.from_pretrained("SyMuPe/Aria-MIDI-MLM", device=device) |
| # model, tokenizer = embedder.model, embedder.tokenizer |
| |
| # Extract embeddings from a MIDI file |
| result = embedder("performance.mid", max_seq_len=512, hop_size=256, layer=-1) |
| # result is MusicEmbeddingResult(...) containing: |
| # - midi, seq, embeddings, memory_tokens, token_embeddings, hidden_states, sequences and window_indices |
| |
| print(result.embeddings.shape) # (windows, seq_len, emb_dim) |
| ``` |
|
|
| ## License |
|
|
| The model weights are distributed under the [CC-BY-NC-SA 4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/deed.en) license. |
|
|
| ## Citation |
|
|
| If you use this model in your research, please cite: |
|
|
| ```bibtex |
| @inproceedings{borovik2025symupe, |
| title = {{SyMuPe: Affective and Controllable Symbolic Music Performance}}, |
| author = {Borovik, Ilya and Gavrilev, Dmitrii and Viro, Vladimir}, |
| year = {2025}, |
| booktitle = {Proceedings of the 33rd ACM International Conference on Multimedia}, |
| pages = {10699--10708}, |
| doi = {10.1145/3746027.3755871} |
| } |
| ``` |
|
|
| ```bibtex |
| @article{borovik2026pianocore, |
| title = {{PianoCoRe: Combined and Refined Piano MIDI Dataset}}, |
| author = {Borovik, Ilya}, |
| year = {2026}, |
| journal = {Transactions of the International Society for Music Information Retrieval}, |
| volume = {9}, |
| number = {1}, |
| pages = {144--163}, |
| doi = {10.5334/tismir.333} |
| } |
| ``` |