Recently, I found two new representations for symbolic music generation.
But using their open source code is not as easy as using muspy.
I hope that these two representations could be added to muspy's built-in representations.
REMI
https://ailabs.tw/human-interaction/pop-music-transformer/

The MIDI-like representation is actually the event-based representation in muspy
And of course, the REMI representation got beter performance in their paper
Compound Words
https://ailabs.tw/human-interaction/compound-word-transformer-generate-pop-piano-music-of-full-song-length/

The Compound Words representation further improves the REMI representation.
The most import point I think is that, Compound Words representation makes the sequence much shorter.
And this makes it posible for attention window to cover the whole sequence.

Recently, I found two new representations for symbolic music generation.
But using their open source code is not as easy as using muspy.
I hope that these two representations could be added to muspy's built-in representations.
REMI
https://ailabs.tw/human-interaction/pop-music-transformer/

The MIDI-like representation is actually the event-based representation in muspy
And of course, the REMI representation got beter performance in their paper
Compound Words
https://ailabs.tw/human-interaction/compound-word-transformer-generate-pop-piano-music-of-full-song-length/


The Compound Words representation further improves the REMI representation.
The most import point I think is that, Compound Words representation makes the sequence much shorter.
And this makes it posible for attention window to cover the whole sequence.