Please read the topic category description to understand what this is all about
Description
If you treat music notation as a form of “text”, you can use language modelling to generate new songs! The goal of this project is to explore how well Transformers perform at music modelling.
Model(s)
None that we could find on the Hub, but see here for some pretrained music Transformers
Datasets
None that we could find on the Hugging Face Hub, but see here for some ideas
Challenges
This task probably involves pretraining a Transformer, which can potentially take multiple days using the GPU resources provided by AWS. An interesting alternative would be to see whether one can integrate an existing pretrained model within Transformers and use that to either fine-tune.
Desired project outcomes
- Create a Streamlit or Gradio app on Spaces that people can remix famous songs with newly generated ones
- Don’t forget to push all your models and datasets to the Hub so others can build on them!
Additional resources
- https://towardsdatascience.com/creating-a-pop-music-generator-with-the-transformer-5867511b382a
- https://arxiv.org/abs/2002.00212
- https://openai.com/blog/jukebox/
- https://research.google/teams/brain/magenta/
Discord channel
To chat and organise with other people interested in this project, head over to our Discord and:
-
Follow the instructions on the
#join-course
channel -
Join the
#pop-music-transformer
channel
Just make sure you comment here to indicate that you’ll be contributing to this project