Please read the topic category description to understand what this is all about
Emoticons are often used as a proxy for emotional content in social media posts or instant messaging chats. As a result, emojis are often used as a label to train text classifiers. The goal of this project is to create a Transformer-based implementation of DeepEmoji, a research project from MIT that studied this task with LSTMs.
Any BERT-like model would be a good candidate for fine-tuning on an emoji dataset and you can get inspiration from models like these:
To get better performance, you may want to perform domain adaptation by fine-tuning the language model on in-domain data. We recommend trying this approach only after building a baseline classifier.
- Create a Streamlit or Gradio app on Spaces that can predict the top 5 emojis associated with a piece of text
- Don’t forget to push all your models and datasets to the Hub so others can build on them!
- https://huggingface.co/spaces/ml6team/emoji_predictor (a Space for inspiration)
To chat and organise with other people interested in this project, head over to our Discord and:
- Follow the instructions on the
- Join the
Just make sure you comment here to indicate that you’ll be contributing to this project