Example of how to pretrain T5?

T5 pre-training is now supported in JAX/FLAX. You can check out the example script here: transformers/examples/flax/language-modeling at master 路 huggingface/transformers 路 GitHub. It actually includes 2 scripts:

@patrickvonplaten also demonstrates how to run the script in this video (starts around 13:35).

This script was developed for the JAX/FLAX community event. It would be really cool if someone contributes the PyTorch version of it. It would mean translating the script from FLAX to PyTorch, which is probably straightforward.

2 Likes