Hum, there should be tqdm
bars all around the training. Perhaps it didn’t even start in case you see nothing.
By the way, a couple of months after this, I suggest going for a different task to learn how TPUs work. NLI is by no means a good example of this. Pre-training or fine-tuning a language model is a good use case for this hardware. Check run_mlm.py example, it can be easily adapted using the same ideas of my former notebook.