i have trained different BERT models to evaluate which of these models perform best with my dataset.
At the moment I got the problem that I always get different results when running the model again (e.g. run a new colab session).
For a reproducible evaluation I want to get always the same results (e.g. F1) but I don’t know how to do this.
I have ensured that my train/test/validation split is always the same. I have read that it could have something to do with the Dropout of the model, but what do I have to change if I want the output to be every time the same?
Thanks in advance!