I’m going through the tutorial steps here to make sure I can run transformers on my gpu, but I’m running in to an OutOfMemoryError when trying to run the “Vanilla Training” step.
I’ve searched around and mostly people are suggesting lowering the batch sizes. So I tried lowering the batch size to 1 but still get the error. The data fits in the gpu so this is something training related, right? Is the tutorial expecting me to have a GPU with 16GB of memory? Or is something else going on with my setup?