Here is my issue. I have a model trained with
huggingface (yay) that works well. Now I need to perform inference and compute predictions on a very large dataset of several millions of observations.
What is the best way to proceed here? Should I write the prediction loop myself? Which routines in
datasets should be useful here? My computer has lots of RAM (100GB+), 20+ cpus and a big GPU card.