Lets say we got the GPT-3 model from OpenAI. (I know GPT-3 is closed source)
Then we can do fine-tune the GPT-3 model.
In that case what would be the difference between fine-tune vs few-shot learning.
The case for few-shot learning is it does not need the model to be trained and just need to show a few examples.
But pre-training is easy as well. We just need to train on a smaller dataset. Also what I understand is for pre-training you dont need a complex GPU setup either.
So if someone had access to GPU, why would he ever used few-shot learning?
If they get the GPT-3 model weights can anyone would be able to fine-tune it if they have access to a couple of RTX 3080 GPU?
Or would it need setup like the big companies?