Hi, I had a beginner question: does HF Transformers have a way to distill a LLM w/ knowledge distillation? For example, could I make something like distilgpt2
?
Thanks!
Hi, I had a beginner question: does HF Transformers have a way to distill a LLM w/ knowledge distillation? For example, could I make something like distilgpt2
?
Thanks!