This thread should be used to ask questions about how examples/seq2seq/distillation.py
works, and to ask questions about the associated paper after it gets released.
What is the reasoning behind choosing alternating layers ?
no teacher distillation scores for XSUM ?
no teacher is working for non seq-2-seq task as well as we saw with MNLI, should we also see if it works other tasks as well ?
Alternating layers seems to perform the best by a moderate amount.
Definitely interested to see results for other tasks!
relocated to examples/research_projects/seq2seq-distillation/distillation.py
?
Yes, that project is now moved to research_projects
dir.
Hey @sshleifer, I was trying to fine-tune the distill-pegasus-cnn-16-4
model provided by you but I am not sure of the hyper-parameters. Could you please share the hyper-parameters that you used to train this model (and achieve the results shown in Table 5 from your paper?
Thanks a lot!
Naman
Hi! have a question regarding the article «Pre-trained Summarization Distillation» (https://arxiv.org/pdf/2010.13002.pdf). In section 6.2, it is said «Table 10 shows results from fine-tuning teacher models…». However, throughout the paper it is stated that the experiments with pseudo-labeling only when fine-tuning the student model were performed. Is it a typo and the result from fine-tuning student models is indeed depicted?
Thanks in advance!
Hi @sshleifer . Any thoughts on if the T5 distillation would still be feasible with PEFT techniques such as LORA? I have a fine tuned T5-11B using LORA and want to distill this model to something feasible like T5-base or even T5-large. But I’m not sure if the teacher model , which essentially has a LoRA adapter work on a similar way ? Any thoughts / ideas regarding this would be great help. Thanks