I everyone I am trying to fine tuning llama2 using lora on colab, I need to resume training from the last checkpoint available but I don’t know how to do it.
Did you ever found a solution?
I everyone I am trying to fine tuning llama2 using lora on colab, I need to resume training from the last checkpoint available but I don’t know how to do it.
Did you ever found a solution?