Hello,
I’m using the PeftModel.from_pretrained(base_model, lora_model_id)
method to load a LoRA adapter on a base LLM. The method seems to be directly modifying the base_model
weights.
Is there a way to “unload” an adapter to get the original base_model
weights back? I want to be able to switch between adapters in real-time for multi-task inference.
Thanks!