I’m using the
PeftModel.from_pretrained(base_model, lora_model_id) method to load a LoRA adapter on a base LLM. The method seems to be directly modifying the
Is there a way to “unload” an adapter to get the original
base_model weights back? I want to be able to switch between adapters in real-time for multi-task inference.