When working with LoRA, it is common to use .add_adapter method to supply a particular layers of various NLP models such as DeBERTa, FLAN, T5, etc., with adapter inside those selected layers.
This method differs from get_peft_model, because the first one do not create a wrapper outside basic the model which, sometimes, comes in handy.
The problem is that this .add_adapter method is not supported for CLIP, ViT and other popular vision models.