I have use PEFT lora technique to fine tune a mistral model i used bit and byte for my quantization so wanted to where its saving ,mu quantize model and how i can use quantize model with my adapter layer. if any reference please share article
Related topics
Topic | Replies | Views | Activity | |
---|---|---|---|---|
Finetuned LLM model conversion to GGUF - performance drop | 4 | 1518 | July 31, 2024 | |
Loading an LoRA adapter trained on quantized model on a non-quantized model | 0 | 1306 | November 7, 2023 | |
Can I load a model fine-tuned with LoRA 4-bit quantization as an 8-bit model? | 0 | 285 | November 27, 2023 | |
Fine tuning LoRa merge | 0 | 320 | January 10, 2024 | |
Can the Pegasus Model be Adapted Using the PEFT-LORA Technique? | 0 | 233 | August 8, 2023 |