Hi everyone,
I’m working on building a rubber duck–style code assistant that takes user code and returns explanations, debugging suggestions, and thoughtful guiding questions. I want to fine-tune a model to handle this well.
My key goals are:
- The model should perform well for code analysis, especially for C++ and Python.
- I want to fine-tune using LoRA or any lightweight method, ideally on Google Colab (free tier) or a modest local setup (Ryzen 5, 8GB RAM).
- I eventually want to deploy the model for backend use in my app — so I need something that can be deployed at zero cost or with open-source tools (like
ollama
,text-generation-webui
, etc.).
Currently considering models like DeepSeek-Coder, CodeLlama, or StarCoder.
Can anyone suggest:
- Which open-source model would be best for my needs?
- Are there quantized versions that run well on basic systems?
- Any fine-tuning tools or tricks for keeping it lightweight?
Thanks in advance!