Hey, thank you so much for the reply, and sorry for the late response! I’m currently checking out Gemma 3n 4B but I’m running into an issue—it requires a lot of memory to run locally after downloading it from Hugging Face. Do I need to quantize the model before running it?
1 Like