i did download qwen2.5-coder-32b-instruct-q4_0.gguf on my lm studio how to make my ollama run it ? i dont know if i have to move my model to somwhere else or i have to tell ollama run it the model is in my C:\Users\user\.cache\lm-studio\models\Qwen\Qwen2.5-Coder-32B-Instruct-GGUF folder
Thank you very much for this review! Good to know quantization has not affected much the quality of the model. I will start using it for my coding endevours :)
🔥Qwen2.5 Coder 32B Instruct - Best Coding Model To-Date - Install Locally - ruclips.net/video/tMd0FcPSei4/видео.htmlsi=BET-0lYt68gUO25I
i did download qwen2.5-coder-32b-instruct-q4_0.gguf on my lm studio how to make my ollama run it ? i dont know if i have to move my model to somwhere else or i have to tell ollama run it the model is in my C:\Users\user\.cache\lm-studio\models\Qwen\Qwen2.5-Coder-32B-Instruct-GGUF folder
Thank you very much for this review! Good to know quantization has not affected much the quality of the model. I will start using it for my coding endevours :)
Hehe..it does run on my 3060 but with 3-4 t/s.
Wow! that is very impressive. I wonder how much better the unquantized version is.
Nice info!!! What is the context window you are using? Also, do you think the 15b version is as good?
please cover this topic - LLaMA-Mesh by Nvidia
Thank You