Qwen2.5 Coder 32B on Ollama - Run Locally on Less VRAM

Поделиться
HTML-код
  • Опубликовано: 17 ноя 2024

Комментарии • 8

  • @fahdmirza
    @fahdmirza  День назад +1

    🔥Qwen2.5 Coder 32B Instruct - Best Coding Model To-Date - Install Locally - ruclips.net/video/tMd0FcPSei4/видео.htmlsi=BET-0lYt68gUO25I

    • @mostafamostafa-fi7kr
      @mostafamostafa-fi7kr День назад

      i did download qwen2.5-coder-32b-instruct-q4_0.gguf on my lm studio how to make my ollama run it ? i dont know if i have to move my model to somwhere else or i have to tell ollama run it the model is in my C:\Users\user\.cache\lm-studio\models\Qwen\Qwen2.5-Coder-32B-Instruct-GGUF folder

  • @andrepaes3908
    @andrepaes3908 День назад

    Thank you very much for this review! Good to know quantization has not affected much the quality of the model. I will start using it for my coding endevours :)

  • @bamit1979
    @bamit1979 День назад +2

    Hehe..it does run on my 3060 but with 3-4 t/s.

  • @metaltech3944
    @metaltech3944 День назад

    Wow! that is very impressive. I wonder how much better the unquantized version is.

  • @DigitalDesignET
    @DigitalDesignET День назад

    Nice info!!! What is the context window you are using? Also, do you think the 15b version is as good?

  • @Mnbir
    @Mnbir День назад +1

    please cover this topic - LLaMA-Mesh by Nvidia

  • @ronigaldino355
    @ronigaldino355 День назад

    Thank You