I Tested NVIDIA Nemotron 70B and Found the BEST Open Source LLM

Поделиться
HTML-код
  • Опубликовано: 9 янв 2025

Комментарии • 25

  • @Alen_115
    @Alen_115 2 месяца назад

    I was eagerly waiting for your this video. I also think the model is awesome. I was waiting for your review

  • @HyperUpscale
    @HyperUpscale 2 месяца назад +1

    Thank you for the video!

  • @zoranProCode
    @zoranProCode 2 месяца назад +4

    how fast is it? can we use it on some fast platform like groq?

  • @amit4infinity
    @amit4infinity 22 дня назад

    Much appreciated

  • @LeaSam-u9o
    @LeaSam-u9o 2 месяца назад

    Wow impressed with how fast it was able to debug the codes!

  • @rambapat588
    @rambapat588 11 дней назад

    Can you make a video about how to host a nim on google cloud?

  • @greatsarmad5147
    @greatsarmad5147 2 месяца назад +1

    can you test it like apple did, make some small changes in names or numbers in test questions and see if response quality gets worse or not?

  • @ZaferCan
    @ZaferCan Месяц назад

    i've tried many of open source models, but still can't figure it out how to make a chatbot with tool calls with chat history. After answering according to its general knowledge and not using a tool even though it should, llm model cannot even use very simple tools. Because I think it looks at conversation history and says "oh, okay, I can handle this without using a tool"

  • @elyakimlev
    @elyakimlev 2 месяца назад +1

    I loaded a 4.3b exl2 quantized version of it on my 2x RTX 3090. I'm getting 15-18 tokens a second. This thing is super fast and smart! The only downside is that it's censored. Hopefully someone makes an uncensored version soon.

    • @rportella9357
      @rportella9357 2 месяца назад

      "Downside."

    • @KaletheQuick
      @KaletheQuick 2 месяца назад

      ​@@rportella9357Yes, that is a downside.

  • @RodCoelho
    @RodCoelho 2 месяца назад

    Where can I test this modrl like you did?

  • @limjuroy7078
    @limjuroy7078 2 месяца назад

    I heard it will use more tokens for a query die to the built-in CoT?

  • @nrusimha11
    @nrusimha11 2 месяца назад +1

    Thank you for the video. What hardware did you run this on?

    • @MervinPraison
      @MervinPraison  2 месяца назад +1

      I used hugging chat.
      Ideally you need a high spec computer with good GPU

  • @aniketchavan6868
    @aniketchavan6868 2 месяца назад

    can we use it with groq

  •  2 месяца назад

    Nice!

  • @tonamewith
    @tonamewith 2 месяца назад +1

    Do you get the same results if you redo the whole test?

    • @rportella9357
      @rportella9357 2 месяца назад +1

      i bet not, cuz i've see some people get the strawberry question right.

  • @theNotLogo
    @theNotLogo 2 месяца назад

    I have an issue. Please help me people..i waited foe my new cpu with i914900kf 128 gb Kingston 6000 mhz ran nvme 980 pro and rtx 4090 BUT 70B models can fit to my vram so how to use all my power? Or is there any model for coding tasks which will give good coding results and fit all my hardware?

  • @60pluscrazy
    @60pluscrazy 2 месяца назад

    🎉

  • @igormel5137
    @igormel5137 2 месяца назад +2

    Я протестировал его для решения своих задач на Python, а также попросил написать несколько простых игр вроде арканоид. Очень слабая модель. Мне кажется даже Qwen 2.5 6B получше справляется.

    • @shouyiwangwasai1
      @shouyiwangwasai1 2 месяца назад +1

      I am a Chinese user, and I also feel that this model is weak. Why is everyone praising it?