Mistral 7B LLM AI Leaderboard: GPU Contender Nvidia Tesla M40 24GB

Поделиться
HTML-код
  • Опубликовано: 21 окт 2024

Комментарии • 10

  • @InstaKane
    @InstaKane Месяц назад

    Great video, looking forward to seeing that leaderboard get filled up!

    • @RoboTFAI
      @RoboTFAI  Месяц назад

      Thanks! Lots more coming to fill it out!

  • @pro100vald
    @pro100vald Месяц назад

    Thank you for providing comprehensive testing! Can you please share the name of the software that you use to do the tests?

    • @RoboTFAI
      @RoboTFAI  Месяц назад

      The testing suite is a custom built app I built for my lab that uses streamlit, langchain, python, etc.

  • @blast_0230
    @blast_0230 Месяц назад

    m40 and p40 card is good at fp32. They are not built for fp 16 or int8. So if you have two p40 or m40 you should try an fp32 llm and you will surprise by the result.
    also great video.

    • @RoboTFAI
      @RoboTFAI  Месяц назад

      I agree! They are still good budget cards for expanding VRAM, that's where I started in my lab. Have 3 of them sitting here. They are aging out of cuda a bit, and power hungry, along with passive cooling so gotta handle that also.

  • @FrankHouston-v5e
    @FrankHouston-v5e Месяц назад +1

    M40 AT $85 is great when paired with my 4080 super giving me a 40 Gb GPU 🧐.

    • @InstaKane
      @InstaKane Месяц назад

      Nice, I’m thinking of doing the same, does the M40 GPU slow down the overall inference speed?

    • @RoboTFAI
      @RoboTFAI  Месяц назад

      Yep I agree, for what they cost if you can feed them power and keep them cool, they are still good value depending on your use cases! Tensor splitting can be fun to play with when using mixed cards.
      Mixed cards will effect your speeds (have a video covering tensor splitting and mixed gpus) but it's great way to expand your total VRAM for larger models!

    • @FrankHouston-v5e
      @FrankHouston-v5e Месяц назад

      @@InstaKane M40 will be 1/5 the speed of 4080 super but I’m using 8 gig to build a 24 gb card. At level 3 my Dell T7820 with 128gb Llamafile 40 core Xeon CPU. Building this machine to launch my ML/AI channel 🥹.