Overview of an Example LLM Inference Setup

Поделиться
HTML-код
  • Опубликовано: 27 окт 2024

Комментарии • 12

  • @FarhadOmid
    @FarhadOmid 2 месяца назад

    Great work, Jordan! Gonna start scraping the parts together...

  • @rodrimora
    @rodrimora 2 месяца назад +2

    I feel jalous of that 8xH100 server. Currently using a 4x3090 at home. I actually use a pretty similar setup with vLLM for the full precision models and exllama or llama.cpp for quantized models + OpenwebUI as a frontend.

    • @MadeInJack
      @MadeInJack 2 месяца назад +1

      Why would you need more than that? Be glad for what you already have or you won't find happiness :)

    • @ricardocosta9336
      @ricardocosta9336 2 месяца назад

      Bitch i have a p40 and im over the moon. Being poor in ml is hard.

  • @fakebizPrez
    @fakebizPrez 2 месяца назад

    Sweet rig. Is that your daily driver? 😀😀

  • @niceshotapps1233
    @niceshotapps1233 2 месяца назад +2

    - what are you using it for?
    - .... stuff

    • @0101-s7v
      @0101-s7v 2 месяца назад

      AI, apparently. (LLM = Large Language Model)

  • @KCM25NJL
    @KCM25NJL 2 месяца назад +2

    The cost of such a setup is circa $500,000........ amma get me 2 :)

  • @ZiggyDaZigster
    @ZiggyDaZigster 2 месяца назад

    30k GC? 8 of them?

  • @nesdi6653
    @nesdi6653 2 месяца назад

    Word

    • @nesdi6653
      @nesdi6653 2 месяца назад

      Why not podman tho