Let's go live and look at ExaOne3.5

Поделиться
HTML-код
  • Опубликовано: 18 дек 2024

Комментарии • 10

  • @ianfoster99
    @ianfoster99 2 дня назад

    Great video. When you started to do the English Korean translation it reminded me of a recent video about LCMs which Meta has been researching. May be worth doing a session on them

  • @alexmendoza9394
    @alexmendoza9394 2 дня назад

    Can you show us how to implement LLMs in multiple machine as a cluster of 4 maybe?

  • @republicofamerica1229
    @republicofamerica1229 2 дня назад +1

    Hello Matt, results in open webui with Ollama using two or more models at the same time are very unstable. I was always using 3 models or 4 at the same time and got very frustrated. I moved back to OpenAI. My Ollama server was a test unit again and I noticed that one model at the time was providing amazing answers.

    • @technovangelist
      @technovangelist  2 дня назад +1

      Ollama does multiple models very well at the same time, especially when you have multiple gpus. The machine I am using has 8 H100 GPUs, so it performs very well.

  • @chrisBruner
    @chrisBruner 2 дня назад

    Some obvious answers to the questions. Time Travel, obviously the one guy got spread over two timelines, therefore arriving before and after the other person. (Probably not intact though). Lightbulb answer. Many of the lightbulbs being replaced were previously burnt out.

  • @DasJev
    @DasJev 10 часов назад

    I think the benchmark has to align with your own task, if no benchmark aligns then you have to create your own one, testing what you want to accomplish. Speed is easy, but maybe you have some spezialized task like doing taxes in some complicated country, no benchmark would cover that. Maybe you are looking for an engaging conversation where the scientific accuracy of answers are secondary to you just having fun - like a starship commad AI story game.
    The question is though for me, how do I automate it? Do I use a llm backend, then some pything and excel? or is there already a harness for that?

  • @carinebruyndoncx5331
    @carinebruyndoncx5331 2 дня назад +1

    "Lies, damn lies and statistics" 😊

  • @JustinJohnson13
    @JustinJohnson13 2 дня назад

    Nothing ever works in a live demo. Murphy's law at it's finest.

    • @technovangelist
      @technovangelist  2 дня назад

      except i would say that it all worked very well

  • @DreamsAPI
    @DreamsAPI 2 дня назад

    How is that when possible? That amount of RAM is insane to the nut 🧠 😀