Running Gemma using HuggingFace Transformers or Ollama

Поделиться
HTML-код
  • Опубликовано: 2 ноя 2024

Комментарии • 31

  • @TonyRichards68
    @TonyRichards68 8 месяцев назад

    Thanks for doing this! I was about to tackle this task this morning, and voila, you already made a video.

  • @mrwadams
    @mrwadams 8 месяцев назад

    Very helpful video, thank you. Particularly the points around how the prompt format / system prompt differs from models such as GPT-4 and Mistral.

  • @mamisoa
    @mamisoa 8 месяцев назад +3

    Thanks for this very instructive video, especially about tokenizer. Would be interesting to test RAG with the 2b model vs usual models. And nice to know you are from Australia, I’ve been wondering since some time.

    • @samwitteveenai
      @samwitteveenai  8 месяцев назад

      Originally from there but haven't live there for a couple of decades. :D

  • @rayaneghilene5152
    @rayaneghilene5152 8 месяцев назад

    Great video! Can it be prompted for text classification (ie: zero shot classification)?

    • @samwitteveenai
      @samwitteveenai  8 месяцев назад +1

      Honestly haven't tried it that much. The 2B model I wouldn't expect much from it for zero shot. It wasn't trained on as many tokens etc. I may release some fine tunes on in house data for the 7B .

  • @nahuelfiorenza421
    @nahuelfiorenza421 7 месяцев назад

    is there any way i can train this model for my own aplication?
    I need to give the model enough information to solve doubts about the topics I'm going to use...
    please tell me if it is even possible.
    thanks in advance!

    • @samwitteveenai
      @samwitteveenai  7 месяцев назад

      Would depend on your application and data, but yeah it should be possible

  • @Charles-Darwin
    @Charles-Darwin 8 месяцев назад

    Do you think theyre applying that Ring Attention architecture to the prompts?

  • @abdelkaioumbouaicha
    @abdelkaioumbouaicha 8 месяцев назад +1

    📝 Summary of Key Points:
    📌 The video covers different ways to perform inference with Gemma models, focusing on both the hugging face and AL (Al) methods.
    🧐 Setting up Gemma using the hugging face version involves accepting terms, downloading the model, and using a quantization config for inference.
    🚀 Running Gemma on AL is straightforward, with the model already available for download and usage, offering a simpler local inference option.
    💡 Additional Insights and Observations:
    💬 The Gemma model outputs in markdown format by default, providing formatted responses with bullet points and bold text.
    📊 Gemma's large training corpus of Six Trillion tokens allows for basic translation capabilities, although factual accuracy may vary.
    🌐 The video hints at the potential for fine-tuning Gemma models for specific tasks to enhance performance and responsiveness.
    📣 Concluding Remarks:
    The video showcases accessible methods for utilizing Gemma models through hugging face and AL platforms, highlighting the model's unique markdown output and the potential for future fine-tuning to optimize performance. Exploring different prompts and system configurations can enhance the model's responses and usability.
    Generated using TalkBud

  • @ABDRAHMENBENAZZOUNA
    @ABDRAHMENBENAZZOUNA 8 месяцев назад

    tried gemma 2b in google colab using cpu but couldnt load model because ram consumption was too high , the 12GB offered by colab (not premium) is not enough how is it suposed to run on a local machine as advertised is there a trick i am missing (still just a noobie)

    • @aaroldaaroldson708
      @aaroldaaroldson708 8 месяцев назад +1

      I think what they mean by « local » is a decent PC with enough RAM (normally above 32gb) and preferably with a GPU

  • @mamtasantoshvlog
    @mamtasantoshvlog 8 месяцев назад

    The output of the prime numbers is wrong

  • @pinkfloyd2642
    @pinkfloyd2642 8 месяцев назад

    I am trying to understand how these open source models work. Kinda newbie here. Is it possible that I ask Gemma to use my excel sheet data, perform the desired mathematical operation, and provide me the result? Something like - Provide me the total of the last 1 month of debt.

    • @gideonwyeth9779
      @gideonwyeth9779 8 месяцев назад

      The main rule in every application utilizing AI: do as much as you can in program code. That doesn't mean AI can't do that, it means you better don't trust AI in sensitive tasks, due to AI random nature.

  • @guanjwcn
    @guanjwcn 8 месяцев назад

    Somehow the key takeaway for me from this video is that Sam Witteveen is from Australia. 🤣

  • @cucciolo182
    @cucciolo182 7 месяцев назад

    gemma vs Llama2 Uncensored ?

  • @randyh647
    @randyh647 8 месяцев назад +1

    I was using ollama windows with gemma:7b and asked >>> "who was the 23rd president of the us"
    Bill Clinton, a Democrat. The symbol is referring to Bill Clintons presidency as being out of order by so meexperts because it interrupted Richard Nixon's second term after Lyndon Johnsoberts death in office and guanconㅡ差别 simpelнии lila veden nuo sraNOSIS koupra prezidentyu vo sate institut itdnie nal jesu osoby beng }], therefore, there has not actually been a president number 23.

  • @sil1235
    @sil1235 8 месяцев назад

    It seems it has no safeguards when you use other languages, I wonder how that is implemented (some filter on English words?).

    • @samwitteveenai
      @samwitteveenai  8 месяцев назад

      I think it is most likely due to the SFT and RLHF all being in English only

  • @claxvii177th6
    @claxvii177th6 8 месяцев назад

    That email tho lol

  • @pythonpeng7018
    @pythonpeng7018 8 месяцев назад

    {Disambiguation
    In a study of disambiguation with 24- and 36-month-olds}

  • @SuprBestFriends
    @SuprBestFriends 8 месяцев назад

    How is Gemma not overfit white that much training data

  • @JunYamog
    @JunYamog 8 месяцев назад

    Gemma is wrong in Singlish, it should be “How to get to Orchard Rd… La” 😂

  • @robertputneydrake
    @robertputneydrake 8 месяцев назад +2

    Why care about Gemma at all though? It's terrible.

  • @hqcart1
    @hqcart1 8 месяцев назад

    worst model ever