Fine-Tuning Mistral 7B

Поделиться
HTML-код
  • Опубликовано: 8 ноя 2023
  • This session is led by Chris and Greg!
    You'll learn what you need to know about Mistral 7B, and how to get it fine-tuned for your application!
    Agenda with additional resources: docs.google.com/document/d/1s...
  • НаукаНаука

Комментарии • 14

  • @vibhugoel5525
    @vibhugoel5525 4 месяца назад

    This made my day. Perfect and clear explanation.

  • @hungle2514
    @hungle2514 6 месяцев назад +1

    THank you so much for this tutorial.

  • @AI-Makerspace
    @AI-Makerspace  6 месяцев назад +5

    Google Colab: colab.research.google.com/drive/1JtrVh--bcPR-CR8QNOyXd3Z5eZt0WgOw?usp=sharing
    Slides: www.canva.com/design/DAFzn7Uynrc/IMrrg6GSL_2NWpAnWXfobQ/edit?DAFzn7Uynrc&

  • @seinaimut
    @seinaimut 6 месяцев назад +1

    thanks for this tutorial bro ...

  • @consig1iere294
    @consig1iere294 6 месяцев назад

    New to this stuff. Is it possible for me to use my own gpu to train? If yes, how? Thanks!

    • @AI-Makerspace
      @AI-Makerspace  6 месяцев назад +3

      With a combination of Quantization strategies (4bit from bitsandbytes, AWQ, and more) plus LoRA (or other adapter methods) it's more than possible to fine-tune large language models on a consumer GPU!
      If it's your own GPU on prem, you'll just have to deal with some hardware config that is more streamlined when leveraging compute from cloud providers!

  • @manmanzhang7034
    @manmanzhang7034 4 месяца назад

    Thanks, this is super , in your generate_reponse(promt), for generate_ids what is the value for pad_token_id? pad_token_id=tokenizer or pad_token_id=tokenizer.eos_token? I actually tried both of them, none of them works, anything I missed here? is there any other parameter after pad_token_id?

    • @AI-Makerspace
      @AI-Makerspace  4 месяца назад +1

      pad_token_id=tokenizer.eos_token_id is what you'd want!

  • @thisurawz
    @thisurawz 4 месяца назад +3

    Can you do a video on finetuning a multimodal LLM (Video-LlaMA, LLaVA, or CLIP) with a custom multimodal dataset containing images and texts for relation extraction or a specific task? Can you do it using open-source multimodal LLM and multimodal datasets like video-llama or else so anyone can further their experiments with the help of your tutorial. Can you also talk about how we can boost the performance of the fine-tuned modal using prompt tuning in the same video?

    • @AI-Makerspace
      @AI-Makerspace  4 месяца назад +2

      We'll add this suggestion to our backlog of potential future events for sure! Keep the ideas coming!

    • @thisurawz
      @thisurawz 4 месяца назад

      @@AI-Makerspace Thanks

  • @horyekhunley
    @horyekhunley 3 месяца назад

    If i have hardware constraints, can i use a small model such as tiny-llama?
    Also, how can i perform RAG on a csv dataset?

    • @AI-Makerspace
      @AI-Makerspace  3 месяца назад

      You could!
      For the RAG question - you could use a CSVRetriever!