LLAMA-3.1 🦙: EASIET WAY To FINE-TUNE ON YOUR DATA 🙌

Поделиться
HTML-код
  • Опубликовано: 25 окт 2024

Комментарии • 56

  • @VerdonTrigance
    @VerdonTrigance 2 месяца назад +18

    The most complex problm is preparing dataset with QnA from which u gonna learn. And this is what I'd like to see.

    • @engineerprompt
      @engineerprompt  2 месяца назад +5

      here is a previous video I did on creating custom datasets: ruclips.net/video/z2QE12p3kMM/видео.html

  • @unclecode
    @unclecode 2 месяца назад +8

    Fascinating! The two Australian brothers did a fantastic job of introducing the Unsloth to the community.

  • @paul1979uk2000
    @paul1979uk2000 2 месяца назад +6

    The more I'm seeing of A.I. advancement, I'm coming to the concluding that better isn't always better, and the real battleground isn't so much which is the best when many of the better A.I. models are so close to each other in quality, the real battleground for me is the quality for the size, so a bit like we do for hardware, performance per watt, but in this case, performance per billion parameters, if you can maintain or have better quality at a smaller size, that is a major advantage, especially if it's open source and can run locally on your hardware.
    So as good as the big A.I. models are, they are too tightly controlled and very limited in how you can run them, in most cases online because of how big they are, the real game changer I think is with the smaller open source models that you can run locally, the advantage they've got is that they can be fully integrated and specialised in the OS, apps and games, they also have the advantage of less privacy, security and other concerns like that.
    If the current advancements of A.I. models continues and hardware continues to progress, I suspect the online big models are not going to matter that much as the smaller ones we can run locally will be able to do most of the things we want, and that's when things get really interesting as A.I. gets far more integrated into our daily lives, something that's really limited with these online centralised A.I. models and for countless reasons.
    At the end of the day, what's going to win out isn't going to be the best, good enough will do for most of us, what will really win out is what is smaller, capable and can be run locally, which basically rules out the big online A.I. services as there are too many privacy and security concerns with them, especially as A.I. becomes more capable and integrated into our lives.

  • @MYPL89
    @MYPL89 2 месяца назад +1

    You have no idea how this video helped me!! THANK YOU SO MUCH

  • @truptimohanty9386
    @truptimohanty9386 14 часов назад

    Thank you so much for this wonderful video! I have a couple of questions: For max_seq_length = 2048 # Choose any! We auto-support RoPE scaling internally!, could you clarify, whether it handles cases where the input sequence length exceeds 2048 tokens? Also, when determining the max sequence length for custom data, should it include the combined length of the instruction, input, and output? Thank you again for your insights!

  • @DevtalTalks
    @DevtalTalks Месяц назад

    Excellent explanation!!

  • @Kiran.KillStreak
    @Kiran.KillStreak 2 месяца назад +2

    Thanks for video, every minute detailed video .superb.

  • @avataraang3334
    @avataraang3334 Месяц назад

    Thank You Brother, Truly

  • @TeamDman
    @TeamDman 2 месяца назад

    Haven't watched yet but thank you for all your guides on this, I know where to come when I need to do this myself !!

  • @sizilienrockt9457
    @sizilienrockt9457 2 месяца назад

    like this and this show the easy way for ppls who not are student for ai . not newbie frendly to complex tutorial

  • @developerashish6849
    @developerashish6849 2 месяца назад

    This is awesome, and tutorial is so easy to understand too

  • @bakegleeson8653
    @bakegleeson8653 Месяц назад

    Thank you 😀

  • @johnclay7422
    @johnclay7422 2 месяца назад +1

    hello sir !!!! wonderful contribution!!! can you practically train the model on the data so that we can learn . I am new to this field and your channel is amazing. thanks

    • @engineerprompt
      @engineerprompt  2 месяца назад

      here is a previous video I did on creating custom datasets: ruclips.net/video/z2QE12p3kMM/видео.html

  • @robertjalanda
    @robertjalanda 2 месяца назад

    Thank you for this video tutorial very helpful!

  • @georgebongo-o6n
    @georgebongo-o6n 17 дней назад

    I have problem in creating my own datasets manually. Like CSV file format, how can structure it in CSV file and read it to the fine tunning process?

  • @nikitagarashchuk2430
    @nikitagarashchuk2430 2 месяца назад

    Thanks for video!!
    Can you inform me how to deploy a fine tuned model?

    • @engineerprompt
      @engineerprompt  2 месяца назад +1

      check out this playlist on deployment:
      ruclips.net/video/OuQBxBrO2ms/видео.html&ab_channel=PromptEngineering

  • @akki_the_tecki
    @akki_the_tecki 2 месяца назад +2

    why nobody speaks about, "How should we convert my CONFIDENTIAL RAW text/ PDF into Datasets"????

    • @pladselsker8340
      @pladselsker8340 2 месяца назад

      This is so funny, been looking for this yesterday and today now. Maybe I'm just now realizing after 20 years of google searching experience that I'm bad at googling.

    • @engineerprompt
      @engineerprompt  2 месяца назад

      here is a previous video I did on creating custom datasets: ruclips.net/video/z2QE12p3kMM/видео.html

  • @GandalfTheBrown117
    @GandalfTheBrown117 2 месяца назад

    Thank you!!

  • @karthikb.s.k.4486
    @karthikb.s.k.4486 2 месяца назад

    Nice . Can we run this in our local machine and what config needed to run in local mackbook. Or colab is preferred please let me know.Also can you suggest is mackbook good for handling LLMS

    • @unclecode
      @unclecode 2 месяца назад +1

      Training should be conducted on a CUDA device, but the resulting model can be used on MPS devices (MacBook M series) and CPUs. For fine-tuning models on Mac using MLX-a powerful, open-source array framework for Apple silicon-there's a vibrant community supporting it.

  • @ThaLiquidEdit
    @ThaLiquidEdit 2 месяца назад

    Could you make a video on how to create a training set to fine-tune a model? I want to fine-tune a model like LLAMA-3.1 that creates YAML sections for different tasks similar to ansible. For example when I prompt: "Create a user alice" it should generate a YAML in a specific format like
    user:
    action: create
    username: alice
    Can you show how we can create such a training set. I can't create thousands of training data manually.

    • @AlpMercan14
      @AlpMercan14 2 месяца назад

      You can use function calling also it may be enough rather than fully finetuneing

    • @engineerprompt
      @engineerprompt  2 месяца назад

      You can achieve this through prompting. fine-tuning should be a last resort. You dont' need it in most cases.

    • @ThaLiquidEdit
      @ThaLiquidEdit 2 месяца назад

      @@engineerprompt Could you give an example? You mean like explain the format of the YAML file, make an example and e.g. write "Whenever I create an user, output this YAML file"?

  • @sujjee
    @sujjee 2 месяца назад

    hey can you please explain how to fine to model and deploy to own server if privacy is concern. please make a tutorial on it

    • @engineerprompt
      @engineerprompt  2 месяца назад

      Here is a playlist on deployments:
      ruclips.net/video/OuQBxBrO2ms/видео.html&ab_channel=PromptEngineering

  • @alexo7431
    @alexo7431 2 месяца назад

    Thanks for sharing

  • @preetamrawat4939
    @preetamrawat4939 2 месяца назад

    Hello bro how r u. I just started here but confused where to begin. Can you guide me in a specific direction.
    Thank you :)

    • @engineerprompt
      @engineerprompt  2 месяца назад

      what exactly is your confusion. Are you interested in getting started with LLMs or fine-tuning them?

    • @preetamrawat4939
      @preetamrawat4939 2 месяца назад

      @@engineerprompt which should I start 1st and from which video or playlist should I start?

  • @deepadharshinipalrajan8849
    @deepadharshinipalrajan8849 2 месяца назад

    Are we able to fine tune the model directly which is available in the ollama server?

  • @fl028
    @fl028 2 месяца назад

    Is it normal that the fined tuned version response with the ### Instruction, ### Input, ### Response pattern. Do I have a alternative in the training section, when i want only the response?

  • @awu878
    @awu878 2 месяца назад

    The colab link doesn't seems to work

  • @RajaReivan
    @RajaReivan Месяц назад

    can i have a validation set in sfttrainer?

  • @caslor2002
    @caslor2002 2 месяца назад

    nice video but as most of the other in the same topic use an all ready dataset... i would prefer to see a video juat for a basic construction of a custom alpaca dataset... I think is what is missing from the most of the same kind tutorials.. the logic and the method to create your own alpaca dataset, what if a question has more than one answer? what if a simple question need to be clarified by the user depending of two probabilities ? and then follows the answer based on the clarification user inputs etc ....

    • @engineerprompt
      @engineerprompt  2 месяца назад +1

      here is a previous video I did on the topic: ruclips.net/video/z2QE12p3kMM/видео.html

    • @caslor2002
      @caslor2002 2 месяца назад

      @@engineerprompt Thanks for your reply.. just checked the Link, awesome video... Thanks for sharing your knowledge!!

  • @frazuppi4897
    @frazuppi4897 2 месяца назад +1

    the issues is that benchmark are broken, seeing the graph is pointless at this point

    • @engineerprompt
      @engineerprompt  2 месяца назад

      I agree but unfortunately that's the only thing we have at the moment.

    • @frazuppi4897
      @frazuppi4897 2 месяца назад

      @@engineerprompt well it is not true, you can craft your own benchmark, some people are doing it, and share with us what you thing from it

  • @AlpMercan14
    @AlpMercan14 2 месяца назад

    Can you show how can ı finetune it and store the new one locally. Like ı already have llama3.1 on my local. I want to finetune it and use it

    • @engineerprompt
      @engineerprompt  2 месяца назад +1

      Towards the end of the video, I show how to save and load the models locally. You can use that part of the code

  • @deepadharshinipalrajan8849
    @deepadharshinipalrajan8849 2 месяца назад +1

    Does unsloth support CPU configuration?

    • @engineerprompt
      @engineerprompt  2 месяца назад +1

      it needs GPU

    • @deepadharshinipalrajan8849
      @deepadharshinipalrajan8849 2 месяца назад

      ​@@engineerprompt Are we able to train the model which is available in ollama directly by taking that as base model?

  • @One.manuel
    @One.manuel 2 месяца назад +1

    You are not fine tuning a damn thing bro

    • @mike6335
      @mike6335 2 месяца назад

      Meaning… he wants a step by step vs a high level how to.

  • @ss56075
    @ss56075 2 месяца назад

    When I am using this code "model.push_to_hub_merged("My_Modal_Path", tokenizer, save_method="merged_16bit")" it shows this error "TypeError: argument of type 'NoneType' is not iterable". All files are saved successfully, but when unsloth trying to upload it shows this error.