Generative AI Fine Tuning LLM Models Crash Course

Поделиться
HTML-код
  • Опубликовано: 27 ноя 2024

Комментарии • 69

  • @yogeshmagar452
    @yogeshmagar452 6 месяцев назад +25

    Krish Naik respect Button❤

  • @dvb22sagar
    @dvb22sagar 4 месяца назад

    Full Respect to you Krish, Great video !!

  • @svitlanatuchyna7154
    @svitlanatuchyna7154 4 месяца назад

    Amazing as always! So great tutorials and clear explanations! Thank you!

  • @Mohama589
    @Mohama589 6 месяцев назад +1

    full respect bro , from morocco MA.

  • @anuradhabalasubramanian9845
    @anuradhabalasubramanian9845 5 месяцев назад

    Awesome presentation Krish !!!! You are a superstar!!!

  • @muhammadhassan2484
    @muhammadhassan2484 4 месяца назад +1

    Summary of the course.
    Course Overview: This crash course by Krish Naak covers theoretical concepts and practical implementation of fine-tuning large language models (LLMs), including techniques such as quantization, LoRA, and CLA PFT.
    Fine-Tuning Techniques: The course discusses different fine-tuning methods like quantization-aware training, matrix decomposition, and domain-specific fine-tuning for various applications like chatbots.
    Technical Concepts: Explains floating-point precision (FP32, FP16), tensor data types in TensorFlow, and quantization methods (e.g., 4-bit normal float) used to optimize model performance and memory usage.
    Implementation Steps: Demonstrates the process of preparing datasets, configuring training parameters (like optimizer, learning rate), and using the LoRA configuration for fine-tuning models such as LLaMA 2.
    Practical Application: Provides a hands-on example of loading datasets, setting up the training environment, and fine-tuning a model using custom data, with plans to push the fine-tuned model to platforms like Hugging Face.

  • @prekshamishra9750
    @prekshamishra9750 6 месяцев назад +1

    Krish...yet again!! I was just looking for your finetuning video here and you uploaded this..I cant thank you enough..really 👍😀

    • @sohampatil8681
      @sohampatil8681 6 месяцев назад

      Can we connect brother. I am new into generative AI and wanted to know the basics .

  • @lalaniwerake881
    @lalaniwerake881 6 месяцев назад

    Amazing content, big fan of you :) Much love from Hawaii

  • @eduxlearntodayapplytomorro6616
    @eduxlearntodayapplytomorro6616 2 месяца назад

    GREAT WORK SIR! Love from ghaziabad

  • @BabaAndBaby11
    @BabaAndBaby11 6 месяцев назад

    Thank you very much Krish for uploading this.

  • @senthilkumarradhakrishnan744
    @senthilkumarradhakrishnan744 6 месяцев назад

    just getting your video at the right time !! Cudos brother

  • @souvikchandra
    @souvikchandra 5 месяцев назад

    Thank you so much for such an comprehensive tutorial. Really love your teaching style. Could you also refer some books on LLM fine tuning.

    • @ayonsomaddar5858
      @ayonsomaddar5858 3 месяца назад

      Generative AI on AWS by Chris Fregly and Shelbee Eigenbrode
      its a good one

  • @RaahgirHarshal
    @RaahgirHarshal 18 дней назад

    32GB RAM is less for Krish Bhai.....I want to be like Krish Bhai.

  • @tejasahirrao1103
    @tejasahirrao1103 6 месяцев назад +1

    Please make a complete playlist to secure a job in the field of Ai

  • @deepaksingh-qd7xm
    @deepaksingh-qd7xm 6 месяцев назад +2

    i dont know why i feel training a whole model from scratch is much more easier for me than to fine tune it ..............

    • @kartik1409
      @kartik1409 5 месяцев назад +1

      Ya if u see training the model from scratch for your dataset might look better and optimal but the energy is used in training a model from scratch is too much so finetuning a pretrained model is considered a better option than training model for specific data everytime....

  • @abhishekvarma4449
    @abhishekvarma4449 6 месяцев назад

    Thanks Krish it's very helpful

  • @deekshitht786
    @deekshitht786 5 месяцев назад +1

    You are awesome ❤

  • @tejasahirrao1103
    @tejasahirrao1103 6 месяцев назад +2

    Thank you krish

  • @viranchivedpathak4231
    @viranchivedpathak4231 4 месяца назад

    Great RUclips ssshhhaaaaanel for LLM

  • @mudassarm30
    @mudassarm30 4 дня назад

    Different different :) video, from others!

  • @dear_nidhi
    @dear_nidhi 6 месяцев назад

    we want more video on fine tuning projects

  • @sadiazaman7903
    @sadiazaman7903 6 месяцев назад

    Thank you for an amazing course as always. Can we please get these notes as well. they are really good for quick revision.

  • @yourmirror6
    @yourmirror6 6 месяцев назад

    Brilliant brilliant 🙌

  • @o__bean__o
    @o__bean__o 6 месяцев назад

    Krish bro ❤

  • @shalabhchaturvedi6290
    @shalabhchaturvedi6290 6 месяцев назад

    Big salute!

  • @foysalmamun5106
    @foysalmamun5106 6 месяцев назад

    Hi @krishnaik06,
    Thank you again for anther Crash Course.
    may I know which tools/software are you using for presentation?

  • @virkumardevtale9671
    @virkumardevtale9671 6 месяцев назад

    Thanks you very much sir🎉🎉🎉

  • @miltonvai5816
    @miltonvai5816 5 месяцев назад

    Hello Sir, Hello, could you create a tutorial on fine-tuning vision-language models like LLaVA or Multimodal LLMs like IDEFICS for Visual Question Answering on datasets like VQA-RAD, including evaluation metrics?
    Please make a full step by step tutorial. Thanks in Advance!

  • @shakilkhan4306
    @shakilkhan4306 6 месяцев назад

    Thanks man!

  • @AntonyPraveenkumar
    @AntonyPraveenkumar 6 месяцев назад

    Hi Krish, the video is really good and more understanding. but I have one reason how to you choose this the right dataset and why? why you choosing that format_func function to format the dataset into the some kind of format. if you have any tutorial or blog please share the link.

  • @Ari_Shukla
    @Ari_Shukla 6 месяцев назад +1

    Can you make a good video around how to decide hyper parameters when training gpt 3.5

  • @AshwaniKumar-r4p
    @AshwaniKumar-r4p 12 дней назад

    I completed a fine training a llama 3.2 model with a custom data set and created a gguf file after training. but when we run this GGUF file the response of the model does not match what I trained with the data
    tell me how can trained model and get the same response as in available in the dataset ?

  • @anupampandey1235
    @anupampandey1235 6 месяцев назад

    hello krish sir thank you for amazing lecture can please share the notes of session

  • @Jeganbaskaran
    @Jeganbaskaran 6 месяцев назад

    Krish, most of the fine tuning done by the existing dataset from HF. however converting the dataset as per the format its a challenging for any domain dataset. How we can train our own data to finetune the model so that accuracy ll be even better. Any thoughts?

  • @rotimiakanni8436
    @rotimiakanni8436 6 месяцев назад

    Hi Krish. What device do you use to write on...like a board

  • @arunkrishna1036
    @arunkrishna1036 6 месяцев назад +1

    After the fine tuning process in this video, isn't it the same old model that is used here test the queries? We should have tested the queries with the "new_model" isn't it?

  • @irshadali3515
    @irshadali3515 4 месяца назад

    Hi krish, one question from my side..can I fine tune gpt2 for text classification

  • @adityavipradas3252
    @adityavipradas3252 6 месяцев назад

    RAG or fine-tuning? How should one decide?

  • @maximusrayvego2673
    @maximusrayvego2673 5 месяцев назад

    hey could you tell me what are the pre req to follow this crash course? it would be greatly beneficial!!

  • @rebhuroy3713
    @rebhuroy3713 6 месяцев назад

    Hi Krish, i Have seen entire video. i am confused with 2terms. some times you said its possible to train with my own data (own data refers from a url , pdfs , simple text etc) but when you are trying to train the llm model you are giving inputs as in certain format like### question : ans.
    Now if i want to train my llm in real life scenario i don't have my data in this instruction format right in that case what to do. And its not possible to transform my raw text to into that format right how to handle that situation . is it a only way to fine tune in specific format or i can train given in raw text format i know a process where i need to convert my text to chunks then pass to llm. those are really confusing can you clear those things

  • @thecaptain2000
    @thecaptain2000 Месяц назад

    err -2/4 is not equal to -5 but to -0.5, unless I missed something, did I?

  • @JunaidKhan80121
    @JunaidKhan80121 6 месяцев назад +1

    Can anybody tell me how to fine-tune llm for multiple tasks?

  • @MKtechnoverse
    @MKtechnoverse Месяц назад

    can we fine tune this is our local system does it supports
    i have spec:
    16gb RAM
    ryzen 7 4000 8 core
    NVIDIA RTX 3050
    512 SSD

  • @avd_suresh
    @avd_suresh 6 месяцев назад

    Hello Krishna sir ,
    Please make a playlist for genai and lanchain

    • @krishnaik06
      @krishnaik06  6 месяцев назад

      Already made please check

    • @avd_suresh
      @avd_suresh 6 месяцев назад

      @@krishnaik06 Thank you for replying me

  • @sibims653
    @sibims653 6 месяцев назад

    What documentation did you refer to in this video?

  • @rakeshpanigrahy7000
    @rakeshpanigrahy7000 6 месяцев назад

    Hi sir, I have tried your llama finetuning notebook to run on colab with free T4 gpu but it is throwing OOM error. So could you please guide

  • @rishiraj2548
    @rishiraj2548 6 месяцев назад

    🙏💯💯

  • @SheiphanJoseph
    @SheiphanJoseph 6 месяцев назад

    Please also provide the source. Research paper/Blog you might have referred for this video.

  • @gsayak
    @gsayak 4 месяца назад

    Anyone getting stuck at the import it is
    "from transformers import AutoModelForCausalLM"
    I kept getting error for reading it "casual" instead of "causal"😭

  • @RAZYOUSUFI-z7h
    @RAZYOUSUFI-z7h 3 месяца назад

    Krish, how to retrieve data from an API, like OpenWheatherData, instead of retrieving from Google and Wikipedia?

  • @jaaferklila1375
    @jaaferklila1375 2 месяца назад +1

    can we apply Lora for bert please reply

  • @charithk2160
    @charithk2160 6 месяцев назад

    hey krish , can you by any chance share the notes used in the video. would be really helpful. thanks !!

  • @abhaypatel2585
    @abhaypatel2585 5 месяцев назад

    actually sir this step cant able to run
    !pip install -q datasets
    !huggingface-cli login
    due to this dataset cant be load nd getting error in other step
    so is thier is any solution for this ?????

  • @hassanahmad1483
    @hassanahmad1483 6 месяцев назад

    How to deploy these?...I have seen deployment of custom LLM models...how to do this?

  • @abdullahmehmood2941
    @abdullahmehmood2941 2 месяца назад

    i am unable to make gradient ai account it says not allowed while signing up

  • @stalinjohn721
    @stalinjohn721 5 месяцев назад

    how to finetune and quantize the phi3 mini model ,

  • @navjeetkaur3014
    @navjeetkaur3014 5 месяцев назад

    @krishnaik06 WANDB_Disabled is for disabling weights and Bais of the current model

  • @nitinjain4519
    @nitinjain4519 6 месяцев назад

    Can anyone suggest how to analyze audio for soft skills in speech using Python and llm models?

  • @rahulmanocha4533
    @rahulmanocha4533 6 месяцев назад

    If i would like to join data science community group where i can get the access, please let me know.

  • @shivam_shahi
    @shivam_shahi 6 месяцев назад

    EK HI DIL HAI
    KITNE BAAR JITOGE SIR?

  • @nitishbyahut25
    @nitishbyahut25 6 месяцев назад

    Pre-requisites?

  • @salihedneer8975
    @salihedneer8975 6 месяцев назад

    Prerequisite ?

  • @RensLive
    @RensLive 6 месяцев назад

    I understand this video just like your hairs sometime nothing some time something ❤🫠