Train a Small Language Model for Disease Symptoms | Step-by-Step Tutorial

Поделиться
HTML-код
  • Опубликовано: 21 авг 2024

Комментарии • 42

  • @Hellow_._
    @Hellow_._ 7 месяцев назад +2

    subscribed just after watching one video. Thanks to creator. plz keep posting videos related to GenAI.

  • @chuanjiang6931
    @chuanjiang6931 17 дней назад +1

    outputs = model(input_ids=inputs, labels=targets), is there a reason you used the same tensor for both inputs and targets. Should targets be moved 1 place after inputs?

  • @aneesarom
    @aneesarom 4 дня назад +1

    why inputs and targets are same?

  • @zeta_meow_meow
    @zeta_meow_meow 7 месяцев назад +1

    saving, will watch later when i am good in ml and python

  • @sohailhosseini2266
    @sohailhosseini2266 7 месяцев назад +1

    Thanks for sharing!

    • @AIAnytime
      @AIAnytime  7 месяцев назад

      Thanks for watching!

  • @namashaggarwal7430
    @namashaggarwal7430 7 месяцев назад +3

    Awesome tutorial ❤
    Can you please make a tutorial on how to fine tune a model, especially on textual and image data?

  • @siriyakcr
    @siriyakcr 4 месяца назад

    Much needed

  • @Mr.Tom.handle
    @Mr.Tom.handle 2 месяца назад

    Need a video or instructions to integrating the model into mobile application

  • @VaibhavPatil-rx7pc
    @VaibhavPatil-rx7pc 7 месяцев назад

    Excellent information 🎉

    • @AIAnytime
      @AIAnytime  7 месяцев назад

      Glad it was helpful!

  • @nunoalexandre6408
    @nunoalexandre6408 8 месяцев назад

    Love it!!!!!!!!!!!!!!!!!!!!!!!

  • @user-iu4id3eh1x
    @user-iu4id3eh1x 8 месяцев назад +1

    Oh wow .... Thank you for this tutorial ❤

    • @AIAnytime
      @AIAnytime  8 месяцев назад +2

      Glad you like it!

    • @sriharsha5701
      @sriharsha5701 7 месяцев назад

      @@AIAnytime i am having 35 columns if i ned to get all information based on ID how i would train please explain

  • @mingyukang6592
    @mingyukang6592 7 месяцев назад +1

    Sorry, I have a beginner question.
    1. What is the difference between using data in a data frame and using it by creating a model and loading it like now?
    2. Is the 'SmallMedLM.pt'' model created in the example simply a vector database of "QuyenAnhDE/Diseases_Symptoms" data?
    3. Or do you mean that ‘SmallMedLM.pt’ is a GPT2-based LLM fine-tuned with symptoms matching the disease?
    4. So, does this mean that the trained 'SmallMedLM.pt' can be used when creating a chatbot related to a specific disease or symptom?

  • @ruksharalam173
    @ruksharalam173 5 месяцев назад +1

    what extensions are you using for code autocomplete in colab?

  • @subramanianchenniappan4059
    @subramanianchenniappan4059 7 месяцев назад

    Please tell if you have any training on how to train and deploy a model given a dataset . I am an experienced developer want to learn machine learning and AI and related stuff😊😊

  • @subashpaudelhimself964
    @subashpaudelhimself964 3 месяца назад

    provide a code to push this model to hugging face too sir.

  • @user-qi4jw1lf9i
    @user-qi4jw1lf9i 7 месяцев назад

    WOW IT IS JUST A AWESOME I HAVE 36000 DOC AND WAS USING LAMINI ' YOUR VEDIO WITHOUT INTERNET BUT IT IS HOW CAN I BUILD MODEL USING THAT DOCS SO FETCHING TIME WILL BE LOW

  • @sriharsha5701
    @sriharsha5701 7 месяцев назад

    i am having 35 colum,s if i ned to get a;; information based on ID how i would train please explain

  • @alroygama6166
    @alroygama6166 7 месяцев назад

    I have a classificationstask where i have a para of text and it classifies labels of the para.
    Can i use this similar approach to tune it?
    Also will llama be a better choice?

  • @adityashinde436
    @adityashinde436 7 месяцев назад

    my usecase is that the input will be some product keyword and output should be product category it belongs to (example - input: white shirt, output: apparel). for this purpose which model will be suitable, distilgpt2 is good or do you recommend some other models in text generation section or do you recommend me to check model from other section like roberta or distilbert?

  • @ruksharalam173
    @ruksharalam173 5 месяцев назад

    What are the differences between encode and encode_plus?

  • @jorgerios4091
    @jorgerios4091 6 месяцев назад

    Buddies, it runs on colab but when I try it in my Windows VS code environment I get the following error: RuntimeError: PyTorch is not linked with support for mps devices. Anybody else got the same and how did you fix it?

    • @jorgerios4091
      @jorgerios4091 6 месяцев назад

      Got it now, Gemini told me mps is just for Apple 🤡

  • @jorgerios4091
    @jorgerios4091 8 месяцев назад

    The steps presented for training the model are incredibly valuable, thanks for your guidance. Would there be any notable differences in the results if we were to utilize a LaMini model instead of GPT2?

    • @AIAnytime
      @AIAnytime  8 месяцев назад +3

      In my experiments, T5 was better .... I am working on that to create a video shortly for a similar use case.

    • @ARkhan-xw8ud
      @ARkhan-xw8ud 7 месяцев назад

      is there any metrics to check the performance of the LLMs
      @@AIAnytime

    • @AIAnytime
      @AIAnytime  7 месяцев назад +2

      Yes look at my Evaluation of LLMs and RAGs video. It's detailed.

  • @dragneel4451
    @dragneel4451 7 месяцев назад

    Can I train this model on movies dataset with columns Name, Synopsis, Genres and ask the model to recommendation similar kinds of movies?

    • @AIAnytime
      @AIAnytime  7 месяцев назад +1

      Absolutely. You must do it. Try T5 model instead distilGPT

  • @akash_a_desai
    @akash_a_desai 7 месяцев назад +1

    Convert it to ggml format so we can use it in cpu

    • @AIAnytime
      @AIAnytime  7 месяцев назад +1

      Won't recommend to convert this into GGUF as the model is already very small (around 380MB). Compressing it more will further degrade the performance. But if the data quality is better and other LMs like t5 and lamini are used then we can try GGUF. The current one smoothly runs on CPU tho.

    • @nothing_is_real_0000
      @nothing_is_real_0000 7 месяцев назад

      @@AIAnytime Agree. Any model within 1 GB is very much within "small" model range

  • @sumitawasthi7068
    @sumitawasthi7068 7 месяцев назад

    Why did Manchester United bring a map to the game against West Ham?

  • @Sergatx
    @Sergatx 8 месяцев назад

    Why didnt you do this with GPT 3.5 instead ?

    • @pjoshi_15
      @pjoshi_15 7 месяцев назад

      If you don't want to expose your data to open AI

    • @AIAnytime
      @AIAnytime  7 месяцев назад +4

      1. Gpt 3.5 in a closed source. You need to pay. 2. Data protection and privacy. 3. Inference token costs.

    • @asiddiqi123
      @asiddiqi123 5 месяцев назад

      ​@@AIAnytime
      So you don't pay Hugging face?

  • @CooperDaniel87
    @CooperDaniel87 7 месяцев назад +1

    MAN U 💔