Fine Tuning a Model in Gemini and Vertex AI | Steps to make a LLM

Поделиться
HTML-код
  • Опубликовано: 12 сен 2024

Комментарии • 36

  • @uncle_slashes
    @uncle_slashes 6 месяцев назад +8

    Based on the information in Google's doc - "The PaLM 2 for Text (text-bison, text-unicorn) foundation models are optimized for a variety of natural language tasks such as sentiment analysis, entity extraction, and content creation. The types of content that the PaLM 2 for Text models can create include document summaries, answers to questions, and labels that classify content." This video is not demonstrating Gemini.

  • @outcast2313
    @outcast2313 4 месяца назад +1

    great video! Thanks for this. Seems like its a lot simpler than having to understand how the tokenizer and architecture works and try to get your training dataset in that format. That being said, once tuned can the fine tuned model weights be downloaded to your local machine

  • @cnota69
    @cnota69 6 месяцев назад +2

    thank you so much for the instruction

    • @LevelUp_Plus
      @LevelUp_Plus  6 месяцев назад +1

      Thank you for watching, glad it was useful.

  • @mr.daniish
    @mr.daniish 6 месяцев назад

    Loved it!

    • @LevelUp_Plus
      @LevelUp_Plus  6 месяцев назад

      Thank you, and glad you enjoyed it. I had fun making the video.

  • @FatimaZahra-zb8ub
    @FatimaZahra-zb8ub 22 дня назад

    hey after tuning my data it is successfully tuned but when I test my model and try generating prompt I am getting this error
    ERROR Cannot read properties of null (reading 'length')

  • @AhmedZ-v4e
    @AhmedZ-v4e 20 дней назад

    How do you train it with images ?

  • @John-jt3dn
    @John-jt3dn 4 месяца назад +1

    It took vertex AI to fine tune the data 2 hours 12 minutes on my side just for 10 examples while i was testing. Is there a way to speed this up? My original data has 14467 examples. It would take ages....

    • @LevelUp_Plus
      @LevelUp_Plus  4 месяца назад +1

      Hi, Thanks for watching and appreciate the comment. In Vertex AI, I have seen a way to allocate a more powerful server/TPUS's to the environment . Alternatively you may try the Google colab version which allows you to select the processing power you would like. I have only very briefly experimented with runtimes and different CPU configurations.

  • @abdulrasool925
    @abdulrasool925 2 месяца назад

    I finetuned gemini 1.0 pro for text classification. While finetuning i also provided a validation dataset. It showed some metrics called "eval_fraction_of_correct_next_step_preds" to be around 0.93 at the end of training. I am assuming it as the accuracy. But when i run the same model for my validation set locally and caluculate accuracy it is coming around 0.77 Then what is the metric that is shown in the vertex ai studio while training?

  • @pradachan
    @pradachan 5 месяцев назад +1

    can you turn off the browser when you just uploaded jsonl file to get trained, would it get trained or i have to keep the pc on>?

    • @LevelUp_Plus
      @LevelUp_Plus  5 месяцев назад

      Hi thanks for watching, and yes you can close the browser and your PC, the the engine will run in the background and alert you once done.

  • @jonzh4720
    @jonzh4720 5 месяцев назад

    Thanks. Can make video for evaluating finetune results such as accuracy in vertex?

    • @LevelUp_Plus
      @LevelUp_Plus  5 месяцев назад

      Hi thanks for watching, I will have a look into this. thanks for the suggestion.

  • @HpVictus15
    @HpVictus15 2 месяца назад

    Can we fine tune again a fine tuned model with different dataset?

  • @John-jt3dn
    @John-jt3dn 4 месяца назад

    Hello.. Thank you for the video.. I have a question.. for inputs i have examples questions and 5 other input parameters and or course and output. How can i integrated there parameters into the trainin data set? Which format should i use? Thanks.

    • @LevelUp_Plus
      @LevelUp_Plus  4 месяца назад +1

      Hi thanks for watching, I would check the documentation as a first port of call. There may be some experimentation here. So keep the number of samples low to test it first.

  • @noureddinhamouda3438
    @noureddinhamouda3438 6 месяцев назад

    thank you very much for this video, can you record another video for the chat-bison, I'm always getting an error when the training comes to the dataset-encoder.

    • @LevelUp_Plus
      @LevelUp_Plus  6 месяцев назад +2

      Hi, the encoder issue is due to the training dataset you are using, it needs to be formatted perfectly. below is a sample of one line items I used in my JSONL file. {"input_text": "question: What is Tyrone's view on the future of AI in insurance? context: AI is possibly the most exciting technology to ever have been created in my view, a solution that allows access to model and information at unprecedented rates","output_text": "The Gen AI future is looking awesome, so excited and can't wait to help more business adopt it"}

  • @Ristontestaus-on8uq
    @Ristontestaus-on8uq 3 месяца назад

    Hi there, thanks a lot for this. Trying to test somethign with it, but I get the "Failed to create pipeline job. Error: Permission 'aiplatform.metadataStores.get' denied on resource" error. I think it has to do with the service accounts and permissions. For me it only shows the compute engine default service account. I've gone to create service accounts but over there the instructions I've seen ask me to download a key, and that's no longer allowed. Could you please nudge me in the right direction on how to get the service account setup properly so that I could select it here? Thanks!

    • @LevelUp_Plus
      @LevelUp_Plus  3 месяца назад

      Hi, if you haven't already you might need to go through the Google Cloud Setup steps to configure your environment, so click the Google Cloud Setup, and go through the wizard. The user access and service accounts are provisioned in there.

  • @MakeChangeNow
    @MakeChangeNow 5 месяцев назад

    what is the use case for making one? and how is it different than making an agent?

    • @LevelUp_Plus
      @LevelUp_Plus  4 месяца назад +1

      Hi Thanks for watching, the use case for refining a LLM is to provide it with context for your data/environment. For instance if your marketing material is based on a certain style or PR questions and answers follow specific format, or if you want the model to behave like your organisation. Or if you are a specialised insurance/finance with use cases that you have as knowledge banks then using your business IP you can train the LLM further to be more accurate.

  • @KadirOrtac-xj9ww
    @KadirOrtac-xj9ww 5 месяцев назад

    i've got an error like this Failed to process dataset for gemini-1.0-pro-002: Dataset validation failed: {"missing_messages_list": while tuning model. What am i supposed to do?

    • @LevelUp_Plus
      @LevelUp_Plus  5 месяцев назад

      Hi thanks for watching, I found the biggest reason for the failures was the JSONL file. Check the format of the file and ensure the format and end lines etc are correct.

  • @ayushsinghal28
    @ayushsinghal28 6 месяцев назад

    how can we fine tune the multimodal Gemini AI pro

    • @LevelUp_Plus
      @LevelUp_Plus  6 месяцев назад

      Hi, the options are being made available in different countries at different stages. So depending on the country you are in, will be when it is available.

  • @DeepakPoojary
    @DeepakPoojary 4 месяца назад

    dude the code u copied at 3:20 and pasted at 3:50 are two different codes , can u please give this code

    • @LevelUp_Plus
      @LevelUp_Plus  4 месяца назад

      Hi Thanks for letting me know - here is the JSON. below is a sample of one line items I used in my JSONL file. {"input_text": "question: What is Tyrone's view on the future of AI in insurance? context: AI is possibly the most exciting technology to ever have been created in my view, a solution that allows access to model and information at unprecedented rates","output_text": "The Gen AI future is looking awesome, so excited and can't wait to help more business adopt it"}

  • @KadirOrtac-xj9ww
    @KadirOrtac-xj9ww 5 месяцев назад

    can i use my tuned model in my web application for free?

    • @LevelUp_Plus
      @LevelUp_Plus  5 месяцев назад +1

      Hi thanks for watching, the costs for the models are in the vertexAI documentation, and will be token based.

    • @berrabeser7155
      @berrabeser7155 4 месяца назад

      ⁠​⁠@@LevelUp_Plus can we use another environment with the same code that we get from the vertex?