BERT vs GPT

Поделиться
HTML-код
  • Опубликовано: 18 янв 2023
  • #machinelearning #shorts #deeplearning #chatgpt #neuralnetwork #datascience

Комментарии • 23

  • @darshantank554
    @darshantank554 Год назад +51

    One is for Natural language understanding and another is for Natural language generation

    • @Ram_jagat
      @Ram_jagat Месяц назад

      true

    • @borntodoit8744
      @borntodoit8744 10 дней назад

      Remember it as INPUT>MODEL >OUTPUT
      MODEL INPUT (NLU) - txt recognition, vision recognition/image/movie, sound recognition/voice,
      MODEL OUTPUT (NLG) -txt generation, image/movie generation, sound/voice generation + tool integration
      MODEL PROCESSING -Basic (classification summeration extraction) advanced (reasoning, planning, orchestration)

  • @VarunTulsian
    @VarunTulsian Год назад +28

    this is very useful. Just wanted to add that the gpt decoder doesn't have the cross attention in the transformer block.

    • @Tech_kenya
      @Tech_kenya 10 месяцев назад

      What is cross attention

    • @methylphosphatePOET
      @methylphosphatePOET 9 месяцев назад +3

      @@Tech_kenya It's when word vectors reference other word vectors as opposed to just referencing themselves.

    • @imran7TW
      @imran7TW 16 дней назад

      @@methylphosphatePOET so kinda the opposite of self attention?

  • @JillRhoads
    @JillRhoads 7 месяцев назад

    I hadnt known that BERT was an acronym and had been wondering why the Sweden LLM was called Bert. I wonder if this is why. Thanks for the info!

  • @maninzn
    @maninzn Год назад +1

    Great explanation. For eg, if I have to read all the client emails and understand their requirements and auto create tasks based on that prediction, which model should I go for? BERT or GPT?

  • @contactdi8426
    @contactdi8426 3 месяца назад

    Can you please explain about their training process?

  • @cs101-qm2ud
    @cs101-qm2ud 5 месяцев назад

    Wonderfully put.

  • @nicholaszustak6299
    @nicholaszustak6299 10 месяцев назад +1

    So BERT doesn’t have a decoder? Did I misunderstand

  • @vladislavkorecky618
    @vladislavkorecky618 11 месяцев назад +5

    What if I stack both encoders and decoders? Do I get some BERTGPT hybrid?

    • @davronsherbaev9133
      @davronsherbaev9133 10 месяцев назад +1

      there is also Whiser model, that has similar facebook BART decoder part, but has audio decoder.

  • @Dr_Larken
    @Dr_Larken Год назад +1

    Bert also Drives a trans am!
    Jokes aside I do appreciate your videos!

  • @saimadhaviyalla5682
    @saimadhaviyalla5682 9 месяцев назад +1

    transformer models are usually parallelly run right?

    • @eugeneku3239
      @eugeneku3239 8 месяцев назад

      Not when it's decoding. No.

  • @hubgit9556
    @hubgit9556 11 месяцев назад

    good

  • @obieda_ananbeh
    @obieda_ananbeh Год назад

    Awesome 👏

  • @usama57926
    @usama57926 Год назад

    I love you ❤