BERT for Topic Modeling - EXPLAINED!

Поделиться
HTML-код
  • Опубликовано: 22 дек 2024

Комментарии • 35

  • @nextXstep
    @nextXstep 3 года назад +1

    Couldn't find a single article explaining the codes behind bertopic, the explanation in this video is absolutely perfect thanks!!

  • @pierrickleroy780
    @pierrickleroy780 2 года назад +1

    Another clear explanation of multiple sota useful concepts, cheers man I really like your videos and way of communicating things !

  • @IRiViI
    @IRiViI 3 года назад +2

    I just wanted to say that I love your videos!

  • @chetanetrx2005
    @chetanetrx2005 3 года назад +4

    Nice video..
    Can you please explain how sentence transformer works at inference time when we have only one sentence?

  • @MLA263
    @MLA263 Год назад

    This is amazing, thank you, you hero

  • @lepton555
    @lepton555 3 года назад +1

    Awesome job, man! Absolutely horrific! Thank you!

  • @kabeerjaffri4015
    @kabeerjaffri4015 3 года назад +4

    Hey can you make a video on unsupervised temporal action localization in video. like when you search google for somthing and they show time intervals where video content matches your query.
    I think its a great topic and may spark your interest. BTW great content as usual

  • @lucyledezma709
    @lucyledezma709 2 года назад

    Great video!!!🤗🤗🤗

  • @gunavardhinip
    @gunavardhinip Год назад

    Brother will you explaint bart for text summarization

  • @teetanrobotics5363
    @teetanrobotics5363 3 года назад

    Amazing content bro. Please keep updating your playlists.

  • @yazdipour
    @yazdipour 3 года назад

    love ur bert/nlp contents

    • @CodeEmporium
      @CodeEmporium  3 года назад

      Glad you do. I'll try making more of this :)

  • @emanuelgerber
    @emanuelgerber 3 года назад

    Thanks for making this video! Helps a lot

  • @boubacarbah1455
    @boubacarbah1455 2 года назад

    Hello , i'm trying to reproduce your exercice. But i got a problem when i tried to import BERTOPIC " from import bertopic ".I get this error " no module named "llvmlite.binding.dylib". And i could not fix it; Si i wonder if you have a solution ?

  • @Daniel-gy1rc
    @Daniel-gy1rc 2 года назад

    can u make a video about Google-Palm?

  • @datasci4547
    @datasci4547 3 года назад

    Very clear explanation. Thank you :)

  • @alexnim4873
    @alexnim4873 3 года назад

    Amazing video bro. Very helpful. Thanks so much

  • @wilfredomartel7781
    @wilfredomartel7781 Год назад

    Just a doubt. So tripplet dataset is better for improving embeddings. May a video of how to fine tuning a non english transformer?

    • @CodeEmporium
      @CodeEmporium  Год назад +1

      Sure thing. The goal is to make a series in BERT training and code soon from scratch. In the mean time, maybe you’ll enjoy the playlist called “Transformers from Scratch” where I build a translator for a non-English language. Though there is no “pre-training” and “fine tuning”, many components of BERT are similar to the transformer. So I recommend checking that out

    • @wilfredomartel7781
      @wilfredomartel7781 Год назад

      @@CodeEmporium thank you to take your time in resondijg my message. Maybe a link suggestion my friend.

    • @CodeEmporium
      @CodeEmporium  Год назад

      Transformers from scratch
      ruclips.net/p/PLTl9hO2Oobd97qfWC40gOSU8C0iu0m2l4

  • @pranaymehta8471
    @pranaymehta8471 3 года назад

    Have you tried to work with BERTopic with datasets that are bigger in size ? For instance 100k, 500k data sizes ? From what I have seen, the sentence transformer takes a lot of time to create the N dimensional embeddings . I am not sure if berTopic runs things in parallel.

    • @CodeEmporium
      @CodeEmporium  3 года назад +2

      Good question. Sorry I'm late to this. I've worked with Sentence Transformers in general and i can say the speed and quality really depends on which sentence transformer you use. You'll just need to choose the one that balances both qualities if speed is Essential (like for online applications as opposed to postmortem analysis)
      Also BERT and hence the sentence transformer process information in parallel. I know Sentence Transformers have this method called "embedding()" where you can pass in a list and we fetch the embeddings in parallel

  • @user-or7ji5hv8y
    @user-or7ji5hv8y 3 года назад

    Really well explained

  • @maryamaziz3841
    @maryamaziz3841 3 года назад

    Hi, How can use BERT to word embedding

  • @anonxnor
    @anonxnor 3 года назад

    You should be able to do this with the CLS token embeddings, instead of sentence embeddings from S-BERT, if you use regular BERT right?

  • @geoffreyanderson4719
    @geoffreyanderson4719 3 года назад

    My spidey-sense tingles for me when more than half of the topics in the corpus are unclustered. Exploratory data analysis on those might reveal some easily fixed errors. Like, I might want to see what happens when you topic model just those bad bois, letting all the rest of the good bois to go home on the regular schoolbus. Maybe the Breakfast Club misfits have something in common with each other after all.

    • @CodeEmporium
      @CodeEmporium  3 года назад

      Aye good eye. I was mostly just trying to illustrate BERT. But some analysis on this would have been nice too as a follow up

  • @Burakyesilvlogs
    @Burakyesilvlogs 3 года назад

    Great video

  • @greenital
    @greenital 3 года назад

    in love ** divine * :D