Text Representation Using TF-IDF: NLP Tutorial For Beginners - S2 E6

Поделиться
HTML-код
  • Опубликовано: 11 сен 2024

Комментарии • 30

  • @codebasics
    @codebasics  2 года назад +3

    Check out our premium machine learning course with 2 Industry projects: codebasics.io/courses/machine-learning-for-data-science-beginners-to-advanced

  • @user-fz7db4ls3i
    @user-fz7db4ls3i 2 года назад +11

    Your videos are really such great source of knowledge, especially for me as a beginner. I'm trying to find the roadmap to become an NLP engineer, pls don't stop making videos.

  • @robertkumar7768
    @robertkumar7768 Год назад +1

    A big thank you sir for explaining the concepts in simple ways.

  • @mujtabasultani5712
    @mujtabasultani5712 2 года назад +1

    thannnnnnnnnnnnk you very much dear Dhaval, the way you're teaching is amazing..., really beneficial for us, hope you continue the series till end.

  • @shaileshmahto7690
    @shaileshmahto7690 Год назад +3

    According to the explanation at 10:15 , log is used in IDF formula to dampen the effect of term occurring too often. But isn't the effect of frequency of term captured in the TF(Term Frequency) part of the formula(& not the IDF part). IDF, instead captures the value of each term based on WHETHER it occurs in most documents or not. So even if the term occurs 1 million times in one document and never in other documents, its IDF value would be same as if the term occurs only 1 time in one document and never in other documents - since we are checking for number of docs that have term present & not how many times the term occurs. Hence, for my example, both the scenarios will assign the same high IDF value to the term. Hence, I dont see the dampening of the importance of the term that has very high frequency caused due to higher term frequency. Please clarify. Thank you for the practical lessons that are free and easy to understand.

  • @pouriaforouzesh5349
    @pouriaforouzesh5349 Год назад +1

    It could not better than this 🙏

  • @Raaj_ML
    @Raaj_ML 5 месяцев назад +2

    Great tutorial. But the explanation for using log in IDF gives a wrong reason..please check..

  • @minakshisontake3585
    @minakshisontake3585 Год назад +1

    such a great explanation . thank you Dhaval sir.

  • @Lava_Kumar
    @Lava_Kumar 5 месяцев назад +1

    We have to convert it into lowercase also in pre processing

  • @aradhyadhruv9084
    @aradhyadhruv9084 Год назад +1

    Thanks a lot sir and please keep making more videos!!

  • @user-en1eh5ux7d
    @user-en1eh5ux7d Месяц назад

    Thank you very much. Great explanation!

  • @user-nj1jo9tv5e
    @user-nj1jo9tv5e 8 месяцев назад

    Thank you so much for such a clear explanation.

  • @B515R
    @B515R Месяц назад

    AMAZING !! 😍😍

  • @nriezedichisom1676
    @nriezedichisom1676 5 месяцев назад

    Thank you. You are the best

  • @harsh2014
    @harsh2014 Год назад

    Thanks for your great effort !

  • @svensalvatore8702
    @svensalvatore8702 6 месяцев назад

    Sir big fan!

  • @amolkaushal224
    @amolkaushal224 Час назад

    I am facing a issue . I had a dataset which had 60398 test description rows and i cleaned the text did lamitization and stemming . After that i used TFID vectorization to convert text into matrix form.and the matrix shape it is showing (60398x104757). It is getting trained well using SVC . But when i am creating a predictive system and trying to predict same label column for a new data which has less number of rows (10000) after cleaning the new data and transforming into matrix form using TFID. When i am trying to predictict using. Model. Predict(X) i am facing a error " X has 10525 features, but SVC is expecting 104757 features as input". How to correct this error.

  • @semrana1986
    @semrana1986 Месяц назад

    nice work, where is the TF score computed?

  • @souravbarua3991
    @souravbarua3991 Год назад

    Nice and simple explanation. Pls perform checking the model with new text data in the tutorial. Because I tried to check the model with new text data while practicing, it shows an error.

    • @user-on9zx1rg2n
      @user-on9zx1rg2n 10 месяцев назад

      which error had occurred? if u can give the description of the error will try to figure out

  • @n3cr0manz3r6
    @n3cr0manz3r6 2 года назад +3

    Hi Dhaval, it would be of great, if you explain how to deploy this model in your upcoming videos.

    • @codebasics
      @codebasics  2 года назад +4

      ok, actually I have made model deployment videos before in my data science projects. Search "codebasics data science projects" and in those project series you will find deployment videos. I will add separate videos for deployment in this series too when I post videos on end to end NLP projects.

    • @marcellodichiera
      @marcellodichiera 2 года назад

      @@codebasics hope you ll use Streamlit for deployment :) .. always thanks for your precious tutorials ..🙏🙏

  • @vishnuj7470
    @vishnuj7470 Год назад

    Here why are we not using one hot encoding instead of labeling?It could be much better right

  • @jasonpot5669
    @jasonpot5669 6 месяцев назад

    how can i apply tfidf to only one colum means in your dataset, to df['text']?

  • @sanketadamapure802
    @sanketadamapure802 Месяц назад

    22:33 It's False. machine learning models are versatile tools that can process and learn from various data formats.

  • @anirbanc88
    @anirbanc88 Год назад

    15:14 why "already" has 0, if its non existent in the corpus, how is it being added to the vocabulary?

    • @matpro0
      @matpro0 Год назад

      0 is the index, not the count

  • @mohammedjaddoa9783
    @mohammedjaddoa9783 11 месяцев назад

    you used different dataset from kaggel