Lecture 2: Label Errors

Поделиться
HTML-код
  • Опубликовано: 7 сен 2024
  • Introduction to Data-Centric AI, MIT IAP 2023.
    You can find the lecture notes and lab assignment for this lecture at dcai.csail.mit....

Комментарии • 13

  • @ThiagoMotta4Ever
    @ThiagoMotta4Ever Год назад +2

    Thank you for sharing this course, it's fantastic!

  • @williamagyapong6337
    @williamagyapong6337 Год назад +6

    The formula for the t_j’s is not giving me the same values in the presentation and I have a feeling that I’m probably not applying it correctly. Can anyone explain how to get t_dog = 0.7 based on the given noisy labels and the predicted probabilities, for instance?

    • @asdfghjkl743
      @asdfghjkl743 Год назад +1

      None of the threshold's are matching. t_dog = 0.3(1st image)+0.9(6th image)/2 = 0.6. Can someone break down for 1 class if i am wrong

    • @arnabsaha6756
      @arnabsaha6756 Год назад

      @asdfghjkl743 I am getting the same values for t_j's as you. The slides are incorrect.

    • @manigoyal4872
      @manigoyal4872 9 месяцев назад

      sum up the probabilities of each type and divide by the number of images of each type
      for fox, it is (0.7+0.7+0.9+0.8+0.2)/5 = 0.7
      if i am not wrong

    • @HaotianLi6
      @HaotianLi6 9 месяцев назад

      @@manigoyal4872 I think the formula on the prev slide means t_fox is only computed based on the records where its noisy label, i.e. y^tilde=fox, so only 4 images (no.2-5)

  • @turtleyoda7703
    @turtleyoda7703 5 месяцев назад

    Awesome, thank you for sharing!
    One question though:
    IIUC all of this assumes that the model we plan on training is in fact a good estimator of the phenomenom we're trying to model. I understand how the algorithm works in that case.
    However, how do we validate that assumption? What if I'm using a terrible model, how would I know? After using confident learning to clean the datased I'd thing that I now have a better dataset, but I don't think that's achievable through a bad model.

  • @sugoilang
    @sugoilang Год назад +7

    Could anyone told me why the slide in 37:43 align the left most image which has noisy label:dog and higest probability 0.7 as fox to y~ = fox and y* = dog in the table?

    • @dcai-course
      @dcai-course  Год назад +1

      Good find, that's a bug in the slides, images 1 and 5 should be swapped. The first image should have \tilde{y}=dog and y^*=fox.

    • @majovlasanovich9047
      @majovlasanovich9047 Год назад +1

      There is a mistake in the slides, the blue circle examples are switched.

    • @Levy957
      @Levy957 Год назад

      @@majovlasanovich9047 i noticed that right now because i explained that to my dad

    • @williamagyapong6337
      @williamagyapong6337 Год назад

      Good catch

  • @Zeehunt101
    @Zeehunt101 9 месяцев назад

    this course hard to learn, is there anyone just recommend any course should I take and then study this !!