Principal Component Analysis | Dimensionality Reduction Machine Learning | Applied AI Course

Поделиться
HTML-код
  • Опубликовано: 19 ноя 2024
  • НаукаНаука

Комментарии • 13

  • @rkinamdar
    @rkinamdar 7 лет назад +1

    Thanks for excellent deep explaination! Really helpful!!

  • @anmolagarwal8373
    @anmolagarwal8373 3 года назад

    You are providing such a amazing content. I have seen EDA and PCA playlist from your channel and I loved the way you explained. Thank you so much for the content . Apart from it I have a question , Can we see those 200 features which is covering more than 90% information?

    • @AppliedAICourse
      @AppliedAICourse  3 года назад +1

      You can obtain the top information features by using the top eigenvectors as each top eigenvector corresponds to the direction of high variance.

  • @kondurusrikanth7620
    @kondurusrikanth7620 4 года назад

    Super video

  • @shyamss2338
    @shyamss2338 3 года назад

    In the plt.plot(), we only mentioned cum_var_ecplained as a parameter. How did we get n_components as the x-axis?

  • @SushilKumar113
    @SushilKumar113 4 года назад

    Thank you so much for the PCA playlist video

  • @shyamss2338
    @shyamss2338 3 года назад

    Isn't it required for var_explained and cum_var_explaimed to be in a for loop so that the loop goes on from i=0 to i= 783?

  • @manikantabandla3923
    @manikantabandla3923 3 года назад

    The playlist was amazing.
    I have the following question right at the beginning of the introduction of the variance concept.
    How the maintaining of explaining variance very high is related to preserving of original data?
    Can someone help me with this?
    At least try to provide me the relevant blog links.

    • @AppliedAICourse
      @AppliedAICourse  3 года назад +1

      Variance is also a measure of information content in a dataset. As an extreme case, imagine a dataset where all the points have the same value. Such a dataset has very little information content in it and a variance of zero. More the spread in the data points, more the information content in the data set. That’s why in PCA, we try to preserve as much variance as possible while projecting to lower dimensions.

    • @shyamss2338
      @shyamss2338 3 года назад

      @@AppliedAICourse Even if I have same data points in my dataset, why is it considered to have little information content? Isn't all this still data?

  • @manikantabandla3923
    @manikantabandla3923 3 года назад +1

    Can we get back to the original data from reduced data?

    • @AppliedAICourse
      @AppliedAICourse  3 года назад

      You cannot recover the original data if you do not have all the eigenvalues and eigenvectors.

    • @shyamss2338
      @shyamss2338 3 года назад

      @@AppliedAICourse So this is why we intially store the data in a new variable before reducing the dimensionality?