Why T-distributed Stochastic Neighbor Embedding (TSNE) is great for visualization? Math Step By Step

Поделиться
HTML-код
  • Опубликовано: 4 ноя 2024

Комментарии • 5

  • @satyashah3045
    @satyashah3045 3 года назад +1

    Why to keep perplexity value small(low) for dense region?

    • @machinelearningmastery
      @machinelearningmastery  3 года назад +1

      Perplexity intuitively means we are trying to guess the number of neighbors around a point. This is why we keep it low for dense regions since there is a large cluster of points around each point.

    • @timrh3020
      @timrh3020 2 года назад

      @@machinelearningmastery how do we know if regions are dense, a priori?

  • @quonxinquonyi8570
    @quonxinquonyi8570 3 года назад +1

    I have a question why not represent the low dimension probability ‘ q’ with gaussian rather than using t distribution

    • @machinelearningmastery
      @machinelearningmastery  3 года назад

      Good you asked. In my video, I cover the gaussian formula and compare it back to back with the t-distribution. The literature calls the gaussian version of formulation as SNE algorithm. In practice, SNE's embeddings lag behind TSNE in providing clarity when you have large dimensions and some non-linearity. Have a look at the video's formulations for SNE's q version as it is compared to TSNE. Happy to answer should you have any other questions.