13.3.1 L1-regularized Logistic Regression as Embedded Feature Selection (L13: Feature Selection)

Поделиться
HTML-код
  • Опубликовано: 22 дек 2024

Комментарии • 5

  • @amrelsayeh4446
    @amrelsayeh4446 5 месяцев назад

    @sebastian At 13:20, why is the solution between the global minimum and the penalty minimum lie somewhere where one of the weights is zero. In other words, why it should lie at the corner of the penalty function not just at the line. between the global minimum and the penalty minimum.

  • @AyushSharma-jm6ki
    @AyushSharma-jm6ki Год назад

    @sebastian amazing video. Thanks for sharing. I am getting deeper understanding of these topics with your videos.

  • @arunthiru6729
    @arunthiru6729 2 года назад +3

    @sebastian I think using Logistic Regression directly for feature selection based on respective weights/coefficients means we are assuming all dimensions/features are independent. I understand this is not the correct way to do this. Pls advise.

    • @SebastianRaschka
      @SebastianRaschka  2 года назад +1

      Yes, this assumption is correct. ML is full of trade-offs 😅. If you cannot make this assumption, I recommend the sequential feature selection approach

    • @wayne7936
      @wayne7936 7 месяцев назад

      Thanks for pointing this out!