Should I shuffle samples with cross-validation?

Поделиться
HTML-код
  • Опубликовано: 27 окт 2024

Комментарии • 5

  • @dataschool
    @dataschool  5 месяцев назад

    This is a lesson from my NEW course, "Master Machine Learning with scikit-learn." You can enroll here: courses.dataschool.io/master-machine-learning-with-scikit-learn

  • @aleksandartta
    @aleksandartta 5 месяцев назад +1

    Hello Kevin, thank you very much... I have two questions:
    1) after hyper parameters tunning and cross validation, the final model should be some that is trained on the whole dataset (meaning train + validation set)? Am I right?
    2) do we need cross validation if the dataset is very big (and how to know how big :) ? i.e. when cross validation is not necessary?

    • @dataschool
      @dataschool  5 месяцев назад

      Great questions!
      1. Yes, re-train the tuned model on the entire dataset (meaning all samples for which you know the target value).
      2. Yes, cross-validation is a useful model evaluation procedure with any size dataset, with the possible exception of a very tiny dataset. (Below a certain number of samples, no model evaluation procedure is particularly useful.)
      Hope that helps!

  • @sedighehnadaei1895
    @sedighehnadaei1895 5 месяцев назад +1

    As always you did great.thank you so much ❤

    • @dataschool
      @dataschool  5 месяцев назад

      You are so welcome!