Continual Learning and Catastrophic Forgetting

Поделиться
HTML-код
  • Опубликовано: 7 окт 2024
  • A lecture that discusses continual learning and catastrophic forgetting in deep neural networks. We discuss the context, methods for evaluating algorithms, and algorithms based on regularization, dynamic architectures, and Complementary Learning Systems. Specifically, we discuss data permutation tasks, incremental task learning, multimodal learning, the Learning without Forgetting algorithm, Elastic Weight Consolidation, Progressive Neural Networks, and Generative replay.
    This lecture is from Northeastern University's CS 7150 Summer 2020 class on Deep Learning, taught by Paul Hand.
    The notes are available at: khoury.northeas...
    References:
    Parisi et al. 2019:
    Parisi, German I., Ronald Kemker, Jose L. Part, Christopher Kanan, and Stefan Wermter. "Continual lifelong learning with neural networks: A review." Neural Networks (2019).
    Chen and Liu 2018:
    Chen, Zhiyuan, and Bing Liu. "Lifelong machine learning." Synthesis Lectures on Artificial Intelligence and Machine Learning 12, no. 3 (2018): 1-207. Chapter 4.
    Kemker et al. 2017:
    Kemker, Ronald, Marc McClure, Angelina Abitino, Tyler L. Hayes, and Christopher Kanan. "Measuring catastrophic forgetting in neural networks." In Thirty-second AAAI conference on artificial intelligence. 2018.
    van de Ven and Tolias 2019:
    van de Ven, Gido M., and Andreas S. Tolias. "Three scenarios for continual learning." arXiv preprint arXiv:1904.07734 (2019).
    Li and Hoiem 2018:
    Li, Zhizhong, and Derek Hoiem. "Learning without forgetting." IEEE transactions on pattern analysis and machine intelligence 40, no. 12 (2017): 2935-2947.
    Kirkpatrick et al. 2017:
    Kirkpatrick, James, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A. Rusu, Kieran Milan et al. "Overcoming catastrophic forgetting in neural networks." Proceedings of the national academy of sciences 114, no. 13 (2017): 3521-3526.
    Rusu et al. 2016:
    Rusu, Andrei A., Neil C. Rabinowitz, Guillaume Desjardins, Hubert Soyer, James Kirkpatrick, Koray Kavukcuoglu, Razvan Pascanu, and Raia Hadsell. "Progressive neural networks." arXiv preprint arXiv:1606.04671 (2016).
    Shin et al. 2017:
    Shin, Hanul, Jung Kwon Lee, Jaehong Kim, and Jiwon Kim. "Continual learning with deep generative replay." In Advances in Neural Information Processing Systems, pp. 2990-2999. 2017.

Комментарии • 18

  • @weinansun9321
    @weinansun9321 4 года назад +10

    this is a gem on the internet...more people should know about these videos!

  • @prof.laurenzwiskott
    @prof.laurenzwiskott 4 года назад +8

    Very nice lecture giving a good overview. Thanks a lot.
    One remark: There are actually regions in the brain that continuously generate new neurons, in particular the detate gyrus in the hippocampus, and we have actually built a model for avoiding catastrophic interference on that, see Wiskott, L., Rasch, M. & Kempermann, G. A functional hypothesis for adult hippocampal neurogenesis: avoidance of catastrophic interference in the dentate gyrus, Hippocampus, 2006, 16, 329-343.

  • @ahmedbahaaeldin750
    @ahmedbahaaeldin750 4 года назад +6

    this video is genius why isn't it famous ???

    • @lorenzoleongutierrez7927
      @lorenzoleongutierrez7927 3 месяца назад

      Man , here in 2024 and I just discovered this concepts an vid . Amazing !

  • @juandiego2045
    @juandiego2045 3 года назад +1

    Great aproach to the problem, best explanation i´ve found.

  • @3koozy
    @3koozy 3 года назад +1

    Thank you Paul Very Much for this brilliant summary of the "Continual Learning" topic , you saved my day!

  • @boburniyozov62
    @boburniyozov62 3 месяца назад

    indeed very good video. well and easy explained.

  • @niraj5582
    @niraj5582 3 года назад +1

    Excellent content to get a quick overview.

  • @piupunia6373
    @piupunia6373 4 года назад +2

    this video was really informative and well described.

  • @silviasanmartindeporres7033
    @silviasanmartindeporres7033 2 года назад

    Brilliant lecture!

  • @lukepeterson868
    @lukepeterson868 4 года назад +4

    Paul, Do you have a patreon? Your videos are awesome.

  • @PuMpKinSpIcE666
    @PuMpKinSpIcE666 3 года назад +2

    god-tier video

  • @reihanehmirjalili7467
    @reihanehmirjalili7467 5 месяцев назад

    Amazing video!

  • @kil98q
    @kil98q Месяц назад

    Dont know much about this subject. What if you would not need old training data but generate it with the current neural network and the knowledge which objects or things where related. You could dream up the old skill and include the new situation. Nvm.. this has been covered in the video..

  • @abbaskhaliil4717
    @abbaskhaliil4717 3 года назад +2

    Hi dear, nice explanation.. do you have python code for this task?

  • @prasenjitgiri919
    @prasenjitgiri919 3 года назад

    Paul, although a very good explanation but dude c'mon why so low volume!

  • @yeyerrd
    @yeyerrd 2 года назад

    Great talk! Thank you for the video.
    Just two comments regarding typing on ruclips.net/video/vjaq03IYgSk/видео.html:
    1. During initialization, wouldn't Y_o be Y^hat_o? Because that is the output of the network
    2. In the argmin formula isn't Y_o the same as Y_n?