Browse > Article
http://dx.doi.org/10.6109/jkiice.2022.26.1.15

Adaptive Weight Control for Improvement of Catastropic Forgetting in LwF  

Park, Seong-Hyeon (Department of Embedded Systems Engineering, Incheon National University)
Kang, Seok-Hoon (Department of Embedded Systems Engineering, Incheon National University)
Abstract
Among the learning methods for Continuous Learning environments, "Learning without Forgetting" has fixed regularization strengths, which can lead to poor performance in environments where various data are received. We suggest a way to set weights variable by identifying the features of the data we want to learn. We applied weights adaptively using correlation and complexity. Scenarios with various data are used for evaluation and experiments showed accuracy increases by up to 5% in the new task and up to 11% in the previous task. In addition, it was found that the adaptive weight value obtained by the algorithm proposed in this paper, approached the optimal weight value calculated manually by repeated experiments for each experimental scenario. The correlation coefficient value is 0.739, and overall average task accuracy increased. It can be seen that the method of this paper sets an appropriate lambda value every time a new task is learned, and derives the optimal result value in various scenarios.
Keywords
Continuous learning; Catastrophic forgetting; Neural network; Deep learning; Regularization;
Citations & Related Records
Times Cited By KSCI : 1  (Citation Analysis)
연도 인용수 순위
1 I. J. Goodfellow, M. Miraza, D. Xiao, A. Courville, and Y. Bengio, "An empirical investigation of catastrophic forgetting in gradient-based neural networks," arXiv preprint arXiv:1312.6211, 2013.
2 R. M. French, "Catastrophic forgetting in connectionist networks," Trends in cognitive sciences, vol. 3, no. 4, pp. 128-135, Apr. 1999.   DOI
3 G. I. Parisi, R. Kemker, J. L. Part, C. Kanan, and S. Wermter, "Continual lifelong learning with neural networks: A review," Neural Networks, vol. 113, pp. 54-71, 2019.   DOI
4 Z. Li and D. Hoiem, "Learning without forgetting," IEEE transactions on pattern analysis and machine intelligence, vol. 40, no. 12, pp. 2935-2947, Nov. 2017.   DOI
5 Y. Hsu, Y. Liu, A. Ramasamy, and Z. Kira, "Re-evaluating continual learning scenarios: A categorization and case for strong baselines," arXiv preprint, arXiv:1810.12488, 2018.
6 F. Zenke, B. Poole, and S. Ganguli, "Continual learning through synaptic intelligence," in Proceedings of the 34th International Conference on Machine Learning, vol. 70, pp. 3987-3995, 2017.
7 J. Yoon, E. Yang, J. Lee, and S. Hwang, "Lifelong learning with dynamically expandable networks," arXiv preprint arXiv:1708.01547, 2017.
8 H. Shin, J. K. Lee, J. Kim, and J. Kim, "Continual learning with deep generative replay," arXiv preprint arXiv: 1705.08690, 2017.
9 S. H. park and S. H. Kang, "Continual Learning using Data Similarity," Institute of Korean Electrical and Electronics Engineers, vol. 24, no. 2, pp. 514-522, 2020.
10 J. Kirkpatrick, R. Pascanu, N. Rabinowitz, J. Veness, G. Desjardins, A. A. Rusu, K. Milan, J. Quan, T. Ramalho, A. Grabska-Barwinska, D. Hassabis, C. Clopath, D. Kumaran, and R. Hadsell, "Overcoming catastrophic forgetting in neural networks," Proceedings of the national academy of sciences, vol. 114, no. 13, pp. 3521-3526, 2017.   DOI