Sequential Task Tuning: Overcoming Catastrophic Forgetting with Memory-Augmented Language Models

Authors

  • Dhruba Kumar School of Computer Studies, University of the Philippines Baguio, Philippines

Abstract

Sequential task tuning in language models often faces the challenge of catastrophic forgetting, where performance on previously learned tasks deteriorates as new tasks are introduced. This paper explores the use of memory-augmented language models to mitigate this issue. By integrating external memory components that store and retrieve knowledge from previous tasks, we propose a framework that enhances the model’s ability to retain and recall information across a sequence of tasks. Experimental results demonstrate significant improvements in task retention and overall model performance, highlighting the potential of memory-augmented architectures in overcoming catastrophic forgetting.

Downloads

Published

2024-08-11

How to Cite

Kumar, D. (2024). Sequential Task Tuning: Overcoming Catastrophic Forgetting with Memory-Augmented Language Models. MZ Journal of Artificial Intelligence, 1(2). Retrieved from http://mzjournal.com/index.php/MZJAI/article/view/236