[edit]
Combatting Language Forgetting in Low-Resourced Settings
DLI 2025 Research Track, PMLR 302:1-11, 2026.
Abstract
Neural machine translation becomes a continual learning challenge as language evolves over time. While Transformer-based models excel at capturing linguistic patterns from large corpora, they require continual updates to adapt to new language use without losing previously acquired knowledge. In this work, we introduce Latent Replay Buffers to the NLP domain for the first time by implementing and fine-tuning our Latent Replay Transformer. We conduct initial experiments for low-resource languages on Small-100, a distilled version of a multilingual transformer trained on 100 languages, to be well-suited for deployment in memory- and data-constrained environments. Our findings reveal an intriguing trade-off in the selection of latent activations to store for effective replay. We release our code to support both the Continual Learning and NLP for Low-Resourced Languages communities. Keywords: Continual Learning, Transformers, Low-Resourced Languages.