[edit]
Bridging the Language Gap: Fine-Tuning Llama for Machine Translation in Low-Resource African Languages
Proceedings of the AI for African Languages Conference 2025, PMLR 314:37-40, 2026.
Abstract
We adapt a pretrained large language model to support Kikuyu, a low-resource African language. A dataset of 140,000 English-Swahili-Kikuyu sentence pairs was collected across multiple domains, with a 30,000 sentence English-Kikuyu subset used for training. After preprocessing and normalization, the Llama 3.2 (3B) model was fine-tuned using parameter-efficient techniques. The resulting system achieves a BLEU score of 25.21, demonstrating the effectiveness of transfer learning for low-resource machine translation.