AI in Yoruba STEM Education for Early Childhood Learning: A Study on Translation Quality and Context
The integration of AI translation models into education is essential for expand-ing access to learning materials in low-resource languages like Yorub´a, spoken by more than 45 million people in Nigeria and beyond. Early childhood education in a child’s native language is crucial to cognitive and academic development, sig- nificantly enhancing learning outcomes. However, assessing the contextual accu- racy and domain-specific relevance of AI-generated translations remains a critical challenge. This study evaluates the translation quality of state-of-the-art multi-lingual models: Llama, Claude, DeepSeek and AfriTeVa for STEM education in Yor
ub´a. We used a curated STEM dataset digitized from textbooks and translated by expert linguists. Next, we evaluated AI-generated translations against human
references using BLEU, CHRF, TER, and AfriCOMET. To ensure a fair com- parison, all evaluation scores were normalized using the Min Max normalization method. Our results reveal significant gaps in AI-generated STEM translations. Although models like DeepSeek and AfriTeVa perform relatively well in lexical accuracy and fluency, they struggle with domain-specific terminology and contextual integrity. Claude and Llama show higher BLEU scores, but do not maintain scientific accuracy and pedagogical relevance. The findings underscore the need to fine-tune AI models specifically for Yorub´a STEM translation to improve contextual understanding and technical accuracy. Future work should focus on developing domain-adapted, culturally aware translation models and enhancing Yor
ub´a STEM datasets to ensure AI tools better support early childhood STEM education in underrepresented languages.