Persian-Phi: Efficient Cross-Lingual Adaptation of Compact LLMs via Curriculum Learning

📅 2025-12-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Low-resource language LLM training faces bottlenecks in high computational cost and poor scalability. To address this, we propose a lightweight adaptation paradigm for Persian: leveraging the Phi-3 Mini architecture, we design a synergistic training framework comprising (i) bilingual Tiny Stories “warm-up” embedding alignment, (ii) curriculum-driven monolingual-to-multilingual capability transfer, and (iii) parameter-efficient fine-tuning (PEFT) integrated with progressive instruction tuning. This approach substantially reduces GPU memory consumption and computational overhead while achieving competitive performance—matching that of significantly larger models—on the Hugging Face Open Persian LLM Leaderboard. All models and code are publicly released, enabling reproducible, low-barrier adoption. Our work provides a scalable, resource-efficient technical pathway toward equitable AI access for low-resource languages.

Technology Category

Application Category

📝 Abstract
The democratization of AI is currently hindered by the immense computational costs required to train Large Language Models (LLMs) for low-resource languages. This paper presents Persian-Phi, a 3.8B parameter model that challenges the assumption that robust multilingual capabilities require massive model sizes or multilingual baselines. We demonstrate how Microsoft Phi-3 Mini -- originally a monolingual English model -- can be effectively adapted to Persian through a novel, resource-efficient curriculum learning pipeline. Our approach employs a unique "warm-up" stage using bilingual narratives (Tiny Stories) to align embeddings prior to heavy training, followed by continual pretraining and instruction tuning via Parameter-Efficient Fine-Tuning (PEFT). Despite its compact size, Persian-Phi achieves competitive results on Open Persian LLM Leaderboard in HuggingFace. Our findings provide a validated, scalable framework for extending the reach of state-of-the-art LLMs to underrepresented languages with minimal hardware resources. The Persian-Phi model is publicly available at https://huggingface.co/amirakhlaghiqqq/PersianPhi.
Problem

Research questions and friction points this paper is trying to address.

Adapting monolingual English LLMs to low-resource languages efficiently
Reducing computational costs for multilingual AI in underrepresented languages
Enabling competitive performance with compact models via curriculum learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Curriculum learning pipeline for cross-lingual adaptation
Warm-up stage with bilingual narratives to align embeddings
Parameter-Efficient Fine-Tuning for continual pretraining and instruction tuning
🔎 Similar Papers
No similar papers found.
A
Amir Mohammad Akhlaghi
Shahid Beheshti University, Tehran, Iran
A
Amirhossein Shabani
Shahid Beheshti University, Tehran, Iran
M
Mostafa Abdolmaleki
Shahid Beheshti University, Tehran, Iran
Saeed Reza Kheradpisheh
Saeed Reza Kheradpisheh
Shahid Beheshti University
Spiking Neural NetworksDeep LearningComputational Neuroscience