🤖 AI Summary
To address the enterprise deployment bottleneck of high-quality Arabic large language models—stemming from scarce, culturally grounded training data—this paper proposes a culture-aware synthetic data augmentation and iterative preference alignment framework. Methodologically: (1) we introduce the first synthetic data generation paradigm explicitly incorporating Arabic cultural cognition, augmented with human-in-the-loop annotation to produce high-fidelity training corpora; (2) we design a multi-stage post-training pipeline tailored to enterprise requirements, integrating supervised fine-tuning, DPO-driven iterative preference optimization, instruction alignment, and cultural sensitivity distillation. Experimental results demonstrate that our open-sourced 7B-parameter model significantly outperforms same-scale baselines across critical dimensions—including Arabic cultural understanding, instruction following, RAG response quality, and context fidelity—achieving production-ready performance for enterprise applications.
📝 Abstract
Building high-quality large language models (LLMs) for enterprise Arabic applications remains challenging due to the limited availability of digitized Arabic data. In this work, we present a data synthesis and refinement strategy to help address this problem, namely, by leveraging synthetic data generation and human-in-the-loop annotation to expand our Arabic training corpus. We further present our iterative post training recipe that is essential to achieving state-of-the-art performance in aligning the model with human preferences, a critical aspect to enterprise use cases. The culmination of this effort is the release of a small, 7B, open-weight model that outperforms similarly sized peers in head-to-head comparisons and on Arabic-focused benchmarks covering cultural knowledge, instruction following, RAG, and contextual faithfulness.