🤖 AI Summary
To address the challenges of scarce Japanese training corpora, limited long-context support, and suboptimal computational efficiency in large language model development for Japanese, this work introduces the PLaMo 2 series. Methodologically, it adopts a hybrid Samba architecture, extends context length to 32K via continual pretraining, proposes a novel synthetic Japanese corpus generation technique to alleviate data scarcity, and employs weight reuse coupled with structured pruning for efficient training—enabling an 8B-parameter model to match the performance of hundred-billion-parameter models. Furthermore, it integrates synthetic instruction tuning, DPO-based alignment optimization, vLLM-accelerated inference, and low-loss quantization. Empirically, PLaMo 2 achieves state-of-the-art results among open-source Japanese LMs across multiple benchmarks, demonstrating superior instruction following, linguistic fluency, and domain knowledge acquisition.
📝 Abstract
In this report, we introduce PLaMo 2, a series of Japanese-focused large language models featuring a hybrid Samba-based architecture that transitions to full attention via continual pre-training to support 32K token contexts. Training leverages extensive synthetic corpora to overcome data scarcity, while computational efficiency is achieved through weight reuse and structured pruning. This efficient pruning methodology produces an 8B model that achieves performance comparable to our previous 100B model. Post-training further refines the models using a pipeline of supervised fine-tuning (SFT) and direct preference optimization (DPO), enhanced by synthetic Japanese instruction data and model merging techniques. Optimized for inference using vLLM and quantization with minimal accuracy loss, the PLaMo 2 models achieve state-of-the-art results on Japanese benchmarks, outperforming similarly-sized open models in instruction-following, language fluency, and Japanese-specific knowledge.