EXAONE Deep: Reasoning Enhanced Language Models

📅 2025-03-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of enhancing mathematical reasoning and code generation capabilities in small-parameter language models. We propose Long-CoT, a novel training paradigm centered on high-order reasoning, which constructs extended chain-of-thought data and jointly optimizes model performance via supervised fine-tuning on large-scale, high-quality reasoning corpora. This approach significantly improves logical deduction capacity without inflating model size. Evaluated on benchmarks for mathematical reasoning and program synthesis, EXAONE Deep—available in 2.4B and 7.8B parameter variants—outperforms mainstream models of comparable scale; its 32B variant matches top-tier open-source large language models. Key contributions include: (1) the first reasoning-specific, long-chain-of-thought data-driven training framework, balancing efficiency and effectiveness; and (2) full open release of all model weights under a commercially permissible license, enabling broad adoption in both academic research and industrial applications.

Technology Category

Application Category

📝 Abstract
We present EXAONE Deep series, which exhibits superior capabilities in various reasoning tasks, including math and coding benchmarks. We train our models mainly on the reasoning-specialized dataset that incorporates long streams of thought processes. Evaluation results show that our smaller models, EXAONE Deep 2.4B and 7.8B, outperform other models of comparable size, while the largest model, EXAONE Deep 32B, demonstrates competitive performance against leading open-weight models. All EXAONE Deep models are openly available for research purposes and can be downloaded from https://huggingface.co/LGAI-EXAONE
Problem

Research questions and friction points this paper is trying to address.

Enhance language models for superior reasoning tasks
Outperform comparable models in math and coding benchmarks
Provide openly available models for research purposes
Innovation

Methods, ideas, or system contributions that make the work stand out.

Specialized dataset for reasoning tasks
Smaller models outperform comparable sizes
Openly available for research purposes
🔎 Similar Papers
No similar papers found.
L
LG AI Research
Kyunghoon Bae
Kyunghoon Bae
LG AI Research
Generative AIComputer VisionNatural Language ProcessingContinual LearningExplainable AI
Eunbi Choi
Eunbi Choi
LG AI Research
K
Kibong Choi
LG AI Research
Stanley Jungkyu Choi
Stanley Jungkyu Choi
LG AI Research
AINatural Language ProcessingSpeech RecognitionVision
Y
Yemuk Choi
LG AI Research
Seokhee Hong
Seokhee Hong
LG AI Research
Natural Language Processing
J
Junwon Hwang
LG AI Research
H
Hyojin Jeon
LG AI Research
K
Kijeong Jeon
LG AI Research
G
Gerrard Jeongwon Jo
LG AI Research
H
Hyunjik Jo
LG AI Research
J
Jiyeon Jung
LG AI Research
H
Hyosang Kim
LG AI Research
Joonkee Kim
Joonkee Kim
LG AI Research
Language ModelingReinforcement Learning
Seonghwan Kim
Seonghwan Kim
Dept. Chemistry, KAIST
machine learningchemical reactionrepresentation learning
Soyeon Kim
Soyeon Kim
Korea Advanced Institute of Science and Technology
Responsible AIML FairnessDifferential PrivacyLLM Hallucination
Sunkyoung Kim
Sunkyoung Kim
LG AI Research
Large Language ModelCross-lingual TransferDomain AdaptationQuestion Answering
Yireun Kim
Yireun Kim
LG AI Research
Deep LearningLLMNLPDatabase
Yongil Kim
Yongil Kim
Seoul National University
Dialog SystemMulti-modal learning
Y
Youchul Kim
LG AI Research
E
Edward Hwayoung Lee
LG AI Research
Haeju Lee
Haeju Lee
KAIST, LG AI Research
Honglak Lee
Honglak Lee
LG AI Research / U. Michigan
Machine LearningDeep LearningReinforcement LearningComputer VisionArtificial Intelligence
Jinsik Lee
Jinsik Lee
LG AI Research
Natural Language Processing
K
Kyungmin Lee
LG AI Research
Sangha Park
Sangha Park
Seoul National University
machine learningdeep learningAI safety and reliability
Y
Yongmin Park
LG AI Research
S
Sihoon Yang
LG AI Research
H
Heuiyeen Yeen
LG AI Research
S
Sihyuk Yi
LG AI Research
Hyeongu Yun
Hyeongu Yun
LG AI Research
Large Language Models