A.X K1 Technical Report

📅 2026-01-14
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work proposes A.X K1, a 519-billion-parameter mixture-of-experts (MoE) language model trained from scratch under constrained computational budgets, designed to simultaneously enhance multilingual—particularly Korean—reasoning capabilities and inference efficiency. Leveraging a 10-trillion-token corpus, multi-stage data curation, scaling-law-informed training configurations, and an innovative Think-Fusion training strategy, the model enables users to explicitly control reasoning mode switching. Experimental results demonstrate that A.X K1 achieves state-of-the-art performance among open-source models across multiple benchmarks, significantly outperforming existing approaches—especially on Korean-language tasks—while maintaining high inference efficiency and deployment flexibility.

Technology Category

Application Category

📝 Abstract
We introduce A.X K1, a 519B-parameter Mixture-of-Experts (MoE) language model trained from scratch. Our design leverages scaling laws to optimize training configurations and vocabulary size under fixed computational budgets. A.X K1 is pre-trained on a corpus of approximately 10T tokens, curated by a multi-stage data processing pipeline. Designed to bridge the gap between reasoning capability and inference efficiency, A.X K1 supports explicitly controllable reasoning to facilitate scalable deployment across diverse real-world scenarios. We propose a simple yet effective Think-Fusion training recipe, enabling user-controlled switching between thinking and non-thinking modes within a single unified model. Extensive evaluations demonstrate that A.X K1 achieves performance competitive with leading open-source models, while establishing a distinctive advantage in Korean-language benchmarks.
Problem

Research questions and friction points this paper is trying to address.

Mixture-of-Experts
reasoning capability
inference efficiency
controllable reasoning
large language model
Innovation

Methods, ideas, or system contributions that make the work stand out.

Mixture-of-Experts
controllable reasoning
Think-Fusion training
scaling laws
efficient inference
🔎 Similar Papers
No similar papers found.
Sung Jun Cheon
Sung Jun Cheon
SK Telecom
generative modelslarge language modelsspeech synthesismolecular generation
J
Jaekyung Cho
SK Telecom
Seongho Choi
Seongho Choi
Ph. D Student at Seoul National University
Video QA
Hyunjun Eun
Hyunjun Eun
SK Telecom
Deep LearningComputer Vision
S
Seokhwan Jo
SK Telecom
Jaehyun Jun
Jaehyun Jun
SK Telecom
Deep LearningRepresentation LearningMultimodalityVisual Reasoning
Minsoo Kang
Minsoo Kang
SK Telecom
Machine LearningComputer Vision
J
Jin Kim
SK Telecom
Jiwon Kim
Jiwon Kim
SK Telecom
Deep LearningArtificial IntelligenceMachine Learning
Minsang Kim
Minsang Kim
Korea university
Machine LearningDeep LearningNLPLLMFoundation Models
S
Sungwan Kim
SK Telecom
S
Seungsik Kim
SK Telecom
T
Tae Yoon Kim
SK Telecom
Y
Youngran Kim
SK Telecom
H
Hyeon-Seung Lee
SK Telecom
Sangyeol Lee
Sangyeol Lee
Seoul National University
Financial Time SeriesRisk managementChange point analysis & Statististical Process ControlPredictive Analytics
S
Sungeun Lee
SK Telecom
Y
Youngsoon Lee
SK Telecom
Y
Yujin Lee
SK Telecom
S
Seongmin Ok
SK Telecom
C
Chanyong Park
SK Telecom
H
Hyewoong Park
SK Telecom
J
Junyoung Park
SK Telecom
H
Hyunho Yang
SK Telecom
S
Subin Yi
SK Telecom
S
Soohyun Bae
SK Telecom
D
Dhammiko Arya
SK Telecom
Yongseok Choi
Yongseok Choi
T-Brain, SK Telecom
Machine learningVideo compressionLow-power embedded systems
S
Sangho Choi
SK Telecom
D
Dongyeon Cho
SK Telecom
S
Seung-Mo Cho
SK Telecom
G
Gyoungeun Han
SK Telecom
Y
Yong-jin Han
SK Telecom
S
Seokyoung Hong
SK Telecom
Hyeon Hwang
Hyeon Hwang
Korea University
Natural Language Processing
W
Wonbeom Jang
SK Telecom
Minjeong Ju
Minjeong Ju
Korea Advanced Institute of Science and Technology (KAIST)
Machine LearningDeep LearningArtificial IntelligenceComputer Vision
W
Wonjin Jung
SK Telecom
Keummin Ka
Keummin Ka
Yonsei University
S
Sungil Kang
SK Telecom
D
Dongnam Kim
SK Telecom
J
Joonghoon Kim
SK Telecom
Jonghwi Kim
Jonghwi Kim
Research Intern, AI Model Lab, SK Telecom
Information RetrievalCross-Language Information RetrievalRankingCode SwitchingRAG
S
SaeRom Kim
SK Telecom
S
Sangjin Kim
SK Telecom
S
S. Kim
SK Telecom
Y
Youngjin Kim
SK Telecom
S
Seojin Lee
SK Telecom
S
Sunwoo Lee
SK Telecom
Taehoon Lee
Taehoon Lee
Project Leader @ AI R&D Center, SK Telecom
AI RoboticsMachine LearningComputer Vision
C
Chanwoo Park
SK Telecom
S
Sohee Park
SK Telecom
S
Sooyeon Park
SK Telecom
Y
Yohan Ra
SK Telecom
S
Sereimony Sek
SK Telecom
S
Seungyeon Seo
SK Telecom
G
Gun Song
SK Telecom
S
Sanghoon Woo
SK Telecom
J
J. Yoon
SK Telecom
S
Sungbin Yoon
SK Telecom