Yi-Lightning Technical Report

📅 2024-12-02
🏛️ arXiv.org
📈 Citations: 1
Influential: 1
📄 PDF
🤖 AI Summary
To address the performance–cost imbalance of large language models in Chinese understanding, mathematical reasoning, code generation, and complex prompt handling, this work introduces Yi-Lightning—a high-performance open-source model. Methodologically, it proposes a novel enhanced Mixture-of-Experts (MoE) architecture with dynamic expert routing; designs the RAISE safety framework—comprising Rule-based filtering, Alignment fine-tuning, Iterative refinement, and Safety Evaluation; and integrates synthetic data construction, multi-stage training (pretraining, supervised fine-tuning, RLHF), and optimized KV caching. Contributions include: ranking 6th overall on Chatbot Arena, and 2nd–4th on Chinese, mathematical, coding, and hard-prompt benchmarks; substantial reduction in inference cost; empirical identification of misalignment between static benchmarks and human preferences—spurring evaluation paradigm innovation; and competitive performance against state-of-the-art closed-source models across major academic benchmarks.

Technology Category

Application Category

📝 Abstract
This technical report presents Yi-Lightning, our latest flagship large language model (LLM). It achieves exceptional performance, ranking 6th overall on Chatbot Arena, with particularly strong results (2nd to 4th place) in specialized categories including Chinese, Math, Coding, and Hard Prompts. Yi-Lightning leverages an enhanced Mixture-of-Experts (MoE) architecture, featuring advanced expert segmentation and routing mechanisms coupled with optimized KV-caching techniques. Our development process encompasses comprehensive pre-training, supervised fine-tuning (SFT), and reinforcement learning from human feedback (RLHF), where we devise deliberate strategies for multi-stage training, synthetic data construction, and reward modeling. Furthermore, we implement RAISE (Responsible AI Safety Engine), a four-component framework to address safety issues across pre-training, post-training, and serving phases. Empowered by our scalable super-computing infrastructure, all these innovations substantially reduce training, deployment and inference costs while maintaining high-performance standards. With further evaluations on public academic benchmarks, Yi-Lightning demonstrates competitive performance against top-tier LLMs, while we observe a notable disparity between traditional, static benchmark results and real-world, dynamic human preferences. This observation prompts a critical reassessment of conventional benchmarks' utility in guiding the development of more intelligent and powerful AI systems for practical applications. Yi-Lightning is now available through our developer platform at https://platform.lingyiwanwu.com.
Problem

Research questions and friction points this paper is trying to address.

Super Language Model
Efficiency and Cost-effectiveness
Chinese and Multidisciplinary Proficiency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Super Language Model
Advanced Structural Optimization
Cost-Efficient Training
🔎 Similar Papers
No similar papers found.
0
01.AI Alan Wake
A
Albert Wang
B
Bei Chen
C
C. X. Lv
C
Chao Li
C
Chengen Huang
C
Chenglin Cai
Chujie Zheng
Chujie Zheng
Qwen Team, Alibaba Group
Artifical IntelligenceLarge Language Models
Daniel Cooper
Daniel Cooper
Associate Professor at University of Michigan
Sustainable manufacturinglife cycle engineeringmetal formingindustrial ecologyresource
E
Ethan Dai
F
Fan Zhou
F
Feng Hu
Heng Ji
Heng Ji
Professor of Computer Science, AICE Director, ASKS Director, UIUC, Amazon Scholar
Natural Language ProcessingLarge Language Models
H
Howard Qiu
J
Jiangcheng Zhu
J
Jun Tian
K
Katherine Su
L
Lihuan Zhang
L
Liying Li
M
Ming Song
M
Mou Li
P
Peng Liu
Q
Qichen Hu
S
Shawn Wang
S
Shijun Zhou
Shiyong Li
Shiyong Li
Beijing Institute of Technology
Microwave imaging
T
Tianhang Zhu
Wen Xie
Wen Xie
X
Xiang He
X
Xiaobo Chen
X
Xiaohui Hu
X
Xiaoyi Ren
Xinyao Niu
Xinyao Niu
Unknown affiliation
AIML
Yanpeng Li
Yanpeng Li
Data Scientist Lead, FedEx Dataworks
data sciencenatural language processingbiomedical informaticsclinical informatics
Y
Yongke Zhao
Y
Yongzhen Luo
Y
Yuchi Xu
Y
Yuxuan Sha
Z
Zhaodong Yan
Z
Zhiyuan Liu
Z
Zirui Zhang