Transferable Post-training via Inverse Value Learning

📅 2024-10-28
🏛️ North American Chapter of the Association for Computational Linguistics
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
The rapid growth of post-training data and model scale has led to prohibitive computational costs and deployment challenges. Method: We propose Inverse Value Learning (IVL), a framework that models post-training behavior in the logits space and trains lightweight, transferable value networks—enabling capability enhancement without fine-tuning. IVL formalizes post-training as a value-function learning problem, designed for cross-model transferability across parameter scales, pretraining stages, and vocabulary families. It comprises four key components: inverse modeling at the logits layer, multi-model connectivity adaptation, few-shot demonstration-based training, and overfitting-suppressing regularization. Contribution/Results: Experiments demonstrate high fidelity in intra-family model transfer, effective cross-vocabulary generalization, and competitive performance—achieving over 98% of full-parameter fine-tuning accuracy on several tasks—while reducing computational cost by two orders of magnitude.

Technology Category

Application Category

📝 Abstract
As post-training processes utilize increasingly large datasets and base models continue to grow in size, the computational demands and implementation challenges of existing algorithms are escalating significantly. In this paper, we propose modeling the changes at the logits level during post-training using a separate neural network (i.e., the value network). After training this network on a small base model using demonstrations, this network can be seamlessly integrated with other pre-trained models during inference, enables them to achieve similar capability enhancements. We systematically investigate the best practices for this paradigm in terms of pre-training weights and connection schemes. We demonstrate that the resulting value network has broad transferability across pre-trained models of different parameter sizes within the same family, models undergoing continuous pre-training within the same family, and models with different vocabularies across families. In certain cases, it can achieve performance comparable to full-parameter fine-tuning. Furthermore, we explore methods to enhance the transferability of the value model and prevent overfitting to the base model used during training.
Problem

Research questions and friction points this paper is trying to address.

Reducing computational demands in large-scale post-training processes
Enhancing model capabilities via transferable value networks
Improving transferability across diverse pre-trained model architectures
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses value network for logits-level changes
Transfers enhancements across pre-trained models
Improves transferability and prevents overfitting
🔎 Similar Papers
No similar papers found.
X
Xinyu Lu
Chinese Information Processing Laboratory, Institute of Software, Chinese Academy of Sciences; University of Chinese Academy of Sciences
Xueru Wen
Xueru Wen
School of Computer Science and Technology, University of Chinese Academy of Sciences
Natural Language ProcessingAlignmentLarge Language Model
Yaojie Lu
Yaojie Lu
Institute of Software, Chinese Academy of Sciences
Information ExtractionLarge Language Models
Bowen Yu
Bowen Yu
Qwen Team, Alibaba Group
Post-trainingFoundation Model
H
Hongyu Lin
Chinese Information Processing Laboratory, Institute of Software, Chinese Academy of Sciences
H
Haiyang Yu
Alibaba Group
Le Sun
Le Sun
Institute of Software, CAS
information_retrievalnatural_language_processing
X
Xianpei Han
Chinese Information Processing Laboratory, Institute of Software, Chinese Academy of Sciences
Y
Yongbin Li
Alibaba Group