LANTERN: Scalable Distillation of Large Language Models for Job-Person Fit and Explanation

๐Ÿ“… 2025-10-06
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
In recruitment scenarios, large language models (LLMs) suffer from poor domain adaptability, unstructured outputs, high inference latency, and difficulties in online deployment for jobโ€“candidate matching and explanation generation. Method: We propose a scalable knowledge distillation framework tailored to recruitment, featuring a multi-objective dual-branch architecture that integrates data-level and logit-level distillation. Our approach combines an encoder-decoder structure, post-training optimization, and prompt engineering to efficiently transfer knowledge from a black-box teacher model. Contribution/Results: Experiments demonstrate that our method preserves evaluation accuracy while substantially improving inference efficiency. Online A/B testing shows a 0.24% increase in applicant conversion rate and a 0.28% rise in qualified applications. The framework provides a reusable technical pathway for lightweight LLM deployment in vertical domains.

Technology Category

Application Category

๐Ÿ“ Abstract
Large language models (LLMs) have achieved strong performance across a wide range of natural language processing tasks. However, deploying LLMs at scale for domain specific applications, such as job-person fit and explanation in job seeking platforms, introduces distinct challenges. At LinkedIn, the job person fit task requires analyzing a candidate's public profile against job requirements to produce both a fit assessment and a detailed explanation. Directly applying open source or finetuned LLMs to this task often fails to yield high quality, actionable feedback due to the complexity of the domain and the need for structured outputs. Moreover, the large size of these models leads to high inference latency and limits scalability, making them unsuitable for online use. To address these challenges, we introduce LANTERN, a novel LLM knowledge distillation framework tailored specifically for job person fit tasks. LANTERN involves modeling over multiple objectives, an encoder model for classification purpose, and a decoder model for explanation purpose. To better distill the knowledge from a strong black box teacher model to multiple downstream models, LANTERN incorporates multi level knowledge distillation that integrates both data and logit level insights. In addition to introducing the knowledge distillation framework, we share our insights on post training techniques and prompt engineering, both of which are crucial for successfully adapting LLMs to domain specific downstream tasks. Extensive experimental results demonstrate that LANTERN significantly improves task specific metrics for both job person fit and explanation. Online evaluations further confirm its effectiveness, showing measurable gains in job seeker engagement, including a 0.24% increase in apply rate and a 0.28% increase in qualified applications.
Problem

Research questions and friction points this paper is trying to address.

Deploying LLMs for job-person fit assessment faces domain complexity challenges
Large model size causes high inference latency limiting online scalability
Structured outputs for both classification and explanation require specialized distillation
Innovation

Methods, ideas, or system contributions that make the work stand out.

LANTERN distills LLMs for job-person fit tasks
It uses multi-level knowledge distillation from teacher models
Combines encoder for classification and decoder for explanation
๐Ÿ”Ž Similar Papers
No similar papers found.
Z
Zhoutong Fu
LinkedIn, Mountain View, CA, USA
Yihan Cao
Yihan Cao
LinkedIn
Y
Yi-Lin Chen
LinkedIn, Mountain View, CA, USA
A
Aman Lunia
LinkedIn, Mountain View, CA, USA
Liming Dong
Liming Dong
CSIRO Data61
Software EngineeringSoftware TraceabilityData QualityDevOpsAgentOps
N
Neha Saraf
LinkedIn, Mountain View, CA, USA
R
Ruijie Jiang
LinkedIn, Mountain View, CA, USA
Yun Dai
Yun Dai
OpenAI
deep learningLLMML systemsdistributed training
Qingquan Song
Qingquan Song
LinkedIn
LLMEfficient MLAutoMLTensor AnalysisRecommender Systems
T
Tan Wang
LinkedIn, Mountain View, CA, USA
G
Guoyao Li
LinkedIn, Mountain View, CA, USA
D
Derek Koh
LinkedIn, Mountain View, CA, USA
H
Haichao Wei
LinkedIn, Mountain View, CA, USA
Z
Zhipeng Wang
LinkedIn, Mountain View, CA, USA
A
Aman Gupta
LinkedIn, Mountain View, CA, USA
C
Chengming Jiang
LinkedIn, Mountain View, CA, USA
Jianqiang Shen
Jianqiang Shen
Palo Alto Research Center (PARC)
Artificial IntelligenceMachine LearningIntelligent SystemsSpeech RecognitionNatural Language Processing
L
Liangjie Hong
LinkedIn, Mountain View, CA, USA
W
Wenjing Zhang
LinkedIn, Mountain View, CA, USA