🤖 AI Summary
Large language models (LLMs) struggle to simultaneously achieve stylistic adaptation (e.g., brand tone) and faithful instruction following when instruction-response paired data is scarce.
Method: We propose a two-stage LoRA-based style transfer framework: Stage I disentangles stylistic representations from unpaired, unstructured text; Stage II seamlessly integrates the learned style module into a pretrained instruction-tuned model—without requiring instruction-format annotations or compromising task capability.
Contribution/Results: Our approach eliminates interference between style and task learning, enabling lightweight, controllable, and reusable style adaptation without paired corpora—the first such decoupled style-task adaptation method. Experiments across multiple open-source LLMs and benchmarks demonstrate significant improvements in style consistency (validated by both automated metrics and human evaluation), while fully preserving original instruction-following performance.
📝 Abstract
Adapting LLMs to specific stylistic characteristics, like brand voice or authorial tones, is crucial for enterprise communication but challenging to achieve from corpora which lacks instruction-response formatting without compromising instruction adherence. We introduce StyleAdaptedLM, a framework that efficiently transfers stylistic traits to instruction-following models using Low-Rank Adaptation (LoRA). LoRA adapters are first trained on a base model with diverse unstructured stylistic corpora, then merged with a separate instruction-following model. This enables robust stylistic customization without paired data or sacrificing task performance. Experiments across multiple datasets and models demonstrate improved stylistic consistency while preserving instruction adherence, with human evaluations confirming brand-specific convention uptake. StyleAdaptedLM offers an efficient path for stylistic personalization in LLMs.