🤖 AI Summary
Large language models (LLMs) struggle to align with non-differentiable cultural dimensions—such as Hofstede’s cultural indices—in cross-cultural research, due to the absence of differentiable gradients for cultural measurement.
Method: This paper proposes a zero-shot, parameter-free soft prompt tuning framework that integrates differential evolution (DE)—a gradient-free black-box optimization algorithm—with soft prompting. Without labeled data or model parameter updates, it directly optimizes prompts against fixed LLaMA-3-8B-Instruct weights to align outputs with multi-regional cultural factors.
Contribution/Results: Experiments demonstrate statistically significant improvements over supervised fine-tuning and RLHF baselines in cultural consistency evaluation. The alignment is interpretable—via prompt-culture mappings—and transferable across domains, effectively bridging the gap between computational models and human cultural cognition.
📝 Abstract
Large Language Model (LLM) alignment conventionally relies on supervised fine-tuning or reinforcement learning based alignment frameworks. These methods typically require labeled or preference datasets and involve updating model weights to align the LLM with the training objective or reward model. Meanwhile, in social sciences such as cross-cultural studies, factor analysis is widely used to uncover underlying dimensions or latent variables that explain observed patterns in survey data. The non-differentiable nature of these measurements deriving from survey data renders the former alignment methods infeasible for alignment with cultural dimensions. To overcome this, we propose a parameter efficient strategy that combines soft prompt tuning, which freezes the model parameters while modifying the input prompt embeddings, with Differential Evolution (DE), a black-box optimization method for cases where a differentiable objective is unattainable. This strategy ensures alignment consistency without the need for preference data or model parameter updates, significantly enhancing efficiency and mitigating overfitting. Our method demonstrates significant improvements in LLama-3-8B-Instruct's cultural dimensions across multiple regions, outperforming both the Naive LLM and the In-context Learning (ICL) baseline, and effectively bridges computational models with human cultural nuances.