Large Language Models are Good Multi-lingual Learners : When LLMs Meet Cross-lingual Prompts

📅 2024-09-17
🏛️ International Conference on Computational Linguistics
📈 Citations: 4
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) frequently overlook critical constraints during long-context reasoning, leading to rule violation. To address this, we propose MLPrompt—a cross-lingual prompting strategy that automatically translates error-prone rules into multiple linguistic formulations to enhance model attention to key constraints. MLPrompt introduces the novel prompting paradigm of “multilingual translation for reinforced rule awareness,” integrated with an automated verification mechanism, and supports structured generation tasks including mixed-integer programming (MIP) modeling and text-to-SQL. Experiments across multiple public benchmark datasets demonstrate that MLPrompt significantly outperforms Chain-of-Thought, Tree-of-Thought, and Self-Consistency baselines. In text-to-MIP and text-to-SQL tasks, it yields more robust and accurate structured outputs. These results empirically validate the efficacy of multilingual prompting in improving rule-aware modeling for complex constrained reasoning.

Technology Category

Application Category

📝 Abstract
With the advent of Large Language Models (LLMs), generating rule-based data for real-world applications has become more accessible. Due to the inherent ambiguity of natural language and the complexity of rule sets, especially in long contexts, LLMs often struggle to follow all specified rules, frequently omitting at least one. To enhance the reasoning and understanding of LLMs on long and complex contexts, we propose a novel prompting strategy Multi-Lingual Prompt, namely MLPrompt, which automatically translates the error-prone rule that an LLM struggles to follow into another language, thus drawing greater attention to it. Experimental results on public datasets across various tasks have shown MLPrompt can outperform state-of-the-art prompting methods such as Chain of Thought, Tree of Thought, and Self-Consistency. Additionally, we introduce a framework integrating MLPrompt with an auto-checking mechanism for structured data generation, with a specific case study in text-to-MIP instances. Further, we extend the proposed framework for text-to-SQL to demonstrate its generation ability towards structured data synthesis.
Problem

Research questions and friction points this paper is trying to address.

Improves LLM rule-following in complex multilingual contexts
Enhances structured data generation with auto-checking mechanism
Outperforms existing prompting methods in cross-lingual tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

MLPrompt translates rules into another language
Integrates auto-checking for structured data generation
Extends framework for text-to-SQL synthesis
🔎 Similar Papers
No similar papers found.
T
Teng Wang
Department of Mathematics, The University of Hong Kong, Hong Kong SAR, China
Zhenqi He
Zhenqi He
The Hong Kong University of Science and Technology (HKUST) | The University of Hong Kong (HKU)
Open-World LearningComputer VisionMulti-Modal Learning
W
Wing-Yin Yu
Noah’s Ark Lab, Huawei, Hong Kong SAR, China
X
Xiaojin Fu
Noah’s Ark Lab, Huawei, Hong Kong SAR, China
Xiongwei Han
Xiongwei Han
AI&OR Principal Researcher at Noah's Ark Lab, Huawei
Intelligence ModelingLLMs for OR