π€ AI Summary
Large language models (LLMs) exhibit inaccurate reasoning and poor robustness when handling complex, multi-condition data analysis queriesβsuch as multi-constraint retrieval and transformation tasks.
Method: This paper proposes *Tabular Thinking*, a novel reasoning paradigm that employs a pre-instruction-guided, tabular intermediate representation to explicitly structure multi-step reasoning into rows and columns, aligning with human structured cognition.
Contribution/Results: We formally define and empirically validate four levels of data structuralization and their impact on model performance; further, we introduce a multi-level structured evaluation framework. Experiments demonstrate that Tabular Thinking achieves an average relative performance improvement of 40.29% on multi-condition query tasks, significantly enhancing generalization and stability across diverse requests, constraints, and application scenarios.
π Abstract
Despite the recent advancement of Large Langauge Models (LLMs), they struggle with complex queries often involving multiple conditions, common in real-world scenarios. We propose Thinking with Tables, a technique that assists LLMs to leverage tables for intermediate thinking aligning with human cognitive behavior. By introducing a pre-instruction that triggers an LLM to organize information in tables, our approach achieves a 40.29% average relative performance increase, higher robustness, and show generalizability to different requests, conditions, or scenarios. We additionally show the influence of data structuredness for the model by comparing results from four distinct structuring levels that we introduce.