Charts-of-Thought: Enhancing LLM Visualization Literacy Through Structured Data Extraction

📅 2025-08-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) exhibit limited visualization literacy, hindering reliable chart interpretation. Method: We propose Charts-of-Thought (CoT), a structured prompting framework that guides LLMs through sequential steps—data extraction, cross-validation, and logical reasoning—to systematically enhance multimodal chart understanding. CoT is the first approach to embed interpretable, stepwise reasoning into multimodal visual inference. Contribution/Results: Evaluated using the Visualization Literacy Assessment Test (VLAT) on Claude-3.7, GPT-4.5, and Gemini-2.0, CoT achieves a VLAT score of 50.17 for Claude-3.7-sonnet—significantly surpassing the human baseline (28.82). Across models, average performance improves by 15.2%, with 100% accuracy attained on several chart types. This work establishes a reproducible, generalizable, structured analytical paradigm for visual understanding tasks.

Technology Category

Application Category

📝 Abstract
This paper evaluates the visualization literacy of modern Large Language Models (LLMs) and introduces a novel prompting technique called Charts-of-Thought. We tested three state-of-the-art LLMs (Claude-3.7-sonnet, GPT-4.5 preview, and Gemini-2.0-pro) on the Visualization Literacy Assessment Test (VLAT) using standard prompts and our structured approach. The Charts-of-Thought method guides LLMs through a systematic data extraction, verification, and analysis process before answering visualization questions. Our results show Claude-3.7-sonnet achieved a score of 50.17 using this method, far exceeding the human baseline of 28.82. This approach improved performance across all models, with score increases of 21.8% for GPT-4.5, 9.4% for Gemini-2.0, and 13.5% for Claude-3.7 compared to standard prompting. The performance gains were consistent across original and modified VLAT charts, with Claude correctly answering 100% of questions for several chart types that previously challenged LLMs. Our study reveals that modern multimodal LLMs can surpass human performance on visualization literacy tasks when given the proper analytical framework. These findings establish a new benchmark for LLM visualization literacy and demonstrate the importance of structured prompting strategies for complex visual interpretation tasks. Beyond improving LLM visualization literacy, Charts-of-Thought could also enhance the accessibility of visualizations, potentially benefiting individuals with visual impairments or lower visualization literacy.
Problem

Research questions and friction points this paper is trying to address.

Evaluating LLM visualization literacy using VLAT
Introducing Charts-of-Thought for structured data extraction
Improving LLM performance on visual interpretation tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Structured prompting for data extraction
Systematic verification and analysis process
Enhancing LLM visualization literacy
🔎 Similar Papers
No similar papers found.
A
Amit Kumar Das
Computer Science Department, Stony Brook University, USA
M
Mohammad Tarun
East West University, Bangladesh
Klaus Mueller
Klaus Mueller
Professor of Computer Science, Stony Brook University
VisualizationVisual AnalyticsData ScienceExplainable AIMedical Imaging