Words That Unite The World: A Unified Framework for Deciphering Central Bank Communications Globally

📅 2025-05-15
🏛️ arXiv.org
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
Central bank policy communication often carries ambiguous implicit meanings, and misinterpretations can exacerbate socioeconomic inequality. To address this, we propose the first globally harmonized analytical framework for central bank communication. We introduce the World Central Bank (WCB) dataset—comprising 380,000 sentences from 25 central banks over 28 years—and formalize three core tasks: stance detection, temporal classification, and uncertainty estimation. Methodologically, we pioneer a cross-central-bank joint modeling paradigm (“the whole exceeds the sum of its parts”) and design a high-fidelity semantic annotation protocol featuring dual-dimensional labeling and expert validation. Through systematic evaluation across seven pretrained language models and nine large language models (15,075 benchmark runs), we rigorously assess model generalizability and economic utility. All data, models, and code are publicly released under CC-BY-NC-SA 4.0 on Hugging Face and GitHub.

Technology Category

Application Category

📝 Abstract
Central banks around the world play a crucial role in maintaining economic stability. Deciphering policy implications in their communications is essential, especially as misinterpretations can disproportionately impact vulnerable populations. To address this, we introduce the World Central Banks (WCB) dataset, the most comprehensive monetary policy corpus to date, comprising over 380k sentences from 25 central banks across diverse geographic regions, spanning 28 years of historical data. After uniformly sampling 1k sentences per bank (25k total) across all available years, we annotate and review each sentence using dual annotators, disagreement resolutions, and secondary expert reviews. We define three tasks: Stance Detection, Temporal Classification, and Uncertainty Estimation, with each sentence annotated for all three. We benchmark seven Pretrained Language Models (PLMs) and nine Large Language Models (LLMs) (Zero-Shot, Few-Shot, and with annotation guide) on these tasks, running 15,075 benchmarking experiments. We find that a model trained on aggregated data across banks significantly surpasses a model trained on an individual bank's data, confirming the principle"the whole is greater than the sum of its parts."Additionally, rigorous human evaluations, error analyses, and predictive tasks validate our framework's economic utility. Our artifacts are accessible through the HuggingFace and GitHub under the CC-BY-NC-SA 4.0 license.
Problem

Research questions and friction points this paper is trying to address.

Deciphering policy implications in central bank communications globally
Addressing misinterpretations that disproportionately impact vulnerable populations
Benchmarking language models for stance detection and uncertainty estimation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Created comprehensive global central bank communications dataset
Benchmarked pretrained and large language models extensively
Demonstrated cross-bank training outperforms individual bank models
🔎 Similar Papers
No similar papers found.
Agam Shah
Agam Shah
PhD Candidate, Georgia Institute of Technology
Natural Language ProcessingFinanceData ScienceComputational Science
Siddhant Sukhani
Siddhant Sukhani
Graduate Researcher, Institute of Computational Mathematical Engineering, Stanford University
NLPOptimizationApplied MathematicsArtificial Intelligence
Huzaifa Pardawala
Huzaifa Pardawala
Georgia Institute of Technology
Machine LearningNatural Language ProcessingFinanceOptimizationReward-Based Agents
S
Saketh Budideti
Georgia Institute of Technology
R
Riya Bhadani
Georgia Institute of Technology
R
Rudra Gopal
Georgia Institute of Technology
S
Siddhartha Somani
Georgia Institute of Technology
R
Rutwik Routu
Georgia Institute of Technology
Michael Galarnyk
Michael Galarnyk
Georgia Institute of Technology
Machine LearningGenerative AIFinance
S
Soungmin Lee
Georgia Institute of Technology
A
Arnav Hiray
Georgia Institute of Technology
A
Akshar Ravichandran
Georgia Institute of Technology
E
Eric Kim
Georgia Institute of Technology
P
P. Aluru
Georgia Institute of Technology
J
Joshua Zhang
Georgia Institute of Technology
S
Sebastian Jaskowski
Georgia Institute of Technology
V
Veer Guda
Georgia Institute of Technology
M
Meghaj Tarte
Georgia Institute of Technology
L
Liqin Ye
Georgia Institute of Technology
S
S. Gosden
Georgia Institute of Technology
R
Rutwik Routu
Georgia Institute of Technology
R
Rachel Yuh
Georgia Institute of Technology
S
S. Chava
Georgia Institute of Technology
S
Sahasra Chava
Georgia Institute of Technology
D
Dylan Patrick Kelly
Georgia Institute of Technology
A
Aiden Chiang
Georgia Institute of Technology
H
Harsit Mittal
Georgia Institute of Technology
S
S. Chava
Georgia Institute of Technology