Ticktack : Long Span Temporal Alignment of Large Language Models Leveraging Sexagenary Cycle Time Expression

📅 2025-03-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) suffer from temporal misalignment in millennium-scale reasoning, primarily due to sparse long-range temporal signals in training data, leading to inaccurate time representations and catastrophic forgetting. To address this, we propose a cyclical time modeling framework inspired by the traditional Chinese sexagenary cycle (Gānzhī), which maps Gregorian years onto a 60-year periodic sequence. We introduce polar-coordinate spatiotemporal encoding, enhanced positional encoding, and post-training representation alignment to jointly refine temporal semantics. This work is the first to systematically integrate the sexagenary cycle into LLM temporal representation learning. Evaluated on a newly constructed long-horizon temporal reasoning benchmark, our method achieves a +18.7% improvement in temporal reasoning accuracy, substantially mitigating chronological confusion and knowledge discontinuity across millennia. The approach provides an interpretable, generalizable, and structurally grounded solution for long-term temporal understanding in foundation models.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) suffer from temporal misalignment issues especially across long span of time. The issue arises from knowing that LLMs are trained on large amounts of data where temporal information is rather sparse over long times, such as thousands of years, resulting in insufficient learning or catastrophic forgetting by the LLMs. This paper proposes a methodology named"Ticktack"for addressing the LLM's long-time span misalignment in a yearly setting. Specifically, we first propose to utilize the sexagenary year expression instead of the Gregorian year expression employed by LLMs, achieving a more uniform distribution in yearly granularity. Then, we employ polar coordinates to model the sexagenary cycle of 60 terms and the year order within each term, with additional temporal encoding to ensure LLMs understand them. Finally, we present a temporal representational alignment approach for post-training LLMs that effectively distinguishes time points with relevant knowledge, hence improving performance on time-related tasks, particularly over a long period. We also create a long time span benchmark for evaluation. Experimental results prove the effectiveness of our proposal.
Problem

Research questions and friction points this paper is trying to address.

Addresses temporal misalignment in large language models over long spans.
Proposes using sexagenary cycle for uniform yearly temporal distribution.
Introduces temporal alignment method to improve long-term task performance.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Utilizes sexagenary cycle for uniform time distribution
Employs polar coordinates for temporal modeling
Introduces temporal alignment for improved LLM performance
🔎 Similar Papers
No similar papers found.
Xue Han
Xue Han
Professor of Biomedical Engineering, Boston University
NeuroengineeringNeuroscience
Q
Qian Hu
JIUTIAN Team, China Mobile Research Institute, Beijing, China
Yitong Wang
Yitong Wang
ByteDance Inc.
computer vision
W
Wenchun Gao
JIUTIAN Team, China Mobile Research Institute, Beijing, China
L
Lianlian Zhang
JIUTIAN Team, China Mobile Research Institute, Beijing, China
Q
Qing Wang
JIUTIAN Team, China Mobile Research Institute, Beijing, China
L
Lijun Mei
JIUTIAN Team, China Mobile Research Institute, Beijing, China
C
Chao Deng
JIUTIAN Team, China Mobile Research Institute, Beijing, China
Junlan Feng
Junlan Feng
Chief Scientist at China Mobile Research
Natural LanguageMachine LearningSpeech ProcessingData Mining