🤖 AI Summary
Time-series data often lack descriptive textual annotations, hindering the training of text generation models. Method: This paper introduces a novel forward-and-backward dual-path pairing construction framework, pioneering a backward generation paradigm: starting from human-crafted rule-based textual descriptions, it synthesizes corresponding time-series data in reverse. Based on this paradigm, we construct TACO—the first cross-domain time-series–text paired dataset. We further propose a contrastive learning–based text generation framework that jointly models abstract time-series features and incorporates backward rule guidance. Results: Experiments demonstrate zero-shot textual generation on unseen domains, significantly improving cross-domain generalization and generated text quality. Moreover, the approach enhances interpretability of time-series understanding and boosts human–machine interaction efficiency.
📝 Abstract
Due to scarcity of time-series data annotated with descriptive texts, training a model to generate descriptive texts for time-series data is challenging. In this study, we propose a method to systematically generate domain-independent descriptive texts from time-series data. We identify two distinct approaches for creating pairs of time-series data and descriptive texts: the forward approach and the backward approach. By implementing the novel backward approach, we create the Temporal Automated Captions for Observations (TACO) dataset. Experimental results demonstrate that a contrastive learning based model trained using the TACO dataset is capable of generating descriptive texts for time-series data in novel domains.