🤖 AI Summary
Natural language-to-sign language (NL2SL) translation suffers from severe low-resource constraints—specifically, the absence of large-scale, standardized, and domain-balanced parallel corpora—leading to poor generalization in existing models. To address this, we propose AulSign, the first framework integrating dynamic prompt engineering, in-context learning, and symbolic gesture association. Without relying on built-in sign language knowledge, AulSign models gesture semantics solely through natural language descriptions and introduces a sample selection mechanism to enhance large language models’ implicit understanding of sign language structure. This establishes a scalable, multilingual NL2SL translation paradigm. On SignBank+ and LaCAM CNR-ISTC benchmarks, AulSign significantly outperforms prior state-of-the-art methods, demonstrating high accuracy and strong generalization under low-resource conditions. Our approach opens a new pathway toward accessible human–machine interaction.
📝 Abstract
Translating natural languages into sign languages is a highly complex and underexplored task. Despite growing interest in accessibility and inclusivity, the development of robust translation systems remains hindered by the limited availability of parallel corpora which align natural language with sign language data. Existing methods often struggle to generalize in these data-scarce environments, as the few datasets available are typically domain-specific, lack standardization, or fail to capture the full linguistic richness of sign languages. To address this limitation, we propose Advanced Use of LLMs for Sign Language Translation (AulSign), a novel method that leverages Large Language Models via dynamic prompting and in-context learning with sample selection and subsequent sign association. Despite their impressive abilities in processing text, LLMs lack intrinsic knowledge of sign languages; therefore, they are unable to natively perform this kind of translation. To overcome this limitation, we associate the signs with compact descriptions in natural language and instruct the model to use them. We evaluate our method on both English and Italian languages using SignBank+, a recognized benchmark in the field, as well as the Italian LaCAM CNR-ISTC dataset. We demonstrate superior performance compared to state-of-the-art models in low-data scenario. Our findings demonstrate the effectiveness of AulSign, with the potential to enhance accessibility and inclusivity in communication technologies for underrepresented linguistic communities.