SpellRing: Recognizing Continuous Fingerspelling in American Sign Language using a Ring

📅 2025-02-15
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenges of continuous American Sign Language (ASL) fingerspelling recognition for deaf and hard-of-hearing users, where existing wearable solutions suffer from insufficient accuracy and practicality. We propose SpellRing, a thumb-worn smart ring featuring a novel single-ring multimodal sensing architecture that integrates active acoustic sensing (miniaturized speaker/microphone pair) with a six-axis inertial measurement unit (IMU), enabling joint modeling of static handshapes and dynamic finger motions. A lightweight CTC-based temporal neural network enables real-time, end-to-end recognition. Evaluated on unconstrained continuous ASL, SpellRing achieves 82.45% Top-1 and 92.42% Top-5 word-level accuracy, with a sentence-level word error rate of only 9.9%, robustly supporting a vocabulary of 1,164 words and 100 phrases. This is the first demonstration that a single-point, lightweight wearable can achieve high-accuracy continuous ASL recognition, establishing a portable, unobtrusive paradigm for accessible text input.

Technology Category

Application Category

📝 Abstract
Fingerspelling is a critical part of American Sign Language (ASL) recognition and has become an accessible optional text entry method for Deaf and Hard of Hearing (DHH) individuals. In this paper, we introduce SpellRing, a single smart ring worn on the thumb that recognizes words continuously fingerspelled in ASL. SpellRing uses active acoustic sensing (via a microphone and speaker) and an inertial measurement unit (IMU) to track handshape and movement, which are processed through a deep learning algorithm using Connectionist Temporal Classification (CTC) loss. We evaluated the system with 20 ASL signers (13 fluent and 7 learners), using the MacKenzie-Soukoref Phrase Set of 1,164 words and 100 phrases. Offline evaluation yielded top-1 and top-5 word recognition accuracies of 82.45% (9.67%) and 92.42% (5.70%), respectively. In real-time, the system achieved a word error rate (WER) of 0.099 (0.039) on the phrases. Based on these results, we discuss key lessons and design implications for future minimally obtrusive ASL recognition wearables.
Problem

Research questions and friction points this paper is trying to address.

Continuous fingerspelling recognition in ASL
Smart ring for ASL text entry
Deep learning for handshape and movement tracking
Innovation

Methods, ideas, or system contributions that make the work stand out.

Smart ring for ASL recognition
Active acoustic sensing technology
Deep learning with CTC loss
🔎 Similar Papers
No similar papers found.
H
Hyunchul Lim
Cornell University, Ithaca, New York, USA
N
Nam Anh Dang
Cornell University, Ithaca, New York, USA
D
Dylan Lee
Cornell University, Ithaca, New York, USA
Tianhong Catherine Yu
Tianhong Catherine Yu
PhD Student, Cornell University
J
Jane Lu
Cornell University, Ithaca, New York, USA
Franklin Mingzhe Li
Franklin Mingzhe Li
Carnegie Mellon University
AccessibilityHuman-Computer InteractionUbiquitous ComputingVR/ARHuman-Centered AI
Y
Yiqi Jin
Cornell University, Ithaca, New York, USA
Y
Yan Ma
Stony Brook University, Stony Brook, New York, USA
Xiaojun Bi
Xiaojun Bi
Department of Computer Science, Stony Brook University
Human Computer InteractionMobile User InterfacesText InputHuman Performance Models
F
François Guimbretière
Cornell University, Ithaca, New York, USA
C
Cheng Zhang
Cornell University, Ithaca, New York, USA