Clo-HDnn: A 4.66 TFLOPS/W and 3.78 TOPS/W Continual On-Device Learning Accelerator with Energy-efficient Hyperdimensional Computing via Progressive Search

📅 2025-07-23
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the low energy efficiency and high computational overhead in edge-device continual learning, this work proposes a domain-specific accelerator architecture based on hyperdimensional computing (HDC). Methodologically: (i) it introduces a gradient-free continual learning mechanism that eliminates backpropagation; (ii) it incorporates a low-cost Kronecker-product-based hypervector encoder coupled with weight re-clustering for feature extraction; and (iii) it supports dual-mode operation and progressive search, encoding only a subset of query hypervectors. Experimental results demonstrate that the accelerator achieves 4.66 TFLOPS/W and 3.78 TOPS/W energy efficiency on feature extraction and classification tasks, respectively—outperforming state-of-the-art accelerators by 7.77× and 4.85×. Computational complexity is reduced by up to 61%. This is the first systematic application of HDC to hardware acceleration for edge continual learning, delivering significant improvements in energy efficiency and real-time performance while preserving accuracy.

Technology Category

Application Category

📝 Abstract
Clo-HDnn is an on-device learning (ODL) accelerator designed for emerging continual learning (CL) tasks. Clo-HDnn integrates hyperdimensional computing (HDC) along with low-cost Kronecker HD Encoder and weight clustering feature extraction (WCFE) to optimize accuracy and efficiency. Clo-HDnn adopts gradient-free CL to efficiently update and store the learned knowledge in the form of class hypervectors. Its dual-mode operation enables bypassing costly feature extraction for simpler datasets, while progressive search reduces complexity by up to 61% by encoding and comparing only partial query hypervectors. Achieving 4.66 TFLOPS/W (FE) and 3.78 TOPS/W (classifier), Clo-HDnn delivers 7.77x and 4.85x higher energy efficiency compared to SOTA ODL accelerators.
Problem

Research questions and friction points this paper is trying to address.

Designs an on-device continual learning accelerator
Optimizes accuracy and efficiency with hyperdimensional computing
Achieves high energy efficiency for learning tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hyperdimensional computing for efficient learning
Gradient-free continual learning for knowledge storage
Progressive search reduces complexity by 61%
🔎 Similar Papers
No similar papers found.
C
Chang Eun Song
University of California San Diego, La Jolla, CA, USA
W
Weihong Xu
University of California San Diego, La Jolla, CA, USA
K
Keming Fan
University of California San Diego, La Jolla, CA, USA
S
Soumil Jain
University of California San Diego, La Jolla, CA, USA
G
Gopabandhu Hota
University of California San Diego, La Jolla, CA, USA
H
Haichao Yang
University of California San Diego, La Jolla, CA, USA
L
Leo Liu
Taiwan Semiconductor Manufacturing Company (TSMC)
K
Kerem Akarvardar
Taiwan Semiconductor Manufacturing Company (TSMC)
Meng-Fan Chang
Meng-Fan Chang
National Tsing Hua University, Taiwan
Memory Circuit DesignsCompute-in-MemoryAI ChipsMemristor-Based Neuromorphic Circuitsetc.
C
Carlos H. Diaz
Taiwan Semiconductor Manufacturing Company (TSMC)
Gert Cauwenberghs
Gert Cauwenberghs
Professor of Bioengineering and Co-Director, Institute for Neural Computation, UC San Diego
Neuromorphic engineeringlearning systemsneural interfacesbioinstrumentation
Tajana Rosing
Tajana Rosing
Distinguished Professor, UCSD
computer architecturecyber-physical systemssystem energy efficiency
M
Mingu Kang
University of California San Diego, La Jolla, CA, USA