Ultra-Efficient On-Device Object Detection on AI-Integrated Smart Glasses with TinyissimoYOLO

📅 2023-11-02
🏛️ ECCV Workshops
📈 Citations: 17
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenge of achieving real-time, on-device object detection with extended battery life on resource-constrained wearable devices such as smart glasses, this work proposes a dual-MCU heterogeneous hardware architecture integrating a milliwatt-level RISC-V parallel processor and a custom vision AI accelerator. We design TinyissimoYOLO—a family of sub-million-parameter YOLO variants (v1.3/v5/v8)—and introduce a novel dynamic power-cycling mechanism coupled with multimodal sensor co-scheduling. A lightweight, MCU-optimized inference engine is developed to support efficient deployment. Evaluation shows an end-to-end latency of 56 ms (18 FPS), per-inference latency of 17 ms, and energy consumption of 1.59 mJ; the system achieves a total power draw of 62.9 mW, enabling 9.3 hours of runtime on a 154 mAh battery. Our solution significantly outperforms state-of-the-art embedded frameworks (e.g., MCUNet) and marks the first demonstration of全天候, high-frame-rate, full-stack low-power visual perception on microcontrollers.
📝 Abstract
Smart glasses are rapidly gaining advanced functionality thanks to cutting-edge computing technologies, accelerated hardware architectures, and tiny AI algorithms. Integrating AI into smart glasses featuring a small form factor and limited battery capacity is still challenging when targeting full-day usage for a satisfactory user experience. This paper illustrates the design and implementation of tiny machine-learning algorithms exploiting novel low-power processors to enable prolonged continuous operation in smart glasses. We explore the energy- and latency-efficient of smart glasses in the case of real-time object detection. To this goal, we designed a smart glasses prototype as a research platform featuring two microcontrollers, including a novel milliwatt-power RISC-V parallel processor with a hardware accelerator for visual AI, and a Bluetooth low-power module for communication. The smart glasses integrate power cycling mechanisms, including image and audio sensing interfaces. Furthermore, we developed a family of novel tiny deep-learning models based on YOLO with sub-million parameters customized for microcontroller-based inference dubbed TinyissimoYOLO v1.3, v5, and v8, aiming at benchmarking object detection with smart glasses for energy and latency. Evaluations on the prototype of the smart glasses demonstrate TinyissimoYOLO's 17ms inference latency and 1.59mJ energy consumption per inference while ensuring acceptable detection accuracy. Further evaluation reveals an end-to-end latency from image capturing to the algorithm's prediction of 56ms or equivalently 18 fps, with a total power consumption of 62.9mW, equivalent to a 9.3 hours of continuous run time on a 154mAh battery. These results outperform MCUNet (TinyNAS+TinyEngine), which runs a simpler task (image classification) at just 7.3 fps per second.
Problem

Research questions and friction points this paper is trying to address.

Achieving ultra-efficient on-device object detection for smart glasses
Overcoming limited battery capacity and small form factor constraints
Enabling real-time continuous operation with low energy consumption
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses GAP9 multi-core RISC-V processor for edge computing
Implements sub-million parameter TinyissimoYOLO neural networks
Achieves 18 FPS with 62.9mW power consumption
🔎 Similar Papers
2021-07-01IEEE International Conference on Application-Specific Systems, Architectures, and ProcessorsCitations: 13