EETnet: a CNN for Gaze Detection and Tracking for Smart-Eyewear

📅 2025-06-30
🏛️ IEEE International Joint Conference on Neural Network
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing event-camera-based eye-tracking methods rely on GPU acceleration, making them unsuitable for resource-constrained embedded devices such as smart glasses. Method: We propose a lightweight, fully event-driven convolutional neural network architecture optimized for microcontrollers. It comprises two low-complexity models—grid-based classifier and pixel-level regressor—and supports end-to-end training, evaluation, and INT8 quantization. The system processes raw event streams directly, eliminating image reconstruction and achieving microsecond-level inference latency with significantly reduced power consumption. Contribution/Results: Evaluated on public event-based datasets, our approach matches the accuracy of state-of-the-art GPU-dependent methods. It is the first fully event-driven eye-tracking system capable of real-time operation on microcontroller units (MCUs), successfully deployed on STM32H7-series platforms. This work establishes a viable pathway for edge intelligence in wearable eye-tracking applications.

Technology Category

Application Category

📝 Abstract
Event-based cameras are becoming a popular solution for efficient, low-power eye tracking. Due to the sparse and asynchronous nature of event data, they require less processing power and offer latencies in the microsecond range. However, many existing solutions are limited to validation on powerful GPUs, with no deployment on real embedded devices. In this paper, we present EETnet, a convolutional neural network designed for eye tracking using purely event-based data, capable of running on microcontrollers with limited resources. Additionally, we outline a methodology to train, evaluate, and quantize the network using a public dataset. Finally, we propose two versions of the architecture: a classification model that detects the pupil on a grid superimposed on the original image, and a regression model that operates at the pixel level.
Problem

Research questions and friction points this paper is trying to address.

Develops efficient eye tracking for resource-constrained smart eyewear devices
Addresses limitations of existing GPU-dependent event-based gaze tracking solutions
Creates microcontroller-compatible CNN models for pupil detection and localization
Innovation

Methods, ideas, or system contributions that make the work stand out.

CNN for eye tracking using event-based data
Runs on microcontrollers with limited resources
Includes classification and regression model versions
🔎 Similar Papers
No similar papers found.
A
Andrea Aspesi
Department of Electronics, Information and Bioengineering (DEIB), Politecnico di Milano - Via Ponzio 34/5, 20133 Milan, Italy
Andrea Simpsi
Andrea Simpsi
Politecnico di Milano
Artificial IntelligenceComputer Vision
A
Aaron Tognoli
Department of Electronics, Information and Bioengineering (DEIB), Politecnico di Milano - Via Ponzio 34/5, 20133 Milan, Italy
Simone Mentasti
Simone Mentasti
Politecnico di Milano
Autonomous Drivingsensor fusionimage segmentationdeep learning
L
Luca Merigo
EssilorLuxottica Italia S.p.A. - Piazzale Cadorna 3, 20123 Milan, Italy
Matteo Matteucci
Matteo Matteucci
Full Professor, Department of Electronics Information and Bioengineering, Politecnico di Milano
RoboticsMachine LearningComputer VisionPattern Recognition