A Reliable, Time-Predictable Heterogeneous SoC for AI-Enhanced Mixed-Criticality Edge Applications

📅 2025-02-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Mixed-criticality edge AI applications in robotics, automotive, and aerospace demand stringent temporal predictability and high reliability under shared hardware resources. Method: We propose a 16 nm heterogeneous SoC featuring (i) a novel hardware–software co-designed configurable IP mechanism enabling deterministic interconnect and memory access, and (ii) a highly reliable multi-core AI accelerator coupled with a dual-core vector cluster, supporting strict spatial and temporal isolation between critical and non-critical workloads. Contribution/Results: The AI accelerator achieves 304.9 GOPS at 1.6 TOPS/W; the vector cluster delivers 121.8 GFLOPS at 1.1 TFLOPS/W. The entire SoC operates within ≤1.2 W power envelope while satisfying rigorous worst-case execution time (WCET) constraints—demonstrating simultaneous breakthroughs in energy efficiency and timing determinism for safety-critical edge AI.

Technology Category

Application Category

📝 Abstract
Next-generation mixed-criticality Systems-on-chip (SoCs) for robotics, automotive, and space must execute mixed-criticality AI-enhanced sensor processing and control workloads, ensuring reliable and time-predictable execution of critical tasks sharing resources with non-critical tasks, while also fitting within a sub-2W power envelope. To tackle these multi-dimensional challenges, in this brief, we present a 16nm, reliable, time-predictable heterogeneous SoC with multiple programmable accelerators. Within a 1.2W power envelope, the SoC integrates software-configurable hardware IPs to ensure predictable access to shared resources, such as the on-chip interconnect and memory system, leading to tight upper bounds on execution times of critical applications. To accelerate mixed-precision mission-critical AI, the SoC integrates a reliable multi-core accelerator achieving 304.9 GOPS peak performance at 1.6 TOPS/W energy efficiency. Non-critical, compute-intensive, floating-point workloads are accelerated by a dual-core vector cluster, achieving 121.8 GFLOPS at 1.1 TFLOPS/W and 106.8 GFLOPS/mm2.
Problem

Research questions and friction points this paper is trying to address.

Ensures reliable, time-predictable execution of critical tasks
Integrates software-configurable hardware IPs for predictable resource access
Accelerates mixed-precision AI and floating-point workloads efficiently
Innovation

Methods, ideas, or system contributions that make the work stand out.

16nm heterogeneous SoC design
software-configurable hardware IPs
multi-core AI accelerator
🔎 Similar Papers
No similar papers found.
Angelo Garofalo
Angelo Garofalo
University of Bologna, ETH Zurich
HW efficient Machine LearningHeterogeneous Computing ArchitecturesMixed-Criticality Systems
A
Alessandro Ottaviano
D-ITET department at ETH Zurich, Switzerland
Matteo Perotti
Matteo Perotti
ETH Zürich
Electronic engineeringcomputer architectureRISC-V
Thomas Benz
Thomas Benz
ETH Zurich
Digital DesignsHigh-Performance SoCsMemory Architectures
Yvan Tortorella
Yvan Tortorella
University of Bologna
Digital ASIC DesignHardware AcceleratorsHeterogeneous SoCsSpace ArchitecturesEmbedded Systems
Robert Balas
Robert Balas
D-ITET department at ETH Zurich, Switzerland
Michael Rogenmoser
Michael Rogenmoser
ETH Zurich
Fault ToleranceMulticore processorsReliable SoCs
C
Chi Zhang
D-ITET department at ETH Zurich, Switzerland
Luca Bertaccini
Luca Bertaccini
ETH Zurich
Digital ASIC DesignHardware AcceleratorsHeterogenous SoCsEmbedded Systems
Nils Wistoff
Nils Wistoff
PhD Student, ETH Zurich
processor designsecure computer architecture
M
Maicol Ciani
Department of Electrical, Electronic, and Information Engineering (DEI), University of Bologna, Italy
C
Cyril Koenig
D-ITET department at ETH Zurich, Switzerland
M
Mattia Sinigaglia
Department of Electrical, Electronic, and Information Engineering (DEI), University of Bologna, Italy
Luca Valente
Luca Valente
Università di Bologna
Digital ElectronicsVLSI
Paul Scheffler
Paul Scheffler
ETH Zurich
Computer ArchitectureDigital Hardware DesignInterconnectsASICs
Manuel Eggimann
Manuel Eggimann
Axelera AI
Low Power Digital Design
Matheus Cavalcante
Matheus Cavalcante
Stanford University
F
Francesco Restuccia
Department of Computer Science and Engineering, University of California in San Diego (UCSD), California, USA
Alessandro Biondi
Alessandro Biondi
Scuola Superiore Sant'Anna
real-time systemsembedded systems
Francesco Conti
Francesco Conti
Associate Professor, University of Bologna
Hardware acceleratorsDeep LearningUltra-Low Power Computing
F
Frank K. Gurkaynak
D-ITET department at ETH Zurich, Switzerland
Davide Rossi
Davide Rossi
Associate Professor, University Of Bologna
VLSI systemsUltra-low-power circuitsmulti core architecturereconfigurable computing
Luca Benini
Luca Benini
ETH Zürich, Università di Bologna
Integrated CircuitsComputer ArchitectureEmbedded SystemsVLSIMachine Learning