🤖 AI Summary
To address computational unreliability in analog in-memory computing (IMC) systems caused by stuck-at faults (SAFs) and the high compilation overhead and poor scalability of existing fault-tolerant compilation algorithms (e.g., Fault-Free), this work proposes a row-column hybrid grouping multi-bit weight mapping method. We introduce a novel bidirectional redundancy architecture to enhance robustness against SAFs and formulate fault-tolerant weight decomposition as an integer linear programming (ILP) problem, solved efficiently via theoretical optimization. Evaluated on convolutional neural networks and small language models, our approach achieves up to an 8-percentage-point accuracy improvement, 150× faster compilation, and 2× higher energy efficiency compared to state-of-the-art baselines. This work significantly advances the reliability, deployment efficiency, and energy-delay-product of analog IMC accelerators.
📝 Abstract
This paper addresses two critical challenges in analog In-Memory Computing (IMC) systems that limit their scalability and deployability: the computational unreliability caused by stuck-at faults (SAFs) and the high compilation overhead of existing fault-mitigation algorithms, namely Fault-Free (FF). To overcome these limitations, we first propose a novel multi-bit weight representation technique, termed row-column hybrid grouping, which generalizes conventional column grouping by introducing redundancy across both rows and columns. This structural redundancy enhances fault tolerance and can be effectively combined with existing fault-mitigation solutions. Second, we design a compiler pipeline that reformulates the fault-aware weight decomposition problem as an Integer Linear Programming (ILP) task, enabling fast and scalable compilation through off-the-shelf solvers. Further acceleration is achieved through theoretical insights that identify fault patterns amenable to trivial solutions, significantly reducing computation. Experimental results on convolutional networks and small language models demonstrate the effectiveness of our approach, achieving up to 8%p improvement in accuracy, 150x faster compilation, and 2x energy efficiency gain compared to existing baselines.