🤖 AI Summary
To address the prohibitive memory and computational overhead—as well as poor scalability—arising from transmission matrix computation in large-scale photonic neural network (PNN) training, this paper proposes an iterative slicing-based lightweight unitary matrix computation framework. The method innovatively decomposes large-scale unitary matrices into small, iteratively updated sub-blocks, enabling end-to-end backpropagation while ensuring linear memory growth and efficient gradient computation. By integrating explicit unitary constraint modeling with parameterized photonic circuit simulation, the framework enables, for the first time, full-stack training of PNNs comprising thousands of optical units. Experiments on MNIST, Digits, and Olivetti Faces demonstrate a 3.2–8.7× speedup in training and a reduction in memory consumption to one-fifth of prior approaches, thereby substantially overcoming key scalability bottlenecks in PNN training.
📝 Abstract
PNNs present promising opportunities for accelerating machine learning by leveraging the unique benefits of photonic circuits. However, current state of the art PNN simulation tools face significant scalability challenges when training large-scale PNNs, due to the computational demands of transfer matrix calculations, resulting in high memory and time consumption. To overcome these limitations, we introduce the Slicing method, an efficient transfer matrix computation approach compatible with back-propagation. We integrate this method into LuxIA, a unified simulation and training framework. The Slicing method substantially reduces memory usage and execution time, enabling scalable simulation and training of large PNNs. Experimental evaluations across various photonic architectures and standard datasets, including MNIST, Digits, and Olivetti Faces, show that LuxIA consistently surpasses existing tools in speed and scalability. Our results advance the state of the art in PNN simulation, making it feasible to explore and optimize larger, more complex architectures. By addressing key computational bottlenecks, LuxIA facilitates broader adoption and accelerates innovation in AI hardware through photonic technologies. This work paves the way for more efficient and scalable photonic neural network research and development.