🤖 AI Summary
This work addresses the bottleneck in energy function design for Dense Associative Memory (DenseAM) networks. We propose a novel log-sum-ReLU (LSR) energy function, constructed from the Epanechnikov kernel, which departs from the conventional paradigm relying on exponential separation functions. Our method achieves exponential memory capacity without compromising pattern retrieval accuracy and, for the first time in DenseAM, introduces numerous high-likelihood “emergent” local minima—enabling both perfect memory recall and creative memory generation. Theoretical analysis shows that the non-convex energy landscape exhibits superior separability. Experiments on image data demonstrate that LSR significantly outperforms the log-sum-exp (LSE) baseline in memory capacity and emergent memory novelty, while maintaining comparable log-likelihood performance.
📝 Abstract
We propose a novel energy function for Dense Associative Memory (DenseAM) networks, the log-sum-ReLU (LSR), inspired by optimal kernel density estimation. Unlike the common log-sum-exponential (LSE) function, LSR is based on the Epanechnikov kernel and enables exact memory retrieval with exponential capacity without requiring exponential separation functions. Moreover, it introduces abundant additional emph{emergent} local minima while preserving perfect pattern recovery -- a characteristic previously unseen in DenseAM literature. Empirical results show that LSR energy has significantly more local minima (memories) that have comparable log-likelihood to LSE-based models. Analysis of LSR's emergent memories on image datasets reveals a degree of creativity and novelty, hinting at this method's potential for both large-scale memory storage and generative tasks.