🤖 AI Summary
With SRAM/DRAM costs plateauing, memory has become the primary bottleneck for system cost and energy efficiency.
Method: This paper proposes a memory specialization paradigm that departs from conventional hierarchical memory architectures. It introduces two application-tailored memory classes: Long-term RAM (LtRAM) for long-lived, read-intensive data, and Short-term RAM (StRAM) for short-lived, high-frequency transient accesses. This approach requires explicit OS-level management to enable non-hierarchical memory resource scheduling. Leveraging emerging memory technologies, the paper designs hardware architectures, system interfaces, and integration mechanisms for LtRAM and StRAM.
Contribution/Results: The work identifies key technical challenges—including coherence, migration, and interface standardization—and outlines concrete implementation pathways. By rethinking memory as a heterogeneous, application-aware resource rather than a monolithic hierarchy, it establishes a foundational architectural direction for efficient, scalable computing systems.
📝 Abstract
Both SRAM and DRAM have stopped scaling: there is no technical roadmap to reduce their cost (per byte/GB). As a result, memory now dominates system cost. This paper argues for a paradigm shift from today's simple memory hierarchy toward specialized memory architectures that exploit application-specific access patterns. Rather than relying solely on traditional off-chip DRAM and on-chip SRAM, we envisage memory systems equipped with additional types of memory whose performance trade-offs benefit workloads through non-hierarchical optimization. We propose two new memory classes deserving explicit OS support: long-term RAM (LtRAM) optimized for read-intensive data with long lifetimes, and short-term RAM (StRAM) designed for transient, frequently-accessed data with short lifetimes. We explore underlying device technologies that could implement these classes, including their evolution and their potential integration into current system designs given emerging workload requirements. We identify critical research challenges to realize what we believe is a necessary evolution toward more efficient and scalable computing systems capable of meeting future demands.