๐ค AI Summary
To address the high data migration overhead and underutilized SSD computational capability in storage-disaggregated architectures, this paper proposes a firmware-level lightweight containerized in-storage processing (ISP) framework. Our approach introduces two novel mechanisms: Ethernet-over-NVMe communication and a Virtual Firmware architecture, enabling native OS-level virtualization container support (e.g., Docker) directly within SSD firmwareโthereby achieving secure and efficient storage-side compute offloading. By co-designing NVMe protocol extensions and storage-compute scheduling, our framework delivers a 2.0ร speedup for I/O-intensive workloads and accelerates distributed large language model inference by 7.9ร, while significantly reducing host CPU and memory overhead. This work establishes a new hardware-level co-design paradigm for disaggregated storage systems and AI workloads.
๐ Abstract
ISP minimizes data transfer for analytics but faces challenges in adaptation and disaggregation. We propose DockerSSD, an ISP model leveraging OS-level virtualization and lightweight firmware to enable containerized data processing directly on SSDs. Key features include Ethernet over NVMe for network-based ISP management and Virtual Firmware for secure, efficient container execution. DockerSSD supports disaggregated storage pools, reducing host overhead and enhancing large-scale services like LLM inference. It achieves up to 2.0x better performance for I/O-intensive workloads, and 7.9x improvement in distributed LLM inference.