🤖 AI Summary
Optimizing ML model kernels for heterogeneous hardware (x86/ARM/RISC-V/GPUs) remains challenging due to poor portability, high manual effort, and limitations of existing automated approaches—which rely heavily on hardware-specific heuristics and opaque intermediate representations. Method: This paper introduces PerfDojo, a framework that synergistically integrates large language models (LLMs) and reinforcement learning (RL) to automatically search for high-performance kernels within a mathematically formalized, semantically transparent code representation space—requiring no hardware-specific priors or low-level architectural knowledge. It natively supports modern model features including sparsity and quantization. Contribution/Results: Experiments across diverse architectures demonstrate that PerfDojo consistently outperforms state-of-the-art libraries (e.g., OpenBLAS, cuBLAS, TVM), generating kernels that achieve superior performance, cross-platform portability, and human-interpretable source code.
📝 Abstract
The increasing complexity of machine learning models and the proliferation of diverse hardware architectures (CPUs, GPUs, accelerators) make achieving optimal performance a significant challenge. Heterogeneity in instruction sets, specialized kernel requirements for different data types and model features (e.g., sparsity, quantization), and architecture-specific optimizations complicate performance tuning. Manual optimization is resource-intensive, while existing automatic approaches often rely on complex hardware-specific heuristics and uninterpretable intermediate representations, hindering performance portability. We introduce PerfLLM, a novel automatic optimization methodology leveraging Large Language Models (LLMs) and Reinforcement Learning (RL). Central to this is PerfDojo, an environment framing optimization as an RL game using a human-readable, mathematically-inspired code representation that guarantees semantic validity through transformations. This allows effective optimization without prior hardware knowledge, facilitating both human analysis and RL agent training. We demonstrate PerfLLM's ability to achieve significant performance gains across diverse CPU (x86, Arm, RISC-V) and GPU architectures.