๐ค AI Summary
Existing GPU-accelerated singular value decomposition (SVD) algorithms struggle to balance portability across heterogeneous hardware platforms and efficiency across mixed-precision regimes. Method: This paper proposes a unified, QR-iteration-based GPU-accelerated SVD implementation. It is the first to support both Apple Metal GPUs and FP16 half-precision arithmetic within a single kernel, leveraging Juliaโs multiple dispatch and metaprogramming capabilities alongside GPUArrays/KernelAbstractions to construct a hardware-agnostic parallel abstraction layer. Results: For matrices larger than 1024ร1024, the implementation outperforms MAGMA and SLATE; on large-scale matrices, it achieves 80โ90% of cuSOLVERโs performance. Crucially, it significantly reduces dependence on the CUDA/NVIDIA ecosystem, delivering high-efficiency, cross-platform SVD infrastructure for scientific computing and low-rank adaptation in large language models.
๐ Abstract
This paper presents a portable, GPU-accelerated implementation of a QR-based singular value computation algorithm in Julia. The singular value ecomposition (SVD) is a fundamental numerical tool in scientific computing and machine learning, providing optimal low-rank matrix approximations. Its importance has increased even more in large-scale machine learning pipelines, including large language models (LLMs), where it enables low-rank adaptation (LoRA). The implemented algorithm is based on the classic two-stage QR reduction, consisting of successive matrix reduction to band form and bidiagonal form. Our implementation leverages Julia's multiple dispatch and metaprogramming capabilities, integrating with the GPUArrays and KernelAbstractions frameworks to provide a unified type and hardware-agnostic function. It supports diverse GPU architectures and data types, and is, to our knowledge, the first GPU-accelerated singular value implementation to support Apple Metal GPUs and half precision. Performance results on multiple GPU backends and data types demonstrate that portability does not require sacrificing performance: the unified function outperforms most linear algebra libraries (MAGMA, SLATE, rocSOLVER, oneMKL) for matrix sizes larger than 1024x1024, and achieves 80%-90% of the performance of cuSOLVER for large matrices.