🤖 AI Summary
AI-driven streaming inference workflows face challenges due to dynamically shifting data access patterns under event-driven execution, rendering conventional caching and scheduling techniques ineffective for latency reduction. To address this, we propose an application-aware data-computation correlation modeling mechanism, introducing the novel abstraction of “correlation grouping.” This enables developers to explicitly declare dynamic data dependencies via lightweight runtime annotations—without modifying core logic. Leveraging this abstraction, we design correlation-driven cross-node cooperative scheduling, streaming workflow orchestration, and cluster-wide data-affinity scheduling policies. Evaluated on real-world latency-sensitive ML applications, our approach achieves a 58% reduction in end-to-end latency, a 73% decrease in latency variance, and a 41% improvement in node utilization—all with minimal code changes.
📝 Abstract
ML inference workflows often require low latency and high throughput, yet we lack good options for addressing this need. Techniques that reduce latency in other streaming settings (such as caching and optimization-driven scheduling) are of limited value because ML data dependencies are often very large and can change dramatically depending on the triggering event. In this work, we propose a novel correlation grouping mechanism that makes it easier for developers to express application-specific data access correlations, enabling coordinated management of data objects in server clusters hosting streaming inference tasks. Experiments based on a latency-sensitive ML-based application confirm the limitations of standard techniques while showing that our solution yields dramatically better performance. The proposed mechanism is able to maintain significantly lower and more consistent latency, achieves higher node utilization as workload and scale-out increase, and yet requires only minor changes to the code implementing the application.