Skip to content

Latest commit

 

History

History
21 lines (11 loc) · 1.41 KB

2018-nips-dynamic-space-time-scheduling-for-gpu-inference.md

File metadata and controls

21 lines (11 loc) · 1.41 KB

[2018 NIPS] Dynamic Space-Time Scheduling for GPU Inference

Summary

The authors evaluated different multiplexing (time & space) techniques for ML inferences on GPUs and proposed ideas to achieve the best tradeoff across criterias.

Background & Motivation

Almost all cloud inference service providers/frameworks assign each model an exclusive GPU. This, combined with the small batch sizes used in an online setting, results in low hardware utilization. Current approaches that multiplex workloads have different tradeoffs, and there is no single solution that wins on all criteria.

ApproachUtilizationPerformance (throughput/latency)Predictability/Performance Isolation
Exclusive accessPoorGoodGood
Time multiplexing (CUDA context switching)AveragePoorGood
Spatial multiplexingGoodAveragePoor

Design & Implementation

The authors proposed software-level fusion of kernel operators across multiple inference jobs to get the best of all worlds.

Links & References