Главная
Study mode:
on
1
Intro
2
Generative Models
3
Inference of Generative Language M
4
Serving of Generative Language Mo
5
Problem 1: Request-Level Schedulin
6
Solution 1: Iteration-Level Schedulin
7
Problem 2: Batching
8
Solution 2: Selective Batching
9
Orca System Architecture
10
Scheduling
Description:
Explore a conference talk on Orca, a distributed serving system designed for Transformer-based generative models. Delve into the challenges of serving large-scale language models like GPT-3 and discover innovative solutions such as iteration-level scheduling and selective batching. Learn how these techniques significantly improve latency and throughput compared to existing systems. Gain insights into the architecture and scheduling mechanisms of Orca, which enable efficient processing of multi-iteration workloads for autoregressive token generation. Understand the importance of system support for serving cutting-edge generative AI models and how Orca addresses the limitations of current inference serving systems.

Orca - A Distributed Serving System for Transformer-Based Generative Models

USENIX
Add to list