Главная
Study mode:
on
1
Serving a model for 100 customers
2
Video Overview
3
Choosing a server
4
Choosing software to serve an API
5
One-click templates
6
Tips on GPU selection.
7
Using quantisation to fit in a cheaper GPU
8
Vast.ai setup
9
Serve Mistral with vLLM and AWQ, incl. concurrent requests
10
Serving a function calling model
11
API speed tests, including concurrent
12
Video Recap
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Learn how to serve a custom Large Language Model (LLM) for over 100 customers in this comprehensive 52-minute video tutorial. Explore key topics including server selection, API software choices, and GPU optimization techniques. Discover one-click templates for easy implementation and gain insights on using quantization to maximize GPU efficiency. Follow along with a step-by-step Vast.ai setup and learn to serve Mistral with vLLM and AWQ, handling concurrent requests. Delve into function calling models and conduct API speed tests. Master the skills needed to efficiently deploy and manage LLMs for multiple users in this informative and practical guide.

Serve a Custom LLM for Over 100 Customers - GPU Selection, Quantization, and API Setup

Trelis Research
Add to list