Главная
Study mode:
on
1
Introduction
2
Model Deployment
3
Kubernetes
4
Complexities
5
Kserve
6
Scalability
7
Model Mesh
8
Model Mesh Features
9
Performance Testing Automation
10
Performance Testing Setup
11
Performance Testing Environment
12
QFlow Pipeline
13
K6 Load Tools
14
GRPC
15
Prometheus
16
Demo
17
Testing
18
Testing Log
19
Testing Results
20
Cashmiss Action
21
Should I use Model Mesh
Description:
Explore the intricacies of performance testing for gRPC model inferencing at scale in this informative conference talk. Discover how to set up a Kubernetes cluster with KServe's ModelMesh for high-density deployment of machine learning models. Learn about load testing thousands of models and utilizing Prometheus and Grafana for monitoring key performance metrics. Gain insights into the complexities of model deployment, scalability challenges, and the features of Model Mesh. Delve into the automation of performance testing, including the setup of testing environments, QFlow pipeline, and K6 load tools. Witness a demonstration of the testing process, analyze testing logs and results, and understand the implications of cashmiss actions. Evaluate the benefits of using Model Mesh for your specific use case.

Enhancing the Performance Testing Process for gRPC Model Inferencing at Scale

CNCF [Cloud Native Computing Foundation]
Add to list
0:00 / 0:00