Главная
Study mode:
on
1
Introduction
2
Why Red teaming
3
What is Red teaming
4
What to test
5
Failure modes
6
Automation
7
Red teaming
8
Prompt injection attack
9
Prompt extraction
10
Data transformation
11
Model alignment test
12
Summary
Description:
Explore cutting-edge techniques for evaluating and stress-testing Large Language Models (LLMs) in this lightning talk from the LLMs in Production Conference III. Learn about a comprehensive framework for managing AI risk throughout the model lifecycle, from data collection to production deployment. Discover methods for red-teaming generative AI systems and building validation engines that algorithmically probe models for security, ethics, and safety issues. Gain insights into failure modes, automation strategies, and specific testing approaches such as prompt injection attacks, prompt extraction, data transformation, and model alignment tests. Equip yourself with the knowledge to effectively assess and mitigate potential risks associated with LLMs in production environments.

Evaluating LLMs for AI Risk - Techniques for Red Teaming Generative AI

MLOps.community
Add to list