Главная
Study mode:
on
1
Intro
2
Zero-sum games
3
Generative Adversarial Networks (GANS)
4
Why not using GANs as they are?
5
Membership inference attacks for GANS
6
Differential Privacy (DP) to the rescue!
7
TensorFlow Privacy
8
Differentially private GANS (DP-GAN)
9
DP-GAN: noisy Wasserstein GAN
10
DP-GAN results
11
Context-aware fair data publishing
12
Empirical risk minimization with MI?
13
Generative Adversarial Privacy & Fairness (GAPF)
14
Example: GAPF under log-loss
15
Data-driven GAPF
16
Penalty method
17
Real-life data: GENKI dataset
18
Adversary's neural network
19
Feedforward Neural Network (FNN) encoder
20
Transposed Convolution Neural Network (TCNNP) encoder
21
GENKI fairness vs utility
22
Siamese-GAPF (S-GAPF) What if the sensitive label can take many values?
23
Real-life data: HAR dataset
24
HAR fairness vs utility
25
Gaussian mixture data model
Description:
Explore generative adversarial models for privacy and fairness in this 38-minute lecture by Peter Kairouz from Google AI. Delve into zero-sum games, Generative Adversarial Networks (GANs), and their limitations in privacy preservation. Learn about membership inference attacks on GANs and how Differential Privacy (DP) can mitigate these risks. Discover TensorFlow Privacy and differentially private GANs (DP-GAN), including noisy Wasserstein GAN implementations and their results. Investigate context-aware fair data publishing and empirical risk minimization. Examine Generative Adversarial Privacy & Fairness (GAPF) concepts, including data-driven approaches and real-life applications using the GENKI dataset. Understand adversary's neural networks, feedforward and transposed convolution neural network encoders, and the trade-offs between fairness and utility. Explore Siamese-GAPF (S-GAPF) for handling sensitive labels with multiple values, and analyze fairness vs. utility in the Human Activity Recognition (HAR) dataset. Conclude with insights into Gaussian mixture data models for privacy and fairness in machine learning. Read more

Generative Adversarial Models for Privacy and Fairness

Simons Institute
Add to list
0:00 / 0:00