If playback doesn't begin shortly, try restarting your device.
•
You're signed out
Videos you watch may be added to the TV's watch history and influence TV recommendations. To avoid this, cancel and sign in to YouTube on your computer.
CancelConfirm
Share
An error occurred while retrieving sharing information. Please try again later.
To make a dataset more privacy preserving, a generative adversarial model trained with differential privacy generates a synthetic version of the dataset. This synthetic dataset preserves the original dataset’s statistical properties, while minimizing privacy risks. However, the preservation works less well for groups underrepresented in the data. If an AI were to be trained upon the synthetic dataset, it would be unfairly inaccurate. Future generative models should not only reproduce the original data in a privacy-preserving manner but also guarantee fairness across subgroups.
Reference links:
1. Robin Hood and Matthew Effects: Differential Privacy Has Disparate Impact on Synthetic Data https://proceedings.mlr.press/v162/ga...
2. Exploiting Unintended Feature Leakage in Collaborative Learning https://arxiv.org/pdf/1805.04049.pdf
To find out more, see the Nokia Bell Labs Responsible AI hub: https://www.bell-labs.com/research-in...…...more
Privacy and Synthetic Data: The Good, The Bad, and The Ugly
5Likes
482Views
2023Jun 6
To make a dataset more privacy preserving, a generative adversarial model trained with differential privacy generates a synthetic version of the dataset. This synthetic dataset preserves the original dataset’s statistical properties, while minimizing privacy risks. However, the preservation works less well for groups underrepresented in the data. If an AI were to be trained upon the synthetic dataset, it would be unfairly inaccurate. Future generative models should not only reproduce the original data in a privacy-preserving manner but also guarantee fairness across subgroups.
Reference links:
1. Robin Hood and Matthew Effects: Differential Privacy Has Disparate Impact on Synthetic Data https://proceedings.mlr.press/v162/ga...
2. Exploiting Unintended Feature Leakage in Collaborative Learning https://arxiv.org/pdf/1805.04049.pdf
To find out more, see the Nokia Bell Labs Responsible AI hub: https://www.bell-labs.com/research-in...…...more