Machine learning is increasingly being used in critical decision-making systems, yet is not reliable in the presence of noisy, biased, and adversarial data. Can we trust machine learning models? This course aims to answer this question, by covering the fundamental aspects of reasoning about trust in machine learning, including its robustness to adversarial data and model manipulations, the privacy risks of machine learning algorithms for sensitive data, the fairness measures for machine learning, and transparency in AI. It covers the algorithms that analyze machine learning vulnerabilities; and techniques for building reliable and trustworthy machine learning algorithms.
We will cover the fundamental concepts in:
Some of the references can be found at https://trustworthy-machine-learning.github.io
[Week 01] Course overview; Introduction to trustworthy machine learning
[Week 02] Robustness: Inference in the adversarial setting
[Week 03] Robustness: Robust inference in the adversarial setting
[Week 04] Robustness: Learning in the adversarial setting
[Week 05] Privacy: Introduction to anonymity and data privacy
[Week 06] Privacy: Inference attacks
[Week 07] Privacy: Quantitative reasoning about data privacy in machine learning
[Week 08] Privacy: Differentially private machine learning
[Week 09] Fairness: Bias in machine learning
[Week 10] Fairness: Satisfying fairness criteria in machine learning
[Week 11] Federated learning: Privacy
[Week 12] Federated learning: Robustness and Fairness