Fill out schedule with papers.
This commit is contained in:
parent
fc344fadbb
commit
4bbc39bf34
|
@ -26,12 +26,13 @@ These three components are detailed below.
|
||||||
### Paper presentations
|
### Paper presentations
|
||||||
|
|
||||||
**Paper discussions** are one of the main components of this course. In groups
|
**Paper discussions** are one of the main components of this course. In groups
|
||||||
of two (or very rarely three), you will present 2-3 papers on a related topic
|
of two (or very rarely three), you will present 1-2 papers on a related topic
|
||||||
and lead the discussion; we will have presentations most Wednesdays and Fridays.
|
and lead the discussion. We will have presentations most Wednesdays and Fridays,
|
||||||
Your presentation should last about **60 minutes** long, leaving the remainder
|
Each presentation should be about **60 minutes**, leaving the remainder of the
|
||||||
of the time for a wrap-up discussion. Please sign up for a slot and a paper by
|
time for a wrap-up discussion. Please sign up for a slot by **Monday, September
|
||||||
**Monday, September 9**; while we will try to accommodate everyone's interests,
|
9**; see the [calendar](schedule/lectures.md) for the topic and suggested papers
|
||||||
we may need to adjust the selections for better balance and coverage.
|
for each slot. While we will try to accommodate everyone's interests, we may
|
||||||
|
need to adjust the selections for better balance and coverage.
|
||||||
|
|
||||||
Before every presentation, all students are expected to read the papers closely
|
Before every presentation, all students are expected to read the papers closely
|
||||||
and understand their significance, including (a) the main problems, (b) the
|
and understand their significance, including (a) the main problems, (b) the
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
# Paper Suggestions
|
# Assorted Papers
|
||||||
|
|
||||||
### Differential Privacy
|
### Differential Privacy
|
||||||
- Frank McSherry and Kunal Talwar.
|
- Frank McSherry and Kunal Talwar.
|
||||||
|
@ -10,6 +10,9 @@
|
||||||
- T.-H. Hubert Chan, Elaine Shi, and Dawn Song.
|
- T.-H. Hubert Chan, Elaine Shi, and Dawn Song.
|
||||||
[*Private and Continual Release of Statistics*](https://eprint.iacr.org/2010/076.pdf).
|
[*Private and Continual Release of Statistics*](https://eprint.iacr.org/2010/076.pdf).
|
||||||
ICALP 2010.
|
ICALP 2010.
|
||||||
|
- Ilya Mironov.
|
||||||
|
[*On Significance of the Least Significant Bits For Differential Privacy*](http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.366.5957&rep=rep1&type=pdf).
|
||||||
|
CCS 2012.
|
||||||
- Moritz Hardt, Katrina Ligett, and Frank McSherry.
|
- Moritz Hardt, Katrina Ligett, and Frank McSherry.
|
||||||
[*A Simple and Practical Algorithm for Differentially Private Data Release*](https://papers.nips.cc/paper/4548-a-simple-and-practical-algorithm-for-differentially-private-data-release.pdf).
|
[*A Simple and Practical Algorithm for Differentially Private Data Release*](https://papers.nips.cc/paper/4548-a-simple-and-practical-algorithm-for-differentially-private-data-release.pdf).
|
||||||
NIPS 2012.
|
NIPS 2012.
|
||||||
|
@ -22,44 +25,71 @@
|
||||||
- Cynthia Dwork, Moni Naor, Omer Reingold, and Guy N. Rothblum.
|
- Cynthia Dwork, Moni Naor, Omer Reingold, and Guy N. Rothblum.
|
||||||
[*Pure Differential Privacy for Rectangle Queries via Private Partitions*](https://guyrothblum.files.wordpress.com/2017/06/dnrr15.pdf).
|
[*Pure Differential Privacy for Rectangle Queries via Private Partitions*](https://guyrothblum.files.wordpress.com/2017/06/dnrr15.pdf).
|
||||||
ASIACRYPT 2015.
|
ASIACRYPT 2015.
|
||||||
|
- Martín Abadi, Andy Chu, Ian Goodfellow, H. Brendan McMahan, Ilya Mironov, Kunal Talwar, and Li Zhang.
|
||||||
|
[*Deep Learning with Differential Privacy*](https://arxiv.org/pdf/1607.00133).
|
||||||
|
CCS 2016.
|
||||||
|
- Martín Abadi, Úlfar Erlingsson, Ian Goodfellow, H. Brendan McMahan, Ilya Mironov, Nicolas Papernot, Kunal Talwar, and Li Zhang.
|
||||||
|
[*On the Protection of Private Information in Machine Learning Systems: Two Recent Approaches*](https://arxiv.org/pdf/1708.08022).
|
||||||
|
CSF 2016.
|
||||||
|
- Nicolas Papernot, Martín Abadi, Úlfar Erlingsson, Ian Goodfellow, and Kunal Talwar.
|
||||||
|
[*Semi-supervised Knowledge Transfer for Deep Learning from Private Training Data*](https://arxiv.org/pdf/1610.05755).
|
||||||
|
ICLR 2017.
|
||||||
|
- Nicolas Papernot, Shuang Song, Ilya Mironov, Ananth Raghunathan, Kunal Talwar, and Úlfar Erlingsson.
|
||||||
|
[*Scalable Private Learning with PATE*](https://arxiv.org/pdf/1802.08908).
|
||||||
|
ICLR 2018.
|
||||||
- Matthew Joseph, Aaron Roth, Jonathan Ullman, and Bo Waggoner.
|
- Matthew Joseph, Aaron Roth, Jonathan Ullman, and Bo Waggoner.
|
||||||
[*Local Differential Privacy for Evolving Data*](https://arxiv.org/abs/1802.07128).
|
[*Local Differential Privacy for Evolving Data*](https://arxiv.org/abs/1802.07128).
|
||||||
NIPS 2018.
|
NeurIPS 2018.
|
||||||
- Albert Cheu, Adam Smith, Jonathan Ullman, David Zeber, and Maxim Zhilyaev.
|
- Albert Cheu, Adam Smith, Jonathan Ullman, David Zeber, and Maxim Zhilyaev.
|
||||||
[*Distributed Differential Privacy via Shuffling*](https://arxiv.org/pdf/1808.01394).
|
[*Distributed Differential Privacy via Shuffling*](https://arxiv.org/pdf/1808.01394).
|
||||||
EUROCRYPT 2019.
|
EUROCRYPT 2019.
|
||||||
|
- Úlfar Erlingsson, Vitaly Feldman, Ilya Mironov, Ananth Raghunathan, Kunal Talwar, and Abhradeep Thakurta.
|
||||||
|
[*Amplification by Shuffling: From Local to Central Differential Privacy via Anonymity*](https://arxiv.org/pdf/1811.12469).
|
||||||
|
SODA 2019.
|
||||||
- Jingcheng Liu and Kunal Talwar.
|
- Jingcheng Liu and Kunal Talwar.
|
||||||
[*Private Selection from Private Candidates*](https://arxiv.org/pdf/1811.07971).
|
[*Private Selection from Private Candidates*](https://arxiv.org/pdf/1811.07971).
|
||||||
STOC 2019.
|
STOC 2019.
|
||||||
|
|
||||||
### Adversarial Machine Learning
|
### Adversarial ML
|
||||||
|
- Battista Biggio, Blaine Nelson, and Pavel Laskov.
|
||||||
|
[*Poisoning Attacks against Support Vector Machines*](https://arxiv.org/pdf/1206.6389).
|
||||||
|
ICML 2012.
|
||||||
- Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus.
|
- Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus.
|
||||||
[*Intriguing Properties of Neural Networks*](https://arxiv.org/pdf/1312.6199.pdf).
|
[*Intriguing Properties of Neural Networks*](https://arxiv.org/pdf/1312.6199.pdf).
|
||||||
ICLR 2014.
|
ICLR 2014.
|
||||||
- Ian J. Goodfellow, Jonathon Shlens, and Christian Szegedy.
|
- Ian J. Goodfellow, Jonathon Shlens, and Christian Szegedy.
|
||||||
[*Explaining and Harnessing Adversarial Examples*](https://arxiv.org/abs/1412.6572).
|
[*Explaining and Harnessing Adversarial Examples*](https://arxiv.org/abs/1412.6572).
|
||||||
ICLR 2015.
|
ICLR 2015.
|
||||||
|
- Matt Fredrikson, Somesh Jha, and Thomas Ristenpart.
|
||||||
|
[*Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures*](https://www.cs.cmu.edu/~mfredrik/papers/fjr2015ccs.pdf).
|
||||||
|
CCS 2015.
|
||||||
- Nicholas Carlini and David Wagner.
|
- Nicholas Carlini and David Wagner.
|
||||||
[*Towards Evaluating the Robustness of Neural Networks*](https://arxiv.org/pdf/1608.04644.pdf).
|
[*Towards Evaluating the Robustness of Neural Networks*](https://arxiv.org/pdf/1608.04644.pdf).
|
||||||
S&P 2017.
|
S&P 2017.
|
||||||
- Kevin Eykholt, Ivan Evtimov, Earlence Fernandes, Bo Li, Amir Rahmati, Chaowei Xiao, Atul Prakash, Tadayoshi Kohno, and Dawn Song.
|
- Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov.
|
||||||
[*Robust Physical-World Attacks on Deep Learning Models*](https://arxiv.org/pdf/1707.08945.pdf).
|
[*Membership Inference Attacks against Machine Learning Models*](https://arxiv.org/pdf/1610.05820).
|
||||||
CVPR 2018.
|
S&P 2017.
|
||||||
- Nicholas Carlini and David Wagner.
|
- Nicholas Carlini and David Wagner.
|
||||||
[*Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods*](https://arxiv.org/pdf/1705.07263.pdf).
|
[*Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods*](https://arxiv.org/pdf/1705.07263.pdf).
|
||||||
AISec 2017.
|
AISec 2017.
|
||||||
- Jacob Steinhardt, Pang Wei Koh, and Percy Liang.
|
- Jacob Steinhardt, Pang Wei Koh, and Percy Liang.
|
||||||
[*Certified Defenses for Data Poisoning Attacks*](https://arxiv.org/pdf/1706.03691.pdf).
|
[*Certified Defenses for Data Poisoning Attacks*](https://arxiv.org/pdf/1706.03691.pdf).
|
||||||
NIPS 2017.
|
NIPS 2017.
|
||||||
|
- Kevin Eykholt, Ivan Evtimov, Earlence Fernandes, Bo Li, Amir Rahmati, Chaowei Xiao, Atul Prakash, Tadayoshi Kohno, and Dawn Song.
|
||||||
|
[*Robust Physical-World Attacks on Deep Learning Models*](https://arxiv.org/pdf/1707.08945.pdf).
|
||||||
|
CVPR 2018.
|
||||||
- Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu.
|
- Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu.
|
||||||
[*Towards Deep Learning Models Resistant to Adversarial Attacks*](https://arxiv.org/pdf/1706.06083.pdf).
|
[*Towards Deep Learning Models Resistant to Adversarial Attacks*](https://arxiv.org/pdf/1706.06083.pdf).
|
||||||
ICLR 2018.
|
ICLR 2018.
|
||||||
|
- Aditi Raghunathan, Jacob Steinhardt, and Percy Liang.
|
||||||
|
[*Certified Defenses against Adversarial Examples*](https://arxiv.org/pdf/1801.09344).
|
||||||
|
ICLR 2018.
|
||||||
- Vitaly Feldman.
|
- Vitaly Feldman.
|
||||||
[*Does Learning Require Memorization? A Short Tale about a Long Tail*](https://arxiv.org/pdf/1906.05271).
|
[*Does Learning Require Memorization? A Short Tale about a Long Tail*](https://arxiv.org/pdf/1906.05271).
|
||||||
arXiv 2019.
|
arXiv 2019.
|
||||||
- Nicholas Carlini, Chang Liu, Úlfar Erlingsson, Jernej Kos, and Dawn Song.
|
- Nicholas Carlini, Chang Liu, Úlfar Erlingsson, Jernej Kos, and Dawn Song.
|
||||||
[*The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks*](https://arxiv.org/pdf/1802.08232).
|
[*The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks*](https://arxiv.org/pdf/1802.08232).
|
||||||
USENIX Security 2019.
|
USENIX 2019.
|
||||||
|
|
||||||
### Applied Cryptography
|
### Applied Cryptography
|
||||||
- Benjamin Braun, Ariel J. Feldman, Zuocheng Ren, Srinath Setty, Andrew J. Blumberg, and Michael Walfish.
|
- Benjamin Braun, Ariel J. Feldman, Zuocheng Ren, Srinath Setty, Andrew J. Blumberg, and Michael Walfish.
|
||||||
|
@ -89,12 +119,24 @@
|
||||||
- Henry Corrigan-Gibbs and Dan Boneh.
|
- Henry Corrigan-Gibbs and Dan Boneh.
|
||||||
[*Prio: Private, Robust, and Scalable Computation of Aggregate Statistics*](https://people.csail.mit.edu/henrycg/files/academic/papers/nsdi17prio.pdf).
|
[*Prio: Private, Robust, and Scalable Computation of Aggregate Statistics*](https://people.csail.mit.edu/henrycg/files/academic/papers/nsdi17prio.pdf).
|
||||||
NSDI 2017.
|
NSDI 2017.
|
||||||
|
- Zahra Ghodsi, Tianyu Gu, Siddharth Garg.
|
||||||
|
[*SafetyNets: Verifiable Execution of Deep Neural Networks on an Untrusted Cloud*](https://arxiv.org/pdf/1706.10268).
|
||||||
|
NIPS 2017.
|
||||||
- Valerie Chen, Valerio Pastro, Mariana Raykova.
|
- Valerie Chen, Valerio Pastro, Mariana Raykova.
|
||||||
[*Secure Computation for Machine Learning With SPDZ*](https://arxiv.org/pdf/1901.00329).
|
[*Secure Computation for Machine Learning With SPDZ*](https://arxiv.org/pdf/1901.00329).
|
||||||
NIPS 2018.
|
NeurIPS 2018.
|
||||||
|
- Jialong Zhang, Zhongshu Gu, Jiyong Jang, Hui Wu, Marc Ph. Stoecklin, Heqing Huang, and Ian Molloy.
|
||||||
|
[*Protecting Intellectual Property of Deep Neural Networks with Watermarking*](https://gzs715.github.io/pubs/WATERMARK_ASIACCS18.pdf).
|
||||||
|
AsiaCCS 2018.
|
||||||
|
- Yossi Adi, Carsten Baum, Moustapha Cisse, Benny Pinkas, and Joseph Keshet.
|
||||||
|
[*Turning Your Weakness Into a Strength: Watermarking Deep Neural Networks by Backdooring*](https://arxiv.org/pdf/1802.04633).
|
||||||
|
USENIX 2018.
|
||||||
- Wenting Zheng, Raluca Ada Popa, Joseph E. Gonzalez, Ion Stoica.
|
- Wenting Zheng, Raluca Ada Popa, Joseph E. Gonzalez, Ion Stoica.
|
||||||
[*Helen: Maliciously Secure Coopetitive Learning for Linear Models*](https://arxiv.org/pdf/1907.07212).
|
[*Helen: Maliciously Secure Coopetitive Learning for Linear Models*](https://arxiv.org/pdf/1907.07212).
|
||||||
S&P 2019.
|
S&P 2019.
|
||||||
|
- Bita Darvish Rouhani, Huili Chen, and Farinaz Koushanfar.
|
||||||
|
[*DeepSigns: A Generic Watermarking Framework for IP Protection of Deep Learning Models*](https://arxiv.org/pdf/1804.00750).
|
||||||
|
ASPLOS 2019.
|
||||||
|
|
||||||
### Algorithmic Fairness
|
### Algorithmic Fairness
|
||||||
- Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Rich Zemel.
|
- Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Rich Zemel.
|
||||||
|
@ -106,9 +148,21 @@
|
||||||
- Tolga Bolukbasi, Kai-Wei Chang, James Zou, Venkatesh Saligrama, and Adam Kalai.
|
- Tolga Bolukbasi, Kai-Wei Chang, James Zou, Venkatesh Saligrama, and Adam Kalai.
|
||||||
[*Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings*](https://arxiv.org/pdf/1607.06520).
|
[*Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings*](https://arxiv.org/pdf/1607.06520).
|
||||||
NIPS 2016.
|
NIPS 2016.
|
||||||
|
- Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, and Kai-Wei Chang.
|
||||||
|
[*Men Also Like Shopping: Reducing Gender Bias Amplification using Corpus-level Constraints*](https://arxiv.org/pdf/1707.09457).
|
||||||
|
EMNLP 2017.
|
||||||
- Jon Kleinberg, Sendhil Mullainathan, and Manish Raghavan.
|
- Jon Kleinberg, Sendhil Mullainathan, and Manish Raghavan.
|
||||||
[*Inherent Trade-Offs in the Fair Determination of Risk Scores*](https://arxiv.org/pdf/1609.05807).
|
[*Inherent Trade-Offs in the Fair Determination of Risk Scores*](https://arxiv.org/pdf/1609.05807).
|
||||||
ITCS 2017.
|
ITCS 2017.
|
||||||
|
- Niki Kilbertus, Mateo Rojas-Carulla, Giambattista Parascandolo, Moritz Hardt, Dominik Janzing, and Bernhard Schölkopf.
|
||||||
|
[*Avoiding Discrimination through Causal Reasoning*](https://arxiv.org/pdf/1706.02744).
|
||||||
|
NIPS 2017.
|
||||||
|
- Matt J. Kusner, Joshua R. Loftus, Chris Russell, Ricardo Silva.
|
||||||
|
[*Counterfactual Fairness*](https://arxiv.org/pdf/1703.06856).
|
||||||
|
NIPS 2017.
|
||||||
|
- Razieh Nabi and Ilya Shpitser.
|
||||||
|
[*Fair Inference on Outcomes*](https://arxiv.org/pdf/1705.10378).
|
||||||
|
AAAI 2018.
|
||||||
- Úrsula Hébert-Johnson, Michael P. Kim, Omer Reingold, and Guy N. Rothblum.
|
- Úrsula Hébert-Johnson, Michael P. Kim, Omer Reingold, and Guy N. Rothblum.
|
||||||
[*Multicalibration: Calibration for the (Computationally-Identifiable) Masses*](https://arxiv.org/pdf/1711.08513.pdf).
|
[*Multicalibration: Calibration for the (Computationally-Identifiable) Masses*](https://arxiv.org/pdf/1711.08513.pdf).
|
||||||
ICML 2018.
|
ICML 2018.
|
||||||
|
@ -122,13 +176,19 @@
|
||||||
[*50 Years of Test (Un)fairness: Lessons for Machine Learning*](https://arxiv.org/pdf/1811.10104).
|
[*50 Years of Test (Un)fairness: Lessons for Machine Learning*](https://arxiv.org/pdf/1811.10104).
|
||||||
FAT\* 2019.
|
FAT\* 2019.
|
||||||
|
|
||||||
### Programming Languages and Verification
|
### PL and Verification
|
||||||
- Martín Abadi and Andrew D. Gordon.
|
- Martín Abadi and Andrew D. Gordon.
|
||||||
[*A Calculus for Cryptographic Protocols: The Spi Calculus*](https://www.microsoft.com/en-us/research/wp-content/uploads/2016/11/ic99spi.pdf).
|
[*A Calculus for Cryptographic Protocols: The Spi Calculus*](https://www.microsoft.com/en-us/research/wp-content/uploads/2016/11/ic99spi.pdf).
|
||||||
Information and Computation, 1999.
|
Information and Computation, 1999.
|
||||||
|
- Noah Goodman, Vikash Mansinghka, Daniel M. Roy, Keith Bonawitz, Joshua B. Tenenbaum.
|
||||||
|
[*Church: a language for generative models*](https://arxiv.org/pdf/1206.3255).
|
||||||
|
UAI 2008.
|
||||||
- Frank McSherry.
|
- Frank McSherry.
|
||||||
[*Privacy Integrated Queries*](http://citeseerx.ist.psu.edu/viewdoc/download?rep=rep1&type=pdf&doi=10.1.1.211.4503).
|
[*Privacy Integrated Queries*](http://citeseerx.ist.psu.edu/viewdoc/download?rep=rep1&type=pdf&doi=10.1.1.211.4503).
|
||||||
SIGMOD 2009.
|
SIGMOD 2009.
|
||||||
|
- Marta Kwiatkowska, Gethin Norman, and David Parker.
|
||||||
|
[*Advances and Challenges of Probabilistic Model Checking*](https://www.prismmodelchecker.org/papers/allerton10.pdf).
|
||||||
|
Allerton 2010.
|
||||||
- Jason Reed and Benjamin C. Pierce.
|
- Jason Reed and Benjamin C. Pierce.
|
||||||
[*Distance Makes the Types Grow Stronger: A Calculus for Differential Privacy*](https://www.cis.upenn.edu/~bcpierce/papers/dp.pdf).
|
[*Distance Makes the Types Grow Stronger: A Calculus for Differential Privacy*](https://www.cis.upenn.edu/~bcpierce/papers/dp.pdf).
|
||||||
ICFP 2010.
|
ICFP 2010.
|
||||||
|
@ -141,6 +201,9 @@
|
||||||
- Andrew Miller, Michael Hicks, Jonathan Katz, and Elaine Shi.
|
- Andrew Miller, Michael Hicks, Jonathan Katz, and Elaine Shi.
|
||||||
[*Authenticated Data Structures, Generically*](https://www.cs.umd.edu/~mwh/papers/gpads.pdf).
|
[*Authenticated Data Structures, Generically*](https://www.cs.umd.edu/~mwh/papers/gpads.pdf).
|
||||||
POPL 2014.
|
POPL 2014.
|
||||||
|
- Andrew D. Gordon, Thomas A. Henzinger, Aditya V. Nori, and Sriram K. Rajamani.
|
||||||
|
[*Probabilistic Programming*](https://www.microsoft.com/en-us/research/wp-content/uploads/2016/02/fose-icse2014.pdf).
|
||||||
|
ICSE 2014.
|
||||||
- Gilles Barthe, Marco Gaboardi, Emilio Jesús Gallego Arias, Justin Hsu, Aaron Roth, and Pierre-Yves Strub.
|
- Gilles Barthe, Marco Gaboardi, Emilio Jesús Gallego Arias, Justin Hsu, Aaron Roth, and Pierre-Yves Strub.
|
||||||
[*Higher-Order Approximate Relational Refinement Types for Mechanism Design and Differential Privacy*](https://arxiv.org/pdf/1407.6845.pdf).
|
[*Higher-Order Approximate Relational Refinement Types for Mechanism Design and Differential Privacy*](https://arxiv.org/pdf/1407.6845.pdf).
|
||||||
POPL 2015.
|
POPL 2015.
|
||||||
|
@ -150,9 +213,27 @@
|
||||||
- Chang Liu, Xiao Shaun Wang, Kartik Nayak, Yan Huang, and Elaine Shi.
|
- Chang Liu, Xiao Shaun Wang, Kartik Nayak, Yan Huang, and Elaine Shi.
|
||||||
[*ObliVM: A Programming Framework for Secure Computation*](http://www.cs.umd.edu/~elaine/docs/oblivm.pdf).
|
[*ObliVM: A Programming Framework for Secure Computation*](http://www.cs.umd.edu/~elaine/docs/oblivm.pdf).
|
||||||
S&P 2015.
|
S&P 2015.
|
||||||
|
- Gilles Barthe, Marco Gaboardi, Benjamin Grégoire, Justin Hsu, and Pierre-Yves Strub.
|
||||||
|
[*A Program Logic for Union Bounds*](https://arxiv.org/pdf/1602.05681).
|
||||||
|
ICALP 2016.
|
||||||
|
- Christian Albert Hammerschmidt, Sicco Verwer, Qin Lin, and Radu State.
|
||||||
|
[*Interpreting Finite Automata for Sequential Data*](https://arxiv.org/pdf/1611.07100).
|
||||||
|
NIPS 2016.
|
||||||
|
- Joost-Pieter Katoen.
|
||||||
|
[*The Probabilistic Model Checking Landscape*](https://moves.rwth-aachen.de/wp-content/uploads/lics2016_tutorial_katoen.pdf).
|
||||||
|
LICS 2016.
|
||||||
- Andrew Ferraiuolo, Rui Xu, Danfeng Zhang, Andrew C. Myers, and G. Edward Suh.
|
- Andrew Ferraiuolo, Rui Xu, Danfeng Zhang, Andrew C. Myers, and G. Edward Suh.
|
||||||
[*Verification of a Practical Hardware Security Architecture Through Static Information Flow Analysis*](http://www.cse.psu.edu/~dbz5017/pub/asplos17.pdf).
|
[*Verification of a Practical Hardware Security Architecture Through Static Information Flow Analysis*](http://www.cse.psu.edu/~dbz5017/pub/asplos17.pdf).
|
||||||
ASPLOS 2017.
|
ASPLOS 2017.
|
||||||
|
- Frits Vaandrager.
|
||||||
|
[*Model Learning*](https://m-cacm.acm.org/magazines/2017/2/212445-model-learning/fulltext).
|
||||||
|
CACM 2017.
|
||||||
|
- Timon Gehr, Matthew Mirman, Dana Drachsler-Cohen, Petar Tsankov, Swarat Chaudhuri, and Martin Vechev
|
||||||
|
[*AI2: Safety and Robustness Certification of Neural Networks with Abstract Interpretation*](https://files.sri.inf.ethz.ch/website/papers/sp2018.pdf).
|
||||||
|
S&P 2018.
|
||||||
|
- Marc Fischer, Mislav Balunovic, Dana Drachsler-Cohen, Timon Gehr, Ce Zhang, and Martin Vechev.
|
||||||
|
[*DL2: Training and Querying Neural Networks with Logic*](http://proceedings.mlr.press/v97/fischer19a/fischer19a.pdf).
|
||||||
|
ICML 2019.
|
||||||
|
|
||||||
# Supplemental Material
|
# Supplemental Material
|
||||||
- Cynthia Dwork and Aaron Roth.
|
- Cynthia Dwork and Aaron Roth.
|
||||||
|
@ -165,3 +246,9 @@
|
||||||
[*A Survey of Symbolic Methods in Computational Analysis of Cryptographic Systems*](https://hal.inria.fr/inria-00379776/document).
|
[*A Survey of Symbolic Methods in Computational Analysis of Cryptographic Systems*](https://hal.inria.fr/inria-00379776/document).
|
||||||
- Dan Boneh and Victor Shoup.
|
- Dan Boneh and Victor Shoup.
|
||||||
[*A Graduate Course in Applied Cryptography*](https://crypto.stanford.edu/~dabo/cryptobook/BonehShoup_0_4.pdf).
|
[*A Graduate Course in Applied Cryptography*](https://crypto.stanford.edu/~dabo/cryptobook/BonehShoup_0_4.pdf).
|
||||||
|
- David Hand.
|
||||||
|
[*Statistics and the Theory of Measurement*](http://www.lps.uci.edu/~johnsonk/CLASSES/MeasurementTheory/Hand1996.StatisticsAndTheTheoryOfMeasurement.pdf).
|
||||||
|
- Judea Pearl.
|
||||||
|
[*Causal inference in statistics: An overview*](http://ftp.cs.ucla.edu/pub/stat_ser/r350.pdf).
|
||||||
|
- Yehuda Lindell and Benny Pinkas.
|
||||||
|
[*Secure Multiparty Computation for Privacy-Preserving Data Mining*](https://eprint.iacr.org/2008/197.pdf).
|
||||||
|
|
|
@ -1,4 +1,6 @@
|
||||||
- CSE 291: [Language-Based Security](https://cseweb.ucsd.edu/~dstefan/cse291-winter18/) (Deian Stefan, UCSD)
|
- CSE 291: [Language-Based Security](https://cseweb.ucsd.edu/~dstefan/cse291-winter18/) (Deian Stefan, UC San Diego)
|
||||||
- CSE 711: [Topics in Differential Privacy](https://www.acsu.buffalo.edu/~gaboardi/teaching/CSE711-spring16.html) (Marco Gaboardi, University at Buffalo)
|
- CSE 711: [Topics in Differential Privacy](https://www.acsu.buffalo.edu/~gaboardi/teaching/CSE711-spring16.html) (Marco Gaboardi, University at Buffalo)
|
||||||
- CS 800: [The Algorithmic Foundations of Data Privacy](https://www.cis.upenn.edu/~aaroth/courses/privacyF11.html) (Aaron Roth, UPenn)
|
- CS 800: [The Algorithmic Foundations of Data Privacy](https://www.cis.upenn.edu/~aaroth/courses/privacyF11.html) (Aaron Roth, UPenn)
|
||||||
- CS 229r: [Mathematical Approaches to Data Privacy](http://people.seas.harvard.edu/~salil/diffprivcourse/spring13/) (Salil Vadhan, Harvard)
|
- CS 229r: [Mathematical Approaches to Data Privacy](http://people.seas.harvard.edu/~salil/diffprivcourse/spring13/) (Salil Vadhan, Harvard)
|
||||||
|
- CS 294: [Fairness in Machine Learning](https://fairmlclass.github.io/) (Moritz Hardt, UC Berkeley)
|
||||||
|
- CS 598: [Special Topics in Adversarial Machine Learning](http://www.crystal-boli.com/teaching.html) (Bo Li, UIUC)
|
||||||
|
|
|
@ -1,40 +1,40 @@
|
||||||
# Calendar
|
# Calendar (tentative)
|
||||||
|
|
||||||
Date | Topic | Notes
|
Date | Topic | Presenters | Notes
|
||||||
:----:|-------|:---------:
|
:----:|-------|:----------:|:-----:
|
||||||
| <center> <h4> **Differential Privacy** </h4> </center> |
|
| <center> <h4> **Differential Privacy** </h4> </center> | |
|
||||||
9/4 | [Course welcome](../resources/slides/lecture-welcome.html) <br> **Reading:** Keshav. [*How to Read a Paper*](https://web.stanford.edu/class/ee384m/Handouts/HowtoReadPaper.pdf). | HW1 Out
|
9/4 | [Course welcome](../resources/slides/lecture-welcome.html) <br> **Reading:** [*How to Read a Paper*](https://web.stanford.edu/class/ee384m/Handouts/HowtoReadPaper.pdf) | JH | HW1 Out
|
||||||
9/6 | Basic private mechanisms <br> **Reading:** AFDP 3.2-4 |
|
9/6 | Basic private mechanisms <br> **Reading:** AFDP 3.2-4 | JH |
|
||||||
9/9 | Composition and closure properties <br> **Reading:** AFDP 3.5 | Signups
|
9/9 | Composition and closure properties <br> **Reading:** AFDP 3.5 | JH | Paper Signups
|
||||||
9/11 | What does differential privacy actually mean? <br> **Reading:** McSherry. [Lunchtime for Differential Privacy](https://github.com/frankmcsherry/blog/blob/master/posts/2016-08-16.md) |
|
9/11 | What does differential privacy actually mean? <br> **Reading:** [Lunchtime for Differential Privacy](https://github.com/frankmcsherry/blog/blob/master/posts/2016-08-16.md) | JH |
|
||||||
9/13 | Paper presentations: Differential privacy | HW1 Due
|
9/13 | Differentially private machine learning <br> **Reading:** [*On the Protection of Private Information in Machine Learning Systems: Two Recent Approaches*](https://arxiv.org/pdf/1708.08022) <br> **Reading:** [*Semi-supervised Knowledge Transfer for Deep Learning from Private Training Data*](https://arxiv.org/pdf/1610.05755) | | HW1 Due
|
||||||
| <center> <h4> **Adversarial Machine Learning** </h4> </center> |
|
| <center> <h4> **Adversarial Machine Learning** </h4> </center> | |
|
||||||
9/16 | Overview and basic concepts | HW2 Out
|
9/16 | Overview and basic concepts | JH | HW2 Out
|
||||||
9/18 | Paper presentations: Adversarial attacks |
|
9/18 | Adversarial examples <br> **Reading:** [*Intriguing Properties of Neural Networks*](https://arxiv.org/pdf/1312.6199.pdf) <br> **Reading:** [*Explaining and Harnessing Adversarial Examples*](https://arxiv.org/abs/1412.6572) <br> **Reading:** [*Robust Physical-World Attacks on Deep Learning Models*](https://arxiv.org/pdf/1707.08945.pdf) | |
|
||||||
9/20 | Paper presentations: ??? |
|
9/20 | Data poisoning <br> **Reading:** [*Poisoning Attacks against Support Vector Machines*](https://arxiv.org/pdf/1206.6389) | |
|
||||||
9/23 | Adversarial training |
|
9/23 | Defenses and detection: challenges <br> **Reading:** [*Towards Evaluating the Robustness of Neural Networks*](https://arxiv.org/pdf/1608.04644.pdf) <br> **Reading:** [*Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods*](https://arxiv.org/pdf/1705.07263.pdf) | JH |
|
||||||
9/25 | Paper presentations: Certified defenses |
|
9/25 | Certified defenses <br> **Reading:** [*Certified Defenses for Data Poisoning Attacks*](https://arxiv.org/pdf/1706.03691.pdf) <br> **Reading:** [*Certified Defenses against Adversarial Examples*](https://arxiv.org/pdf/1801.09344) | |
|
||||||
9/27 | Paper presentations: ??? | HW2 Due
|
9/27 | Adversarial training <br> **Reading:** [*Towards Deep Learning Models Resistant to Adversarial Attacks*](https://arxiv.org/pdf/1706.06083.pdf) | | HW2 Due
|
||||||
| <center> <h4> **Applied Cryptography** </h4> </center> |
|
| <center> <h4> **Applied Cryptography** </h4> </center> | |
|
||||||
9/30 | Overview and basic constructions | HW3 Out
|
9/30 | Overview and basic constructions | JH | HW3 Out
|
||||||
10/2 | Paper presentations: Secure Multiparty Computation |
|
10/2 | SMC for machine learning <br> **Reading:** [*Secure Computation for Machine Learning With SPDZ*](https://arxiv.org/pdf/1901.00329) <br> **Reading:** [*Helen: Maliciously Secure Coopetitive Learning for Linear Models*](https://arxiv.org/pdf/1907.07212) | |
|
||||||
10/4 | Paper presentations: ??? |
|
10/4 | Secure data collection at scale <br> **Reading:** [*Prio: Private, Robust, and Scalable Computation of Aggregate Statistics*](https://people.csail.mit.edu/henrycg/files/academic/papers/nsdi17prio.pdf) | |
|
||||||
10/7 | Homomorphic Encryption |
|
10/7 | Verifiable computing <br> **Reading:** [*SafetyNets: Verifiable Execution of Deep Neural Networks on an Untrusted Cloud*](https://arxiv.org/pdf/1706.10268) | JH |
|
||||||
10/9 | Paper presentations: Oblivious computing and side channels |
|
10/9 | Side channels and implementation issues <br> **Reading:** [*On Significance of the Least Significant Bits For Differential Privacy*](http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.366.5957&rep=rep1&type=pdf) | |
|
||||||
10/11 | Paper presentations: ??? | HW3 Due <br> MS1 Due
|
10/11 | Model watermarking <br> **Reading:** [*Protecting Intellectual Property of Deep Neural Networks with Watermarking*](https://gzs715.github.io/pubs/WATERMARK_ASIACCS18.pdf) <br> **Reading:** [*Turning Your Weakness Into a Strength: Watermarking Deep Neural Networks by Backdooring*](https://arxiv.org/pdf/1802.04633) | | HW3 Due <br> MS1 Due
|
||||||
| <center> <h4> **Advanced Topic: Algorithmic Fairness** </h4> </center> |
|
| <center> <h4> **Advanced Topic: Algorithmic Fairness** </h4> </center> | |
|
||||||
10/14 | Overview and basic notions |
|
10/14 | Overview and basic notions <br> **Reading:** Chapter 2 from [Barocas, Hardt, and Narayanan](https://fairmlbook.org/demographic.html) | JH |
|
||||||
10/16 | Paper presentations: Individual and group fairness |
|
10/16 | Individual and group fairness <br> **Reading:** [*Fairness through Awarness*](https://arxiv.org/pdf/1104.3913) <br> **Reading:** [*Equality of Opportunity in Supervised Learning*](https://arxiv.org/pdf/1610.02413) | |
|
||||||
10/18 | Paper presentations: ??? |
|
10/18 | Inherent tradeoffs <br> **Reading:** [*Inherent Trade-Offs in the Fair Determination of Risk Scores*](https://arxiv.org/pdf/1609.05807) | |
|
||||||
10/21 | Challenges in defining fairness |
|
10/21 | Defining fairness: challenges <br> **Reading:** [*50 Years of Test (Un)fairness: Lessons for Machine Learning*](https://arxiv.org/pdf/1811.10104) | JH |
|
||||||
10/23 | Paper presentations: Repairing fairness |
|
10/23 | Fairness in unsupervised learning <br> **Reading:** [*Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings*](https://arxiv.org/pdf/1607.06520) <br> **Reading:** [*Men Also Like Shopping: Reducing Gender Bias Amplification using Corpus-level Constraints*](https://arxiv.org/pdf/1707.09457) | |
|
||||||
10/25 | Paper presentations: ??? |
|
10/25 | Beyond observational measures <br> **Reading:** [*Avoiding Discrimination through Causal Reasoning*](https://arxiv.org/pdf/1706.02744) <br> **Reading:** [*Counterfactual Fairness*](https://arxiv.org/pdf/1703.06856) | |
|
||||||
| <center> <h4> **Advanced Topic: PL and Verification** </h4> </center> |
|
| <center> <h4> **Advanced Topic: PL and Verification** </h4> </center> | |
|
||||||
10/28 | Overview and basic notions |
|
10/28 | Overview and basic notions | JH |
|
||||||
10/30 | Paper presentations: Probabilistic programming languages |
|
10/30 | Probabilistic programming languages <br> **Reading:** [*Probabilistic Programming*](https://www.microsoft.com/en-us/research/wp-content/uploads/2016/02/fose-icse2014.pdf) | |
|
||||||
11/1 | Paper presentations: ??? |
|
11/1 | Automata learning and interpretability <br> **Reading:** [*Model Learning*](https://m-cacm.acm.org/magazines/2017/2/212445-model-learning/fulltext) <br> **Reading:** [*Interpreting Finite Automata for Sequential Data*](https://arxiv.org/pdf/1611.07100) | |
|
||||||
11/4 | Programming languages for differential privacy |
|
11/4 | Programming languages for differential privacy <br> **Reading:** [*Programming Language Techniques for Differential Privacy*](https://dl.acm.org/citation.cfm?id=2893591&dl=ACM&coll=DL) | JH |
|
||||||
11/6 | Paper presentations: Verifying probabilistic programs |
|
11/6 | Verifying neural networks <br> **Reading:** [*AI2: Safety and Robustness Certification of Neural Networks with Abstract Interpretation*](https://files.sri.inf.ethz.ch/website/papers/sp2018.pdf) <br> **Reading:** [*DL2: Training and Querying Neural Networks with Logic*](http://proceedings.mlr.press/v97/fischer19a/fischer19a.pdf) | |
|
||||||
11/8 | Paper presentations: ??? | MS2 Due
|
11/8 | Verifying probabilistic programs <br> **Reading:** [*Advances and Challenges of Probabilistic Model Checking*](https://www.prismmodelchecker.org/papers/allerton10.pdf) <br> **Reading:** [*A Program Logic for Union Bounds*](https://arxiv.org/pdf/1602.05681) | | MS2 Due
|
||||||
| <center> <h4> **No Lectures: Work on Projects** </h4> </center> |
|
| <center> <h4> **No Lectures: Work on Projects** </h4> </center> | |
|
||||||
12/11 (TBD) | Project Presentations |
|
12/11 (TBD) | Project Presentations | |
|
||||||
|
|
Reference in New Issue