- About
- Events
- Calendar
- Graduation Information
- Cornell Learning Machines Seminar
- Student Colloquium
- BOOM
- Fall 2024 Colloquium
- Conway-Walker Lecture Series
- Salton 2024 Lecture Series
- Seminars / Lectures
- Big Red Hacks
- Cornell University - High School Programming Contests 2024
- Game Design Initiative
- CSMore: The Rising Sophomore Summer Program in Computer Science
- Explore CS Research
- ACSU Research Night
- Cornell Junior Theorists' Workshop 2024
- People
- Courses
- Research
- Undergraduate
- M Eng
- MS
- PhD
- Admissions
- Current Students
- Computer Science Graduate Office Hours
- Advising Guide for Research Students
- Business Card Policy
- Cornell Tech
- Curricular Practical Training
- A & B Exam Scheduling Guidelines
- Fellowship Opportunities
- Field of Computer Science Ph.D. Student Handbook
- Graduate TA Handbook
- Field A Exam Summary Form
- Graduate School Forms
- Instructor / TA Application
- Ph.D. Requirements
- Ph.D. Student Financial Support
- Special Committee Selection
- Travel Funding Opportunities
- Travel Reimbursement Guide
- The Outside Minor Requirement
- Diversity and Inclusion
- Graduation Information
- CS Graduate Minor
- Outreach Opportunities
- Parental Accommodation Policy
- Special Masters
- Student Spotlights
- Contact PhD Office
Distributed Differential Privacy in the Shuffle Model
Abstract: How can we analyze sensitive data without compromising those users’ privacy? And how can we do it when data is distributed across millions of devices, and no one is trusted to aggregate the data? These problems are currently being solved using systems that combine differential privacy and lightweight forms of cryptographic secure computation, which impose significant restrictions on what differentially private algorithms we can implement. In this work I will introduce the shuffle model of differential privacy as a way to study how to design differentially private algorithms for these systems and to understand their limitations. I will present both algorithms and lower bounds for this model showing that for low-dimensional problems---computing counts, means, or histograms---this model is extremely powerful, but for high-dimensional problems---selecting relevant features or learning parities---this model is extremely limited.
This talk will be based on joint work with Albert Cheu, Adam Smith, David Zeber, and Maxim Zhilyaev.
Bio: Jonathan Ullman is an Associate Professor in the Khoury College of Computer Sciences at Northeastern University. Before joining Northeastern, he received his PhD from Harvard in 2013, and in 2014 was a Junior Fellow in the Simons Society of Fellows. His research centers on privacy for machine learning and statistics, and its surprising connections to topics like statistical validity, robustness, cryptography, and fairness. He has been recognized with an NSF CAREER award and the Ruth and Joel Spira Outstanding Teacher Award.