Unprecedented quantities of data that could help social scientists understand and ameliorate the challenges of human society are presently locked away inside companies, governments, and other organizations, in part because of worries about privacy violations. Differential Privacy is a technique that might help change this—a technological solution to the political problem of data sharing. Differential privacy provides mathematical guarantees to protect the privacy of individuals who may be in the data while simultaneously making it possible for social scientists to gain insights into societal level patterns and relationships. However, many differentially private solutions are statistically invalid—giving biased estimates or no measures of uncertainty. We aim to fix this. Just as differential privacy can protect individuals, we are developing statistically valid differentially private systems that will also protect researchers and society so that we can be confident of scholarly conclusions, policy advice, and personal recommendations.
As part of this effort, we are partnering with Microsoft and the Privacy Tools Project (at SEAS) to incubate OpenDP, a set of open source software tools for privacy-protective statistically valid analyses of sensitive personal data; the first end-to-end system we've built together is SmartNoise. We are also working together to build a broader OpenDP community with stakeholders and contributors from across academia, industry, and government. Together, we will design, implement, and govern an “OpenDP Commons” that includes a library of differentially private algorithms and other general-purpose tools for use in end-to-end differential privacy systems.
In addition to providing our tools to the public, we are integrating them into IQSS' Dataverse project to enable privacy protected data sharing. These developments are especially helpful in enabling commercial entities to share data with academics to create public good, an effort we organize at Social Science One, which includes a consortium of social science research centers we have organized and various data sharing collaborations with industry.
For generous contributions to the Privacy Insights Project, we thank the Alfred P. Sloan Foundation, Microsoft, Facebook, and private donors.