Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Abhradeep Thakurta is active.

Publication


Featured researches published by Abhradeep Thakurta.


international conference on management of data | 2012

GUPT: privacy preserving data analysis made easy

Prashanth Mohan; Abhradeep Thakurta; Elaine Shi; Dawn Song; David E. Culler

It is often highly valuable for organizations to have their data analyzed by external agents. However, any program that computes on potentially sensitive data risks leaking information through its output. Differential privacy provides a theoretical framework for processing data while protecting the privacy of individual records in a dataset. Unfortunately, it has seen limited adoption because of the loss in output accuracy, the difficulty in making programs differentially private, lack of mechanisms to describe the privacy budget in a programmers utilitarian terms, and the challenging requirement that data owners and data analysts manually distribute the limited privacy budget between queries. This paper presents the design and evaluation of a new system, GUPT, that overcomes these challenges. Unlike existing differentially private systems such as PINQ and Airavat, it guarantees differential privacy to programs not developed with privacy in mind, makes no trust assumptions about the analysis program, and is secure to all known classes of side-channel attacks. GUPT uses a new model of data sensitivity that degrades privacy of data over time. This enables efficient allocation of different levels of privacy for different user applications while guaranteeing an overall constant level of privacy and maximizing the utility of each application. GUPT also introduces techniques that improve the accuracy of output while achieving the same level of privacy. These approaches enable GUPT to easily execute a wide variety of data analysis programs while providing both utility and privacy.


foundations of computer science | 2014

Private Empirical Risk Minimization: Efficient Algorithms and Tight Error Bounds

Raef Bassily; Adam D. Smith; Abhradeep Thakurta

Convex empirical risk minimization is a basic tool in machine learning and statistics. We provide new algorithms and matching lower bounds for differentially private convex empirical risk minimization assuming only that each data points contribution to the loss function is Lipschitz and that the domain of optimization is bounded. We provide a separate set of algorithms and matching lower bounds for the setting in which the loss functions are known to also be strongly convex. Our algorithms run in polynomial time, and in some cases even match the optimal nonprivate running time (as measured by oracle complexity). We give separate algorithms (and lower bounds) for (ε, 0)and (ε, δ)-differential privacy; perhaps surprisingly, the techniques used for designing optimal algorithms in the two cases are completely different. Our lower bounds apply even to very simple, smooth function families, such as linear and quadratic functions. This implies that algorithms from previous work can be used to obtain optimal error rates, under the additional assumption that the contributions of each data point to the loss function is smooth. We show that simple approaches to smoothing arbitrary loss functions (in order to apply previous techniques) do not yield optimal error rates. In particular, optimal algorithms were not previously known for problems such as training support vector machines and the high-dimensional median.


symposium on the theory of computing | 2014

Analyze gauss: optimal bounds for privacy-preserving principal component analysis

Cynthia Dwork; Kunal Talwar; Abhradeep Thakurta; Li Zhang

We consider the problem of privately releasing a low dimensional approximation to a set of data records, represented as a matrix A in which each row corresponds to an individual and each column to an attribute. Our goal is to compute a subspace that captures the covariance of A as much as possible, classically known as principal component analysis (PCA). We assume that each row of A has ℓ2 norm bounded by one, and the privacy guarantee is defined with respect to addition or removal of any single row. We show that the well-known, but misnamed, randomized response algorithm, with properly tuned parameters, provides nearly optimal additive quality gap compared to the best possible singular subspace of A. We further show that when ATA has a large eigenvalue gap -- a reason often cited for PCA -- the quality improves significantly. Optimality (up to logarithmic factors) is proved using techniques inspired by the recent work of Bun, Ullman, and Vadhan on applying Tardoss fingerprinting codes to the construction of hard instances for private mechanisms for 1-way marginal queries. Along the way we define a list culling game which may be of independent interest. By combining the randomized response mechanism with the well-known following the perturbed leader algorithm of Kalai and Vempala we obtain a private online algorithm with nearly optimal regret. The regret of our algorithm even outperforms all the previously known online non-private algorithms of this type. We achieve this better bound by, satisfyingly, borrowing insights and tools from differential privacy!


international conference on the theory and application of cryptology and information security | 2011

Noiseless database privacy

Raghav Bhaskar; Abhishek Bhowmick; Vipul Goyal; Srivatsan Laxman; Abhradeep Thakurta

Differential Privacy (DP) has emerged as a formal, flexible framework for privacy protection, with a guarantee that is agnostic to auxiliary information and that admits simple rules for composition. Benefits notwithstanding, a major drawback of DP is that it provides noisy responses to queries, making it unsuitable for many applications. We propose a new notion called Noiseless Privacy that provides exact answers to queries, without adding any noise whatsoever. While the form of our guarantee is similar to DP, where the privacy comes from is very different, based on statistical assumptions on the data and on restrictions to the auxiliary information available to the adversary. We present a first set of results for Noiseless Privacy of arbitrary Boolean-function queries and of linear Real-function queries, when data are drawn independently, from nearly-uniform and Gaussian distributions respectively. We also derive simple rules for composition under models of dynamically changing data.


theory of cryptography conference | 2013

Testing the lipschitz property over product distributions with applications to data privacy

Kashyap Dixit; Madhav Jha; Sofya Raskhodnikova; Abhradeep Thakurta

In the past few years, the focus of research in the area of statistical data privacy has been in designing algorithms for various problems which satisfy some rigorous notions of privacy. However, not much effort has gone into designing techniques to computationally verify if a given algorithm satisfies some predefined notion of privacy. In this work, we address the following question: Can we design algorithms which tests if a given algorithm satisfies some specific rigorous notion of privacy (e.g., differential privacy)? We design algorithms to test privacy guarantees of a given algorithm


international workshop and international workshop on approximation, randomization, and combinatorial optimization. algorithms and techniques | 2012

Mirror Descent Based Database Privacy

Prateek Jain; Abhradeep Thakurta

\mathcal{A}


SIAM Journal on Computing | 2018

Erasure-Resilient Property Testing

Kashyap Dixit; Sofya Raskhodnikova; Abhradeep Thakurta; Nithin M. Varma

when run on a dataset x containing potentially sensitive information about the individuals. More formally, we design a computationally efficient algorithm


international colloquium on automata, languages and programming | 2016

Erasure-Resilient Property Testing.

Kashyap Dixit; Sofya Raskhodnikova; Abhradeep Thakurta; Nithin M. Varma

{\cal T}_{priv}


knowledge discovery and data mining | 2010

Discovering frequent patterns in sensitive data

Raghav Bhaskar; Srivatsan Laxman; Adam D. Smith; Abhradeep Thakurta

that verifies whether


conference on learning theory | 2012

Differentially Private Online Learning

Prateek Jain; Pravesh Kothari; Abhradeep Thakurta

\mathcal{A}

Collaboration


Dive into the Abhradeep Thakurta's collaboration.

Top Co-Authors

Avatar

Adam D. Smith

Pennsylvania State University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Kashyap Dixit

Pennsylvania State University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Sofya Raskhodnikova

Pennsylvania State University

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge