Mert Pilanci
University of California, Berkeley
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Mert Pilanci.
Siam Journal on Optimization | 2017
Mert Pilanci; Martin J. Wainwright
We propose a randomized second-order method for optimization known as the Newton Sketch: it is based on performing an approximate Newton step using a randomly projected or sub-sampled Hessian. For self-concordant functions, we prove that the algorithm has super-linear convergence with exponentially high probability, with convergence and complexity guarantees that are independent of condition numbers and related problem-dependent quantities. Given a suitable initialization, similar guarantees also hold for strongly convex and smooth objectives without self-concordance. When implemented using randomized projections based on a sub-sampled Hadamard basis, the algorithm typically has substantially lower complexity than Newtons method. We also describe extensions of our methods to programs involving convex constraints that are equipped with self-concordant barriers. We discuss and illustrate applications to linear programs, quadratic programs with convex constraints, logistic regression and other generalized linear models, as well as semidefinite programs.
IEEE Transactions on Information Theory | 2015
Mert Pilanci; Martin J. Wainwright
Random projection (RP) is a classical technique for reducing storage and computational costs. We analyze RP-based approximations of convex programs, in which the original optimization problem is approximated by solving a lower dimensional problem. Such dimensionality reduction is essential in computation-limited settings, since the complexity of general convex programming can be quite high (e.g., cubic for quadratic programs, and substantially higher for semidefinite programs). In addition to computational savings, RP is also useful for reducing memory usage, and has useful properties for privacy-preserving optimization. We prove that the approximation ratio of this procedure can be bounded in terms of the geometry of the constraint set. For a broad class of RPs, including those based on various sub-Gaussian distributions as well as randomized Hadamard and Fourier transforms, the data matrix defining the cost function can be projected to a dimension proportional to the squared Gaussian width of the tangent cone of the constraint set at the original solution. This effective dimension of the convex program is often substantially smaller than the original dimension. We illustrate consequences of our theory for various cases, including unconstrained and
Annals of Statistics | 2017
Yun Yang; Mert Pilanci; Martin J. Wainwright
\ell _{1}
international symposium on information theory | 2014
Mert Pilanci; Martin J. Wainwright
-constrained least squares, support vector machines, low-rank matrix estimation, and discuss implications for privacy-preserving optimization, as well as connections with denoising and compressed sensing.
IEEE Transactions on Signal Processing | 2010
Mert Pilanci; Orhan Arikan; Mustafa Ç. Pınar
Kernel ridge regression (KRR) is a standard method for performing non-parametric regression over reproducing kernel Hilbert spaces. Given
Mathematical Programming | 2015
Mert Pilanci; Martin J. Wainwright; Laurent El Ghaoui
n
international conference on acoustics, speech, and signal processing | 2009
Mert Pilanci; Orhan Arikan; Barlas Oguz; Mustafa Ç. Pınar
samples, the time and space complexity of computing the KRR estimate scale as
signal processing and communications applications conference | 2009
Mert Pilanci; Orhan Arikan; Barlas Oguz; Mustafa Ç. Pınar
\mathcal{O}(n^3)
international conference on acoustics, speech, and signal processing | 2011
Mert Pilanci; Orhan Arikan
and
signal processing and communications applications conference | 2010
Mert Pilanci; Orhan Arikan; Erdal Arikan
\mathcal{O}(n^2)