Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Andre Wibisono is active.

Publication


Featured researches published by Andre Wibisono.


Proceedings of the National Academy of Sciences of the United States of America | 2016

A Variational Perspective on Accelerated Methods in Optimization

Andre Wibisono; Ashia C. Wilson; Michael I. Jordan

Significance Optimization problems arise naturally in statistical machine learning and other fields concerned with data analysis. The rapid growth in the scale and complexity of modern datasets has led to a focus on gradient-based methods and also on the class of accelerated methods, first proposed by Nesterov in 1983. Accelerated methods achieve faster convergence rates than gradient methods and indeed, under certain conditions, they achieve optimal rates. However, accelerated methods are not descent methods and remain a conceptual mystery. We propose a variational, continuous-time framework for understanding accelerated methods. We provide a systematic methodology for converting accelerated higher-order methods from continuous time to discrete time. Our work illuminates a class of dynamics that may be useful for designing better algorithms for optimization. Accelerated gradient methods play a central role in optimization, achieving optimal rates in many settings. Although many generalizations and extensions of Nesterov’s original acceleration method have been proposed, it is not yet clear what is the natural scope of the acceleration concept. In this paper, we study accelerated methods from a continuous-time perspective. We show that there is a Lagrangian functional that we call the Bregman Lagrangian, which generates a large class of accelerated methods in continuous time, including (but not limited to) accelerated gradient descent, its non-Euclidean extension, and accelerated higher-order gradient methods. We show that the continuous-time limit of all of these methods corresponds to traveling the same curve in spacetime at different speeds. From this perspective, Nesterov’s technique and many of its generalizations can be viewed as a systematic way to go from the continuous-time curves generated by the Bregman Lagrangian to a family of discrete-time accelerated algorithms.


IEEE Transactions on Information Theory | 2015

Optimal Rates for Zero-Order Convex Optimization: The Power of Two Function Evaluations

John C. Duchi; Michael I. Jordan; Martin J. Wainwright; Andre Wibisono

We consider derivative-free algorithms for stochastic and nonstochastic convex optimization problems that use only function values rather than gradients. Focusing on nonasymptotic bounds on convergence rates, we show that if pairs of function values are available, algorithms for d-dimensional optimization that use gradient estimates based on random perturbations suffer a factor of at most √d in convergence rate over traditional stochastic gradient methods. We establish such results for both smooth and nonsmooth cases, sharpening previous analyses that suggested a worse dimension dependence, and extend our results to the case of multiple (m ≥ 2) evaluations. We complement our algorithmic development with information-theoretic lower bounds on the minimax convergence rate of such problems, establishing the sharpness of our achievable results up to constant (sometimes logarithmic) factors.


international symposium on information theory | 2017

Information and estimation in Fokker-Planck channels

Andre Wibisono; Varun Jog; Po-Ling Loh

We study the relationship between information- and estimation-theoretic quantities in time-evolving systems. We focus on the Fokker-Planck channel defined by a general stochastic differential equation, and show that the time derivatives of entropy, KL divergence, and mutual information are characterized by estimation-theoretic quantities involving an appropriate generalization of the Fisher information. Our results vastly extend De Bruijns identity and the classical I-MMSE relation.


neural information processing systems | 2013

Streaming Variational Bayes

Tamara Broderick; Nicholas Boyd; Andre Wibisono; Ashia C. Wilson; Michael I. Jordan


arXiv: Statistics Theory | 2013

Maximum Entropy Distributions on Graphs

Christopher J. Hillar; Andre Wibisono


symposium on the theory of computing | 2012

Minimax option pricing meets black-scholes in the limit

Jacob D. Abernethy; Rafael M. Frongillo; Andre Wibisono


Science & Engineering Faculty | 2013

How to Hedge an Option Against an Adversary: Black-Scholes Pricing is Minimax Optimal

Jacob D. Abernethy; Peter L. Bartlett; Rafael M. Frongillo; Andre Wibisono


neural information processing systems | 2012

Finite Sample Convergence Rates of Zero-Order Stochastic Optimization Methods

Andre Wibisono; Martin J. Wainwright; Michael I. Jordan; John C. Duchi


arXiv: Optimization and Control | 2015

On Accelerated Methods in Optimization

Andre Wibisono; Ashia C. Wilson


arXiv: Functional Analysis | 2012

Inverses of symmetric, diagonally dominant positive matrices and applications

Christopher J. Hillar; Shaowei Lin; Andre Wibisono

Collaboration


Dive into the Andre Wibisono's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Po-Ling Loh

University of Wisconsin-Madison

View shared research outputs
Top Co-Authors

Avatar

Rafael M. Frongillo

University of Colorado Boulder

View shared research outputs
Top Co-Authors

Avatar

Shaowei Lin

University of California

View shared research outputs
Top Co-Authors

Avatar

Varun Jog

The Chinese University of Hong Kong

View shared research outputs
Researchain Logo
Decentralizing Knowledge