Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Rong Ge is active.

Publication


Featured researches published by Rong Ge.


symposium on the theory of computing | 2012

Computing a nonnegative matrix factorization -- provably

Sanjeev Arora; Rong Ge; Ravindran Kannan; Ankur Moitra

The Nonnegative Matrix Factorization (NMF) problem has a rich history spanning quantum mechanics, probability theory, data analysis, polyhedral combinatorics, communication complexity, demography, chemometrics, etc. In the past decade NMF has become enormously popular in machine learning, where the factorization is computed using a variety of local search heuristics. Vavasis recently proved that this problem is NP-complete. We initiate a study of when this problem is solvable in polynomial time. Consider a nonnegative m x n matrix


foundations of computer science | 2012

Learning Topic Models -- Going beyond SVD

Sanjeev Arora; Rong Ge; Ankur Moitra

M


international colloquium on automata languages and programming | 2011

New algorithms for learning in presence of errors

Sanjeev Arora; Rong Ge

and a target inner-dimension r. Our results are the following: - We give a polynomial-time algorithm for exact and approximate NMF for every constant r. Indeed NMF is most interesting in applications precisely when r is small. We complement this with a hardness result, that if exact NMF can be solved in time (nm)o(r), 3-SAT has a sub-exponential time algorithm. Hence, substantial improvements to the above algorithm are unlikely. - We give an algorithm that runs in time polynomial in n, m and r under the separablity condition identified by Donoho and Stodden in 2003. The algorithm may be practical since it is simple and noise tolerant (under benign assumptions). Separability is believed to hold in many practical settings. To the best of our knowledge, this last result is the first polynomial-time algorithm that provably works under a non-trivial condition on the input matrix and we believe that this will be an interesting and important direction for future work.


Communications of The ACM | 2011

Computational complexity and information asymmetry in financial products

Sanjeev Arora; Boaz Barak; Markus K. Brunnermeier; Rong Ge

Topic Modeling is an approach used for automatic comprehension and classification of data in a variety of settings, and perhaps the canonical application is in uncovering thematic structure in a corpus of documents. A number of foundational works both in machine learning and in theory have suggested a probabilistic model for documents, whereby documents arise as a convex combination of (i.e. distribution on) a small number of topic vectors, each topic vector being a distribution on words (i.e. a vector of word-frequencies). Similar models have since been used in a variety of application areas, the Latent Dirichlet Allocation or LDA model of Blei et al. is especially popular. Theoretical studies of topic modeling focus on learning the models parameters assuming the data is actually generated from it. Existing approaches for the most part rely on Singular Value Decomposition (SVD), and consequently have one of two limitations: these works need to either assume that each document contains only one topic, or else can only recover the {\em span} of the topic vectors instead of the topic vectors themselves. This paper formally justifies Nonnegative Matrix Factorization (NMF) as a main tool in this context, which is an analog of SVD where all vectors are nonnegative. Using this tool we give the first polynomial-time algorithm for learning topic models without the above two limitations. The algorithm uses a fairly mild assumption about the underlying topic matrix called separability, which is usually found to hold in real-life data. Perhaps the most attractive feature of our algorithm is that it generalizes to yet more realistic models that incorporate topic-topic correlations, such as the Correlated Topic Model (CTM) and the Pachinko Allocation Model (PAM). We hope that this paper will motivate further theoretical results that use NMF as a replacement for SVD -- just as NMF has come to replace SVD in many applications.


neural information processing systems | 2012

Provable ICA with Unknown Gaussian Noise, with Implications for Gaussian Mixtures and Autoencoders

Sanjeev Arora; Rong Ge; Ankur Moitra; Sushant Sachdeva

We give new algorithms for a variety of randomly-generated instances of computational problems using a linearization technique that reduces to solving a system of linear equations. These algorithms are derived in the context of learning with structured noise, a notion introduced in this paper. This notion is best illustrated with the learning parities with noise (LPN) problem --well-studied in learning theory and cryptography. In the standard version, we have access to an oracle that, each time we press a button, returns a random vector a ∈ GF(2)n together with a bit b ∈ GF(2) that was computed as aċu+η, where u ∈ GF(2)n is a secret vector, and η ∈ GF(2) is a noise bit that is 1 with some probability p. Say p = 1/3. The goal is to recover u. This task is conjectured to be intractable. In the structured noise setting we introduce a slight (?) variation of the model: upon pressing a button, we receive (say) 10 random vectors a1, a2, ... , a10 ∈ GF(2)n, and corresponding bits b1, b2, ... , b10, of which at most 3 are noisy. The oracle may arbitrarily decide which of the 10 bits to make noisy. We exhibit a polynomial-time algorithm to recover the secret vector u given such an oracle. We think this structured noise model may be of independent interest in machine learning. We discuss generalizations of our result, including learning with more general noise patterns. We also give the first nontrivial algorithms for two problems, which we show fit in our structured noise framework. We give a slightly subexponential algorithm for the well-known learning with errors (LWE) problem over GF(q) introduced by Regev for cryptographic uses. Our algorithm works for the case when the gaussian noise is small; which was an open problem. Our result also clarifies why existing hardness results fail at this particular noise rate. We also give polynomial-time algorithms for learning the MAJORITY OF PARITIES function of Applebaumet al. for certain parameter values. This function is a special case of Goldreichs pseudorandom generator. The full version is available at http://www.eccc.uni-trier.de/report /2010/066/.


international workshop and international workshop on approximation randomization and combinatorial optimization algorithms and techniques | 2015

Decomposing Overcomplete 3rd Order Tensors using Sum-of-Squares Algorithms

Rong Ge; Tengyu Ma

This paper introduces notions from computational complexity into the study of financial derivatives. Traditional economics argues that derivatives, like CDOs and CDSs, ameliorate the negative costs imposed due to asymmetric information between buyers and sellers. This is because securitization via these derivatives allows the informed party to find buyers for the information-insensitive part of the cash flow stream of an asset (e.g., a mortgage) and retain the remainder. In this paper we show that this viewpoint may need to be revised once computational complexity is brought into the picture. Assuming reasonable complexity-theoretic conjectures, we show that derivatives can actually amplify the costs of asymmetric information instead of reducing them. We prove our results both in the worst-case setting, as well as the more realistic average case setting. In the latter case, to argue that our constructions result in derivatives that “look like” real-life derivatives, we use the notion of computational indistinguishability a la cryptography.


SIAM Journal on Computing | 2016

Computing a Nonnegative Matrix Factorization---Provably

Sanjeev Arora; Rong Ge; Ravi Kannan; Ankur Moitra

We present a new algorithm for independent component analysis which has provable performance guarantees. In particular, suppose we are given samples of the form


algorithmic learning theory | 2015

Tensor Decompositions for Learning Latent Variable Models A Survey for ALT

Animashree Anandkumar; Rong Ge; Daniel J. Hsu; Sham M. Kakade; Matus Telgarsky


foundations of computer science | 2013

Towards a Better Approximation for Sparsest Cut

Sanjeev Arora; Rong Ge; Ali Kemal Sinop

y = Ax + \eta


international symposium on algorithms and computation | 2009

New Results on Simple Stochastic Games

Decheng Dai; Rong Ge

Collaboration


Dive into the Rong Ge's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Sham M. Kakade

University of Washington

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Ankur Moitra

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Chi Jin

University of California

View shared research outputs
Top Co-Authors

Avatar

Majid Janzamin

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge