Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Christos Boutsidis is active.

Publication


Featured researches published by Christos Boutsidis.


Pattern Recognition | 2008

SVD based initialization: A head start for nonnegative matrix factorization

Christos Boutsidis; Efstratios Gallopoulos

We describe Nonnegative Double Singular Value Decomposition (NNDSVD), a new method designed to enhance the initialization stage of nonnegative matrix factorization (NMF). NNDSVD can readily be combined with existing NMF algorithms. The basic algorithm contains no randomization and is based on two SVD processes, one approximating the data matrix, the other approximating positive sections of the resulting partial SVD factors utilizing an algebraic property of unit rank matrices. Simple practical variants for NMF with dense factors are described. NNDSVD is also well suited to initialize NMF algorithms with sparse factors. Many numerical examples suggest that NNDSVD leads to rapid reduction of the approximation error of many NMF algorithms.


SIAM Journal on Computing | 2014

Near-Optimal Column-Based Matrix Reconstruction

Christos Boutsidis; Petros Drineas; Malik Magdon-Ismail

We consider low-rank reconstruction of a matrix using a subset of its columns and present asymptotically optimal algorithms for both spectral norm and Frobenius norm reconstruction. The main tools ...


IEEE Transactions on Information Theory | 2015

Randomized Dimensionality Reduction for

Christos Boutsidis; Anastasios Zouzias; Michael W. Mahoney; Petros Drineas

We study the topic of dimensionality reduction for k-means clustering. Dimensionality reduction encompasses the union of two approaches: 1) feature selection and 2) feature extraction. A feature selection-based algorithm for k-means clustering selects a small subset of the input features and then applies k-means clustering on the selected features. A feature extraction-based algorithm for k-means clustering constructs a small set of new artificial features and then applies k-means clustering on the constructed features. Despite the significance of k-means clustering as well as the wealth of heuristic methods addressing it, provably accurate feature selection methods for k-means clustering are not known. On the other hand, two provably accurate feature extraction methods for k-means clustering are known in the literature; one is based on random projections and the other is based on the singular value decomposition (SVD). This paper makes further progress toward a better understanding of dimensionality reduction for k-means clustering. Namely, we present the first provably accurate feature selection method for k-means clustering and, in addition, we present two feature extraction methods. The first feature extraction method is based on random projections and it improves upon the existing results in terms of time complexity and number of features needed to be extracted. The second feature extraction method is based on fast approximate SVD factorizations and it also improves upon the existing results in terms of time complexity. The proposed algorithms are randomized and provide constant-factor approximation guarantees with respect to the optimal k-means objective value.


foundations of computer science | 2011

k

Christos Boutsidis; Petros Drineas; Malik Magdon-Ismail

We consider low-rank reconstruction of a matrix using a subset of its columns and we present asymptotically optimal algorithms for both spectral norm and Frobenius norm reconstruction. The main tools we introduce to obtain our results are: (i) the use of fast approximate SVD-like decompositions for column-based matrix reconstruction, and (ii) two deterministic algorithms for selecting rows from matrices with orthonormal columns, building upon the sparse representation theorem for decompositions of the identity that appeared in [1].


SIAM Journal on Matrix Analysis and Applications | 2013

-Means Clustering

Christos Boutsidis; Alex Gittens

Several recent randomized linear algebra algorithms rely upon fast dimension reduction methods. A popular choice is the subsampled randomized Hadamard transform (SRHT). In this article, we address the efficacy, in the Frobenius and spectral norms, of an SRHT-based low-rank matrix approximation technique introduced by Woolfe, Liberty, Rohklin, and Tygert. We establish a slightly better Frobenius norm error bound than is currently available, and a much sharper spectral norm error bound (in the presence of reasonable decay of the singular values). Along the way, we produce several results on matrix operations with SRHTs (such as approximate matrix multiplication) that may be of independent interest. Our approach builds upon Tropps in “Improved Analysis of the Subsampled Randomized Hadamard Transform” [Adv. Adaptive Data Anal., 3 (2011), pp. 115--126].


symposium on the theory of computing | 2014

Near Optimal Column-Based Matrix Reconstruction

Christos Boutsidis; David P. Woodruff

The CUR decomposition of an m×n matrix A finds an m×c matrix C with a small subset of c < n columns of A, together with an r×n matrix R with a small subset of r < m rows of A, as well as a c×r low rank matrix U such that the matrix CUR approximates the input matrix A, that is, ||A --- CUR||2F ≤ (1 + ε)||A --- Ak||2F, where ||.||F denotes the Frobenius norm, 0 < ε < 1 is an accuracy parameter, and Ak is the best m × n matrix of rank k constructed via the SVD of A. We present input-sparsity-time and deterministic algorithms for constructing such a CUR matrix decomposition of A where c = O(k/ε) and r = O(k/ε) and rank(U) = k. Up to constant factors, our construction is simultaneously optimal in c, r, and rank(U).


IEEE Transactions on Information Theory | 2013

Improved matrix algorithms via the Subsampled Randomized Hadamard Transform

Christos Boutsidis; Malik Magdon-Ismail

We study feature selection for k-means clustering. Although the literature contains many methods with good empirical performance, algorithms with provable theoretical behavior have only recently been developed. Unfortunately, these algorithms are randomized and fail with, say, a constant probability. We present the first deterministic feature selection algorithm for k-means clustering with relative error guarantees. At the heart of our algorithm lies a deterministic method for decompositions of the identity and a structural result which quantifies some of the tradeoffs in dimensionality reduction.


SIAM Journal on Matrix Analysis and Applications | 2013

Optimal CUR matrix decompositions

Haim Avron; Christos Boutsidis

We study the following problem of subset selection for matrices: given a matrix


ACM Transactions on Knowledge Discovery From Data | 2014

Deterministic Feature Selection for

Saurabh Paul; Christos Boutsidis; Malik Magdon-Ismail; Petros Drineas

\mathbf{X} \in \mathbb{R}^{n \times m}


IEEE Transactions on Information Theory | 2013

k

Christos Boutsidis; Petros Drineas; Malik Magdon-Ismail

(

Collaboration


Dive into the Christos Boutsidis's collaboration.

Top Co-Authors

Avatar

Petros Drineas

Rensselaer Polytechnic Institute

View shared research outputs
Top Co-Authors

Avatar

Malik Magdon-Ismail

Rensselaer Polytechnic Institute

View shared research outputs
Top Co-Authors

Avatar

David P. Woodruff

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Alex Gittens

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Saurabh Paul

Rensselaer Polytechnic Institute

View shared research outputs
Researchain Logo
Decentralizing Knowledge