Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Marten H. Wegkamp is active.

Publication


Featured researches published by Marten H. Wegkamp.


Electronic Journal of Statistics | 2007

Sparsity oracle inequalities for the Lasso

Florentina Bunea; Alexandre B. Tsybakov; Marten H. Wegkamp

This paper studies oracle properties of l1-penalized least squares in nonparametric regression setting with random design. We show that the penalized least squares estimator satisfies sparsity oracle inequalities, i.e., bounds in terms of the number of non-zero components of the oracle vec- tor. The results are valid even when the dimension of the model is (much) larger than the sample size and the regression matrix is not positive definite. They can be applied to high-dimensional linear regression, to nonparamet- ric adaptive regression estimation and to the problem of aggregation of arbitrary estimators. AMS 2000 subject classifications: Primary 62G08; secondary 62C20, 62G05, 62G20. Keywords and phrases: sparsity, oracle inequalities, Lasso, penalized least squares, nonparametric regression, dimension reduction, aggregation, mutual coherence, adaptive estimation.


Annals of Statistics | 2011

Optimal selection of reduced rank estimators of high-dimensional matrices

Florentina Bunea; Yiyuan She; Marten H. Wegkamp

We introduce a new criterion, the Rank Selection Criterion (RSC), for selecting the optimal reduced rank estimator of the coefficient matrix in multivariate response regression models. The corresponding RSC estimator minimizes the Frobenius norm of the fit plus a regularization term proportional to the number of parameters in the reduced rank model. The rank of the RSC estimator provides a consistent estimator of the rank of the coefficient matrix; in general, the rank of our estimator is a consistent estimate of the effective rank, which we define to be the number of singular values of the target matrix that are appropriately large. The consistency results are valid not only in the classic asymptotic regime, when n, the number of responses, and p, the number of predictors, stay bounded, and m, the number of observations, grows, but also when either, or both, n and p grow, possibly much faster than m. We establish minimax optimal bounds on the mean squared errors of our estimators. Our finite sample performance bounds for the RSC estimator show that it achieves the optimal balance between the approximation error and the penalty term. Furthermore, our procedure has very low computational complexity, linear in the number of candidate models, making it particularly appealing for large scale problems. We contrast our estimator with the nuclear norm penalized least squares (NNP) estimator, which has an inherently higher computational complexity than RSC, for multivariate regression models. We show that NNP has estimation properties similar to those of RSC, albeit under stronger conditions. However, it is not as parsimonious as RSC. We offer a simple correction of the NNP estimator which leads to consistent rank estimation. We verify and illustrate our theoretical findings via an extensive simulation study.


IEEE Transactions on Information Theory | 2005

Functional classification in Hilbert spaces

Gérard Biau; Florentina Bunea; Marten H. Wegkamp

Let X be a random variable taking values in a separable Hilbert space X, with label Y/spl isin/{0,1}. We establish universal weak consistency of a nearest neighbor-type classifier based on n independent copies (X/sub i/,Y/sub i/) of the pair (X,Y), extending the classical result of Stone to infinite-dimensional Hilbert spaces. Under a mild condition on the distribution of X, we also prove strong consistency. We reduce the infinite dimension of X by considering only the first d coefficients of a Fourier series expansion of each X/sub i/, and then we perform k-nearest neighbor classification in /spl Ropf//sup d/. Both the dimension and the number of neighbors are automatically selected from the data using a simple data-splitting device. An application of this technique to a signal discrimination problem involving speech recordings is presented.


conference on learning theory | 2006

Aggregation and sparsity via ℓ 1 penalized least squares

Florentina Bunea; Alexandre B. Tsybakov; Marten H. Wegkamp

This paper shows that near optimal rates of aggregation and adaptation to unknown sparsity can be simultaneously achieved via l1 penalized least squares in a nonparametric regression setting. The main tool is a novel oracle inequality on the sum between the empirical squared loss of the penalized least squares estimate and a term reflecting the sparsity of the unknown regression function.


Annals of Statistics | 2012

Joint variable and rank selection for parsimonious estimation of high-dimensional matrices

Florentina Bunea; Yiyuan She; Marten H. Wegkamp

We propose dimension reduction methods for sparse, high-dimensional multivariate response regression models. Both the number of responses and that of the predictors may exceed the sample size. Sometimes viewed as complementary, predictor selection and rank reduction are the most popular strategies for obtaining lower-dimensional approximations of the parameter matrix in such models. We show in this article that important gains in prediction accuracy can be obtained by considering them jointly. We motivate a new class of sparse multivariate regression models, in which the coefficient matrix has low rank and zero rows or can be well approximated by such a matrix. Next, we introduce estimators that are based on penalized least squares, with novel penalties that impose simultaneous row and rank restrictions on the coefficient matrix. We prove that these estimators indeed adapt to the unknown matrix sparsity and have fast rates of convergence. We support our theoretical results with an extensive simulation study and two data analyses.


Annals of Statistics | 2004

Complexity regularization via localized random penalties

Gábor Lugosi; Marten H. Wegkamp

In this article, model selection via penalized empirical loss minimization in nonparametric classification problems is studied. Datadependent penalties are constructed, which are based on estimates of the complexity of a small subclass of each model class, containing only those functions with small empirical loss. The penalties are novel since those considered in the literature are typically based on the entire model class. Oracle inequalities using these penalties are established, and the advantage of the new penalties over those based on the complexity of the whole model class is demonstrated. 1. Introduction. In this article, we propose a new complexity-penalized model selection method based on data-dependent penalties. We consider the binary classification problem where, given a random observation X ∈ R d , one has to predict Y ∈ {0,1}. A classifier or classification rule is a function f : R d → {0,1}, with loss


conference on learning theory | 2007

Sparse density estimation with l 1 penalties

Florentina Bunea; Alexandre B. Tsybakov; Marten H. Wegkamp

This paper studies oracle properties of l1-penalized estimators of a probability density. We show that the penalized least squares estimator satisfies sparsity oracle inequalities, i.e., bounds in terms of the number of non-zero components of the oracle vector. The results are valid even when the dimension of the model is (much) larger than the sample size. They are applied to estimation in sparse high-dimensional mixture models, to nonparametric adaptive density estimation and to the problem of aggregation of density estimators.


Journal of Multivariate Analysis | 2012

Time-dependent copulas

Jean-David Fermanian; Marten H. Wegkamp

For the study of dynamic dependence structures, the authors introduce the concept of a pseudo-copula, which extends Pattons definition of a conditional copula. They state the equivalent of Sklars theorem for pseudo-copulas. They establish the asymptotic normality of nonparametric estimators of pseudo-copulas under strong mixing assumptions, and discuss applications to specification tests. They complement the theory with a small simulation study on the power of the proposed tests.


Annals of Statistics | 2010

SPADES and mixture models

Florentina Bunea; Alexandre B. Tsybakov; Marten H. Wegkamp; Adrian Barbu

This paper studies sparse density estimation via l1 penalization (SPADES). We focus on estimation in high-dimensional mixture models and nonparametric adaptive den- sity estimation. We show, respectively, that SPADES can recover, with high probability, the unknown components of a mixture of probability densities and that it yields minimax adaptive density estimates. These results are based on a general sparsity oracle inequality that the SPADES estimates satisfy. MSC2000 Subject classification: Primary 62G08, Secondary 62C20, 62G05, 62G20


Econometrica | 2002

Weighted Minimum Mean-Square Distance from Independence Estimation

Donald J. Brown; Marten H. Wegkamp

In this paper we introduce a family of semi-parametric estimators, suggested by Manskis minimum mean-square distance from independence estimator. We establish the strong consistency, asymptotic normality and consistency of bootstrap estimates of the sampling distribution and the asymptotic variance of these estimators.

Collaboration


Dive into the Marten H. Wegkamp's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Dragan Radulovic

Florida Atlantic University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Yiyuan She

Florida State University

View shared research outputs
Top Co-Authors

Avatar

Ming Yuan

University of Wisconsin-Madison

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Adrian Barbu

Florida State University

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge