Adam J. Rothman
University of Minnesota
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Adam J. Rothman.
Electronic Journal of Statistics | 2008
Adam J. Rothman; Peter J. Bickel; Elizaveta Levina; J. Zhu
The paper proposes a method for constructing a sparse estima- tor for the inverse covariance (concentration) matrix in high-dimensional settings. The estimator uses a penalized normal likelihood approach and forces sparsity by using a lasso-type penalty. We establish a rate of con- vergence in the Frobenius norm as both data dimension p and sample size n are allowed to grow, and show that the rate depends explicitly on how sparse the true concentration matrix is. We also show that a correlation- based version of the method exhibits better rates in the operator norm. We also derive a fast iterative algorithm for computing the estimator, which relies on the popular Cholesky decomposition of the inverse but produces a permutation-invariant estimator. The method is compared to other es- timators on simulated data and on a real data example of tumor tissue classification using gene expression data.
Journal of the American Statistical Association | 2009
Adam J. Rothman; Elizaveta Levina; J. Zhu
We propose a new class of generalized thresholding operators that combine thresholding with shrinkage, and study generalized thresholding of the sample covariance matrix in high dimensions. Generalized thresholding of the covariance matrix has good theoretical properties and carries almost no computational burden. We obtain an explicit convergence rate in the operator norm that shows the tradeoff between the sparsity of the true model, dimension, and the sample size, and shows that generalized thresholding is consistent over a large class of models as long as the dimension p and the sample size n satisfy log p/n → 0. In addition, we show that generalized thresholding has the “sparsistency” property, meaning it estimates true zeros as zeros with probability tending to 1, and, under an additional mild condition, is sign consistent for nonzero elements. We show that generalized thresholding covers, as special cases, hard and soft thresholding, smoothly clipped absolute deviation, and adaptive lasso, and compare different types of generalized thresholding in a simulation study and in an example of gene clustering from a microarray experiment with tumor tissues.
The Annals of Applied Statistics | 2008
Elizaveta Levina; Adam J. Rothman; J. Zhu
The paper proposes a new covariance estimator for large covariance matrices when the variables have a natural ordering. Using the Cholesky decomposition of the inverse, we impose a banded structure on the Cholesky factor, and select the bandwidth adaptively for each row of the Cholesky factor, using a novel penalty we call nested Lasso. This structure has more flexibility than regular banding, but, unlike regular Lasso applied to the entries of the Cholesky factor, results in a sparse estimator for the inverse of the covariance matrix. An iterative algorithm for solving the optimization problem is developed. The estimator is compared to a number of other covariance estimators and is shown to do best, both in simulations and on a real data example. Simulations show that the margin by which the estimator outperforms its competitors tends to increase with dimension.
Journal of Computational and Graphical Statistics | 2010
Adam J. Rothman; Elizaveta Levina; J. Zhu
We propose a procedure for constructing a sparse estimator of a multivariate regression coefficient matrix that accounts for correlation of the response variables. This method, which we call multivariate regression with covariance estimation (MRCE), involves penalized likelihood with simultaneous estimation of the regression coefficients and the covariance structure. An efficient optimization algorithm and a fast approximation are developed for computing MRCE. Using simulation studies, we show that the proposed method outperforms relevant competitors when the responses are highly correlated. We also apply the new method to a finance example on predicting asset returns. An R-package containing this dataset and code for computing MRCE and its approximation are available online.
Annals of Statistics | 2012
R. Dennis Cook; Liliana Forzani; Adam J. Rothman
We study the asymptotic behavior of a class of methods for sufficient dimension reduction in high-dimension regressions, as the sample size and number of predictors grow in various alignments. It is demonstrated that these methods are consistent in a variety of settings, particularly in abundant regressions where most predictors contribute some information on the response, and oracle rates are possible. Simulation results are presented to support the theoretical conclusion.
Journal of Computational and Graphical Statistics | 2015
Bradley S. Price; Charles J. Geyer; Adam J. Rothman
This article proposes a penalized likelihood method to jointly estimate multiple precision matrices for use in quadratic discriminant analysis (QDA) and model-based clustering. We use a ridge penalty and a ridge fusion penalty to introduce shrinkage and promote similarity between precision matrix estimates. We use blockwise coordinate descent for optimization, and validation likelihood is used for tuning parameter selection. Our method is applied in QDA and semi-supervised model-based clustering.
Electronic Journal of Statistics | 2013
R. Dennis Cook; Liliana Forzani; Adam J. Rothman
[email protected]: An abundant regression is one in which most of the predic-tors contribute information about the response, which is contrary to thecommon notion of a sparse regression where few of the predictors are rele-vant. We discuss asymptotic characteristics of methodology for predictionin abundant linear regressions as the sample size and number of predictorsincrease in various alignments. We show that some of the estimators canperform well for the purpose of prediction in abundant high-dimensionalregressions.AMS 2000 subject classifications: Primary 62J05; secondary 62H12.Keywords and phrases: Inverse regression, least squares, Moore-Penroseinverse, sparse covariance estimation.Received January 2013.
Electronic Journal of Statistics | 2014
Adam J. Rothman; Liliana Forzani
[email protected]: We establish a necessary and sufficient condition for the exis-tence of the precision matrix estimator obtained by minimizing the nega-tive Gaussian log-likelihood plus a weighted bridge penalty. This conditionenables us to connect the literature on Gaussian graphical models to theliterature on penalized Gaussian likelihood.AMS 2000 subject classifications: Primary 62H12; secondary 62H20.Keywords and phrases: High-dimensional data, precision matrix, ridgepenalty, sparsity.Received December 2014.
Journal of Computational and Graphical Statistics | 2018
Aaron J. Molstad; Adam J. Rothman
ABSTRACT We propose a penalized likelihood method to fit the linear discriminant analysis model when the predictor is matrix valued. We simultaneously estimate the means and the precision matrix, which we assume has a Kronecker product decomposition. Our penalties encourage pairs of response category mean matrix estimators to have equal entries and also encourage zeros in the precision matrix estimator. To compute our estimators, we use a blockwise coordinate descent algorithm. To update the optimization variables corresponding to response category mean matrices, we use an alternating minimization algorithm that takes advantage of the Kronecker structure of the precision matrix. We show that our method can outperform relevant competitors in classification, even when our modeling assumptions are violated. We analyze three real datasets to demonstrate our method’s applicability. Supplementary materials, including an R package implementing our method, are available online.
Biometrika | 2018
Aaron J. Molstad; Adam J. Rothman
Summary We propose a framework to shrink a user‐specified characteristic of a precision matrix estimator that is needed to fit a predictive model. Estimators in our framework minimize the Gaussian negative loglikelihood plus an L1 penalty on a linear or affine function evaluated at the optimization variable corresponding to the precision matrix. We establish convergence rate bounds for these estimators and propose an alternating direction method of multipliers algorithm for their computation. Our simulation studies show that our estimators can perform better than competitors when they are used to fit predictive models. In particular, we illustrate cases where our precision matrix estimators perform worse at estimating the population precision matrix but better at prediction.