Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Weijie Su is active.

Publication


Featured researches published by Weijie Su.


The Annals of Applied Statistics | 2015

SLOPE { Adaptive Variable Selection via Convex Optimization

Małgorzata Bogdan; Ewout van den Berg; Chiara Sabatti; Weijie Su; Emmanuel J. Candès

We introduce a new estimator for the vector of coefficients β in the linear model y = Xβ + z, where X has dimensions n × p with p possibly larger than n. SLOPE, short for Sorted L-One Penalized Estimation, is the solution to [Formula: see text]where λ1 ≥ λ2 ≥ … ≥ λ p ≥ 0 and [Formula: see text] are the decreasing absolute values of the entries of b. This is a convex program and we demonstrate a solution algorithm whose computational complexity is roughly comparable to that of classical ℓ1 procedures such as the Lasso. Here, the regularizer is a sorted ℓ1 norm, which penalizes the regression coefficients according to their rank: the higher the rank-that is, stronger the signal-the larger the penalty. This is similar to the Benjamini and Hochberg [J. Roy. Statist. Soc. Ser. B57 (1995) 289-300] procedure (BH) which compares more significant p-values with more stringent thresholds. One notable choice of the sequence {λ i } is given by the BH critical values [Formula: see text], where q ∈ (0, 1) and z(α) is the quantile of a standard normal distribution. SLOPE aims to provide finite sample guarantees on the selected model; of special interest is the false discovery rate (FDR), defined as the expected proportion of irrelevant regressors among all selected predictors. Under orthogonal designs, SLOPE with λBH provably controls FDR at level q. Moreover, it also appears to have appreciable inferential properties under more general designs X while having substantial power, as demonstrated in a series of experiments running on both simulated and real data.


Annals of Statistics | 2016

SLOPE is Adaptive to Unknown Sparsity and Asymptotically Minimax

Weijie Su; Emmanuel J. Candès

We consider high-dimensional sparse regression problems in which we observe


Annals of Statistics | 2017

False discoveries occur early on the Lasso path

Weijie Su; Małgorzata Bogdan; Emmanuel J. Candès

y = X \beta + z


Electronic Journal of Statistics | 2016

Familywise error rate control via knockoffs

Lucas Janson; Weijie Su

, where


Journal of the American Statistical Association | 2018

Group SLOPE - adaptive selection of groups of predictors

Damian Brzyski; Alexej Gossmann; Weijie Su; Małgorzata Bogdan

X


Journal of the American Statistical Association | 2018

Multiple testing when many p-values are uniformly conservative, with application to testing qualitative interaction in educational interventions

Qingyuan Zhao; Dylan S. Small; Weijie Su

is an


arXiv: Methodology | 2013

Statistical estimation and testing via the sorted L1 norm

Małgorzata Bogdan; Ewout van den Berg; Weijie Su; Emmanuel J. Candès

n \times p


arXiv: Statistics Theory | 2015

Private False Discovery Rate Control

Cynthia Dwork; Weijie Su; Li Zhang

design matrix and


Journal of Machine Learning Research | 2016

A differential equation for modeling Nesterov's accelerated gradient method: theory and insights

Weijie Su; Stephen P. Boyd; Emmanuel J. Candès

z


arXiv: Probability | 2017

On the global convergence of a randomly perturbed dissipative nonlinear oscillator.

Wenqing Hu; Chris Junchi Li; Weijie Su

is an

Collaboration


Dive into the Weijie Su's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Dylan S. Small

University of Pennsylvania

View shared research outputs
Top Co-Authors

Avatar

Jingshu Wang

University of Pennsylvania

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Qingyuan Zhao

University of Pennsylvania

View shared research outputs
Researchain Logo
Decentralizing Knowledge