Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Yingying Fan is active.

Publication


Featured researches published by Yingying Fan.


Annals of Statistics | 2008

High-dimensional classification using features annealed independence rules

Jianqing Fan; Yingying Fan

Classification using high-dimensional features arises frequently in many contemporary statistical studies such as tumor classification using microarray or other high-throughput data. The impact of dimensionality on classifications is largely poorly understood. In a seminal paper, Bickel and Levina (2004) show that the Fisher discriminant performs poorly due to diverging spectra and they propose to use the independence rule to overcome the problem. We first demonstrate that even for the independence classification rule, classification using all the features can be as bad as the random guessing due to noise accumulation in estimating population centroids in high-dimensional feature space. In fact, we demonstrate further that almost all linear discriminants can perform as bad as the random guessing. Thus, it is paramountly important to select a subset of important features for high-dimensional classification, resulting in Features Annealed Independence Rules (FAIR). The conditions under which all the important features can be selected by the two-sample t-statistic are established. The choice of the optimal number of features, or equivalently, the threshold value of the test statistics are proposed based on an upper bound of the classification error. Simulation studies and real data analysis support our theoretical results and demonstrate convincingly the advantage of our new classification procedure.


Annals of Statistics | 2009

A unified approach to model selection and sparse recovery using regularized least squares

Jinchi Lv; Yingying Fan

Model selection and sparse recovery are two important problems for which many regularization methods have been proposed. We study the properties of regularization methods in both problems under the unified framework of regularized least squares with concave penalties. For model selection, we establish conditions under which a regularized least squares estimator enjoys a nonasymptotic property, called the weak oracle property, where the dimensionality can grow exponentially with sample size. For sparse recovery, we present a sufficient condition that ensures the recoverability of the sparsest solution. In particular, we approach both problems by considering a family of penalties that give a smooth homotopy between L 0 and L 1 penalties. We also propose the sequentially and iteratively reweighted squares (SIRS) algorithm for sparse recovery. Numerical studies support our theoretical results and demonstrate the advantage of our new methods for model selection and sparse recovery.


Journal of The Royal Statistical Society Series B-statistical Methodology | 2013

Tuning parameter selection in high dimensional penalized likelihood

Yingying Fan; Cheng Yong Tang

Determining how to appropriately select the tuning parameter is essential in penalized likelihood methods for high-dimensional data analysis. We examine this problem in the setting of penalized likelihood methods for generalized linear models, where the dimensionality of covariates p is allowed to increase exponentially with the sample size n. We propose to select the tuning parameter by optimizing the generalized information criterion (GIC) with an appropriate model complexity penalty. To ensure that we consistently identify the true model, a range for the model complexity penalty is identified in GIC. We find that this model complexity penalty should diverge at the rate of some power of


Annals of Statistics | 2014

Adaptive robust variable selection

Jianqing Fan; Yingying Fan; Emre Barut

\log p


Annals of Statistics | 2015

FUNCTIONAL ADDITIVE REGRESSION

Yingying Fan; Gareth M. James; Peter Radchenko

depending on the tail probability behavior of the response variables. This reveals that using the AIC or BIC to select the tuning parameter may not be adequate for consistently identifying the true model. Based on our theoretical study, we propose a uniform choice of the model complexity penalty and show that the proposed approach consistently identifies the true model among candidate models with asymptotic probability one. We justify the performance of the proposed procedure by numerical simulations and a gene expression data analysis.


Annals of Statistics | 2012

VARIABLE SELECTION IN LINEAR MIXED EFFECTS MODELS

Yingying Fan; Runze Li

Heavy-tailed high-dimensional data are commonly encountered in various scientific fields and pose great challenges to modern statistical analysis. A natural procedure to address this problem is to use penalized quantile regression with weighted L1-penalty, called weighted robust Lasso (WR-Lasso), in which weights are introduced to ameliorate the bias problem induced by the L1-penalty. In the ultra-high dimensional setting, where the dimensionality can grow exponentially with the sample size, we investigate the model selection oracle property and establish the asymptotic normality of the WR-Lasso. We show that only mild conditions on the model error distribution are needed. Our theoretical results also reveal that adaptive choice of the weight vector is essential for the WR-Lasso to enjoy these nice asymptotic properties. To make the WR-Lasso practically feasible, we propose a two-step procedure, called adaptive robust Lasso (AR-Lasso), in which the weight vector in the second step is constructed based on the L1-penalized quantile regression estimate from the first step. This two-step procedure is justified theoretically to possess the oracle property and the asymptotic normality. Numerical studies demonstrate the favorable finite-sample performance of the AR-Lasso.


Journal of the American Statistical Association | 2007

Dynamic Integration of Time- and State-Domain Methods for Volatility Estimation

Jianqing Fan; Yingying Fan; Jiancheng Jiang

We suggest a new method, called Functional Additive Regression, or FAR, for efficiently performing high-dimensional functional regression. FAR extends the usual linear regression model involving a functional predictor,


Journal of the American Statistical Association | 2013

Asymptotic Equivalence of Regularization Methods in Thresholded Parameter Space

Yingying Fan; Jinchi Lv

X(t)


Annals of Statistics | 2013

OPTIMAL CLASSIFICATION IN SPARSE GAUSSIAN GRAPHIC MODEL

Yingying Fan; Jiashun Jin; Zhigang Yao

, and a scalar response,


Biometrika | 2014

Asymptotic properties for combined L1 and concave regularization

Yingying Fan; Jinchi Lv

Y

Collaboration


Dive into the Yingying Fan's collaboration.

Top Co-Authors

Avatar

Jinchi Lv

University of Southern California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Yinfei Kong

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Gareth M. James

University of Southern California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Zemin Zheng

University of Science and Technology of China

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Christian M. Hafner

Université catholique de Louvain

View shared research outputs
Top Co-Authors

Avatar

Marc Hallin

Université libre de Bruxelles

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge