Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Lizhen Lin is active.

Publication


Featured researches published by Lizhen Lin.


Computational Statistics & Data Analysis | 2013

Recent progress in the nonparametric estimation of monotone curves-With applications to bioassay and environmental risk assessment

Rabi Bhattacharya; Lizhen Lin

Three recent nonparametric methodologies for estimating a monotone regression function F and its inverse F-1 are (1) the inverse kernel method DNP (Dette et al. (2005), Dette and Scheder (2010)), (2) the monotone spline (Kong and Eubank (2006)) and (3) the data adaptive method NAM (Bhattacharya and Lin (2010), (2011)), with roots in isotonic regression (Ayer et al. (1955), Bhattacharya and Kong (2007)). All three have asymptotically optimal error rates. In this article their finite sample performances are compared using extensive simulation from diverse models of interest, and by analysis of real data. Let there be m distinct values of the independent variable x among N observations y. The results show that if m is relatively small compared to N then generally the NAM performs best, while the DNP outperforms the other methods when m is O(N) unless there is a substantial clustering of the values of the independent variable x.


Journal of the American Statistical Association | 2017

Extrinsic Local Regression on Manifold-Valued Data

Lizhen Lin; Brian St. Thomas; Hongtu Zhu; David B. Dunson

ABSTRACT We propose an extrinsic regression framework for modeling data with manifold valued responses and Euclidean predictors. Regression with manifold responses has wide applications in shape analysis, neuroscience, medical imaging, and many other areas. Our approach embeds the manifold where the responses lie onto a higher dimensional Euclidean space, obtains a local regression estimate in that space, and then projects this estimate back onto the image of the manifold. Outside the regression setting both intrinsic and extrinsic approaches have been proposed for modeling iid manifold-valued data. However, to our knowledge our work is the first to take an extrinsic approach to the regression problem. The proposed extrinsic regression framework is general, computationally efficient, and theoretically appealing. Asymptotic distributions and convergence rates of the extrinsic regression estimates are derived and a large class of examples is considered indicating the wide applicability of our approach. Supplementary materials for this article are available online.


Brain and behavior | 2017

Rat intersubjective decisions are encoded by frequency-specific oscillatory contexts.

Jana Schaich Borg; Sanvesh Srivastava; Lizhen Lin; Joseph Heffner; David B. Dunson; Kafui Dzirasa; Luis de Lecea

It is unknown how the brain coordinates decisions to withstand personal costs in order to prevent other individuals’ distress. Here we test whether local field potential (LFP) oscillations between brain regions create “neural contexts” that select specific brain functions and encode the outcomes of these types of intersubjective decisions.


Biometrika | 2016

Data augmentation for models based on rejection sampling

Vinayak Rao; Lizhen Lin; David B. Dunson

Abstract We present a data augmentation scheme to perform Markov chain Monte Carlo inference for models where data generation involves a rejection sampling algorithm. Our idea is a simple scheme to instantiate the rejected proposals preceding each data point. The resulting joint probability over observed and rejected variables can be much simpler than the marginal distribution over the observed variables, which often involves intractable integrals. We consider three problems: modelling flow-cytometry measurements subject to truncation; the Bayesian analysis of the matrix Langevin distribution on the Stiefel manifold; and Bayesian inference for a nonparametric Gaussian process density model. The latter two are instances of doubly-intractable Markov chain Monte Carlo problems, where evaluating the likelihood is intractable. Our experiments demonstrate superior performance over state-of-the-art sampling algorithms for such problems.


Biometrics | 2016

Flexible link functions in nonparametric binary regression with Gaussian process priors

Dan Li; Xia Wang; Lizhen Lin; Dipak K. Dey

In many scientific fields, it is a common practice to collect a sequence of 0-1 binary responses from a subject across time, space, or a collection of covariates. Researchers are interested in finding out how the expected binary outcome is related to covariates, and aim at better prediction in the future 0-1 outcomes. Gaussian processes have been widely used to model nonlinear systems; in particular to model the latent structure in a binary regression model allowing nonlinear functional relationship between covariates and the expectation of binary outcomes. A critical issue in modeling binary response data is the appropriate choice of link functions. Commonly adopted link functions such as probit or logit links have fixed skewness and lack the flexibility to allow the data to determine the degree of the skewness. To address this limitation, we propose a flexible binary regression model which combines a generalized extreme value link function with a Gaussian process prior on the latent structure. Bayesian computation is employed in model estimation. Posterior consistency of the resulting posterior distribution is demonstrated. The flexibility and gains of the proposed model are illustrated through detailed simulation studies and two real data examples. Empirical results show that the proposed model outperforms a set of alternative models, which only have either a Gaussian process prior on the latent regression function or a Dirichlet prior on the link function.


Computational Statistics & Data Analysis | 2018

Robust and parallel Bayesian model selection

Michael Minyi Zhang; Henry Lam; Lizhen Lin

Abstract Effective and accurate model selection is an important problem in modern data analysis. One of the major challenges is the computational burden required to handle large datasets that cannot be stored or processed on one machine. Another challenge one may encounter is the presence of outliers and contaminations that damage the inference quality. The parallel “divide and conquer” model selection strategy divides the observations of the full dataset into roughly equal subsets and perform inference and model selection independently on each subset. After local subset inference, this method aggregates the posterior model probabilities or other model/variable selection criteria to obtain a final model by using the notion of geometric median. This approach leads to improved concentration in finding the “correct” model and model parameters and also is provably robust to outliers and data contamination.


Journal of Multivariate Analysis | 2017

Scale and curvature effects in principal geodesic analysis

Drew Lazar; Lizhen Lin

There is growing interest in using the close connection between differential geometry and statistics to model smooth manifold-valued data. In particular, much work has been done recently to generalize principal component analysis (PCA), the method of dimension reduction in linear spaces, to Riemannian manifolds. One such generalization is known as principal geodesic analysis (PGA). This paper, in a novel fashion, obtains Taylor expansions in scaling parameters introduced in the domain of objective functions in PGA. It is shown this technique not only leads to better closed-form approximations of PGA but also reveals the effects that scale, curvature and the distribution of data have on solutions to PGA and on their differences to first-order tangent space approximations. This approach should be able to be applied not only to PGA but also to other generalizations of PCA and more generally to other intrinsic statistics on Riemannian manifolds.


Archive | 2016

Nonparametric Curve Estimation

Rabi Bhattacharya; Lizhen Lin; Victor Patrangenaru

This chapter provides an introduction to nonparametric estimations of densities and regression functions by the kernel method.


Archive | 2016

Edgeworth Expansions and the Bootstrap

Rabi Bhattacharya; Lizhen Lin; Victor Patrangenaru

This chapter outlines the proof of the validity of a properly formulated version of the formal Edgeworth expansion, and derives from it the precise asymptotic rate of the coverage error of Efron’s bootstrap. A number of other applications of Edgeworth expansions are outlined.


Archive | 2016

The Nonparametric Bootstrap

Rabi Bhattacharya; Lizhen Lin; Victor Patrangenaru

This chapter introduces Efron’s nonparametric bootstrap, with applications to linear statistics, and semi-linear regression due to Bickel and Freedman.

Collaboration


Dive into the Lizhen Lin's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Vinayak Rao

University College London

View shared research outputs
Top Co-Authors

Avatar

Hui Xiong

University of Arizona

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Michael Minyi Zhang

University of Texas at Austin

View shared research outputs
Researchain Logo
Decentralizing Knowledge