Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Tiangang Cui is active.

Publication


Featured researches published by Tiangang Cui.


Inverse Problems | 2014

Likelihood-informed dimension reduction for nonlinear inverse problems

Tiangang Cui; James F. Martin; Youssef M. Marzouk; Antti Solonen; Alessio Spantini

United States. Dept. of Energy. Office of Advanced Scientific Computing Research (Grant DE-SC0003908)


International Journal for Numerical Methods in Engineering | 2015

Data‐driven model reduction for the Bayesian solution of inverse problems

Tiangang Cui; Youssef M. Marzouk; Karen Willcox

United States. Dept. of Energy. Office of Advanced Scientific Computing Research (Applied Mathematics Program Award DE-FG02-08ER2585)


Journal of Computational Physics | 2016

Dimension-independent likelihood-informed MCMC

Tiangang Cui; Kody J. H. Law; Youssef M. Marzouk

Many Bayesian inference problems require exploring the posterior distribution of high-dimensional parameters that represent the discretization of an underlying function. This work introduces a family of Markov chain Monte Carlo (MCMC) samplers that can adapt to the particular structure of a posterior distribution over functions. Two distinct lines of research intersect in the methods developed here. First, we introduce a general class of operator-weighted proposal distributions that are well defined on function space, such that the performance of the resulting MCMC samplers is independent of the discretization of the function. Second, by exploiting local Hessian information and any associated low-dimensional structure in the change from prior to posterior distributions, we develop an inhomogeneous discretization scheme for the Langevin stochastic differential equation that yields operator-weighted proposals adapted to the non-Gaussian structure of the posterior. The resulting dimension-independent and likelihood-informed (DILI) MCMC samplers may be useful for a large class of high-dimensional problems where the target probability measure has a density with respect to a Gaussian reference measure. Two nonlinear inverse problems are used to demonstrate the efficiency of these DILI samplers: an elliptic PDE coefficient inverse problem and path reconstruction in a conditioned diffusion.


Journal of Computational Physics | 2016

Scalable posterior approximations for large-scale Bayesian inverse problems via likelihood-informed parameter and state reduction

Tiangang Cui; Youssef M. Marzouk; Karen Willcox

Two major bottlenecks to the solution of large-scale Bayesian inverse problems are the scaling of posterior sampling algorithms to high-dimensional parameter spaces and the computational cost of forward model evaluations. Yet incomplete or noisy data, the state variation and parameter dependence of the forward model, and correlations in the prior collectively provide useful structure that can be exploited for dimension reduction in this setting-both in the parameter space of the inverse problem and in the state space of the forward model. To this end, we show how to jointly construct low-dimensional subspaces of the parameter space and the state space in order to accelerate the Bayesian solution of the inverse problem. As a byproduct of state dimension reduction, we also show how to identify low-dimensional subspaces of the data in problems with high-dimensional observations. These subspaces enable approximation of the posterior as a product of two factors: (i) a projection of the posterior onto a low-dimensional parameter subspace, wherein the original likelihood is replaced by an approximation involving a reduced model; and (ii) the marginal prior distribution on the high-dimensional complement of the parameter subspace. We present and compare several strategies for constructing these subspaces using only a limited number of forward and adjoint model simulations. The resulting posterior approximations can rapidly be characterized using standard sampling techniques, e.g., Markov chain Monte Carlo. Two numerical examples demonstrate the accuracy and efficiency of our approach: inversion of an integral equation in atmospheric remote sensing, where the data dimension is very high; and the inference of a heterogeneous transmissivity field in a groundwater system, which involves a partial differential equation forward model with high dimensional state and parameters.


Journal of Hydrologic Engineering | 2013

Uncertainty Quantification for Stream Depletion Tests

Tiangang Cui; Nicholas Dudley Ward

AbstractThis study considers the problem of quantifying stream depletion from pumping test data. Bayesian inference is used to quantify the posterior uncertainty of parameters for a simple vertically heterogeneous aquifer model, in which the pumped semiconfined aquifer is separated by an aquiclude from a phreatic aquifer hydraulically connected to a stream. This study investigates the effects of using different data sets and shows that a single pumping test is generally not sufficient to determine stream depletion within reasonable limits. However, uncertainty quantification conducted within a Bayesian context reveals that by judicious design of aquifer tests, stream depletion can be accurately determined from data.


SIAM Journal on Scientific Computing | 2017

Goal-Oriented Optimal Approximations of Bayesian Linear Inverse Problems

Alessio Spantini; Tiangang Cui; Karen Willcox; Luis Tenorio; Youssef M. Marzouk

We propose optimal dimensionality reduction techniques for the solution of goal-oriented linear-Gaussian inverse problems, where the quantity of interest (QoI) is a function of the inversion parameters. These approximations are suitable for large-scale applications. In particular, we study the approximation of the posterior covariance of the QoI as a low-rank negative update of its prior covariance and prove optimality of this update with respect to the natural geodesic distance on the manifold of symmetric positive definite matrices. Assuming exact knowledge of the posterior mean of the QoI, the optimality results extend to optimality in distribution with respect to the Kullback--Leibler divergence and the Hellinger distance between the associated distributions. We also propose the approximation of the posterior mean of the QoI as a low-rank linear function of the data and prove optimality of this approximation with respect to a weighted Bayes risk. Both of these optimal approximations avoid the explicit...


SIAM Journal on Scientific Computing | 2017

Bayesian Inverse Problems with

Zheng Wang; Johnathan M. Bardsley; Antti Solonen; Tiangang Cui; Youssef M. Marzouk

Prior distributions for Bayesian inference that rely on the


Journal of Structural Biology | 2018

l_1

Cyril Reboul; Simon Kiesewetter; Michael Eager; Matthew J. Belousoff; Tiangang Cui; Hans De Sterck; Dominika Elmlund; Hans Elmlund

l_1


Water Resources Research | 2011

Priors: A Randomize-Then-Optimize Approach

Tiangang Cui; Colin Fox; Michael J. O'Sullivan

-norm of the parameters are of considerable interest, in part because they promote parameter fields with less regularity than Gaussian priors (e.g., discontinuities and blockiness). These


SIAM Journal on Scientific Computing | 2015

Rapid near-atomic resolution single-particle 3D reconstruction with SIMPLE

Alessio Spantini; Antti Solonen; Tiangang Cui; James F. Martin; Luis Tenorio; Youssef M. Marzouk

l_1

Collaboration


Dive into the Tiangang Cui's collaboration.

Top Co-Authors

Avatar

Youssef M. Marzouk

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Alessio Spantini

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Karen Willcox

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Antti Solonen

Lappeenranta University of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

James F. Martin

Baylor College of Medicine

View shared research outputs
Top Co-Authors

Avatar

Kody J. H. Law

Oak Ridge National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Luis Tenorio

Colorado School of Mines

View shared research outputs
Researchain Logo
Decentralizing Knowledge