Featured Researches

Statistics Theory

On Projection Robust Optimal Transport: Sample Complexity and Model Misspecification

Optimal transport (OT) distances are increasingly used as loss functions for statistical inference, notably in the learning of generative models or supervised learning. Yet, the behavior of minimum Wasserstein estimators is poorly understood, notably in high-dimensional regimes or under model misspecification. In this work we adopt the viewpoint of projection robust (PR) OT, which seeks to maximize the OT cost between two measures by choosing a k -dimensional subspace onto which they can be projected. Our first contribution is to establish several fundamental statistical properties of PR Wasserstein distances, complementing and improving previous literature that has been restricted to one-dimensional and well-specified cases. Next, we propose the integral PR Wasserstein (IPRW) distance as an alternative to the PRW distance, by averaging rather than optimizing on subspaces. Our complexity bounds can help explain why both PRW and IPRW distances outperform Wasserstein distances empirically in high-dimensional inference tasks. Finally, we consider parametric inference using the PRW distance. We provide an asymptotic guarantee of two types of minimum PRW estimators and formulate a central limit theorem for max-sliced Wasserstein estimator under model misspecification. To enable our analysis on PRW with projection dimension larger than one, we devise a novel combination of variational analysis and statistical theory.

Read more
Statistics Theory

On Second order correctness of Bootstrap in Logistic Regression

In the fields of clinical trials, biomedical surveys, marketing, banking, with dichotomous response variable, the logistic regression is considered as an alternative convenient approach to linear regression. In this paper, we develop a novel bootstrap technique based on perturbation resampling method for approximating the distribution of the maximum likelihood estimator (MLE) of the regression parameter vector. We establish second order correctness of the proposed bootstrap method after proper studentization and smoothing. It is shown that inferences drawn based on the proposed bootstrap method are more accurate compared to that based on asymptotic normality. The main challenge in establishing second order correctness remains in the fact that the response variable being binary, the resulting MLE has a lattice structure. We show the direct bootstrapping approach fails even after studentization. We adopt smoothing technique developed in Lahiri (1993) to ensure that the smoothed studentized version of the MLE has a density. Similar smoothing strategy is employed to the bootstrap version also to achieve second order correct approximation.

Read more
Statistics Theory

On Smooth Change-Point Location Estimation for Poisson Processes

We are interested in estimating the location of what we call "smooth change-point" from n independent observations of an inhomogeneous Poisson process. The smooth change-point is a transition of the intensity function of the process from one level to another which happens smoothly, but over such a small interval, that its length δ_n is considered to be decreasing to 0 as n→+∞ . We show that if δ_n goes to zero slower than 1/n , our model is locally asymptotically normal (with a rather unusual rate δ_n/n − − − − − √ ), and the maximum likelihood and Bayesian estimators are consistent, asymptotically normal and asymptotically efficient. If, on the contrary, δ_n goes to zero faster than 1/n , our model is non-regular and behaves like a change-point model. More precisely, in this case we show that the Bayesian estimators are consistent, converge at rate 1/n , have non-Gaussian limit distributions and are asymptotically efficient. All these results are obtained using the likelihood ratio analysis method of Ibragimov and Khasminskii, which equally yields the convergence of polynomial moments of the considered estimators. However, in order to study the maximum likelihood estimator in the case where δ_n goes to zero faster than 1/n , this method cannot be applied using the usual topologies of convergence in functional spaces. So, this study should go through the use of an alternative topology and will be considered in a future work.

Read more
Statistics Theory

On Suboptimality of Least Squares with Application to Estimation of Convex Bodies

We develop a technique for establishing lower bounds on the sample complexity of Least Squares (or, Empirical Risk Minimization) for large classes of functions. As an application, we settle an open problem regarding optimality of Least Squares in estimating a convex set from noisy support function measurements in dimension d≥6 . Specifically, we establish that Least Squares is mimimax sub-optimal, and achieves a rate of Θ ~ d ( n −2/(d−1) ) whereas the minimax rate is Θ d ( n −4/(d+3) ) .

Read more
Statistics Theory

On Weighted Extropies

The extropy is a measure of information introduced by Lad et al. (2015) as dual to entropy. As the entropy, it is a shift-independent information measure. We introduce here the notion of weighted extropy, a shift-dependent information measure which gives higher weights to large values of observed random variables. We also study the weighted residual and past extropies as weighted versions of extropy for residual and past lifetimes. Bivariate versions extropy and weighted extropy are also provided. Several examples are presented through out to illustrate the various concepts introduced here.

Read more
Statistics Theory

On a family that unifies Generalized Marshall-Olkin and Poisson-G family of distribution

Unifying the generalized Marshall-Olkin (GMO) and Poisson-G (P-G) a new family of distribution is proposed. Density and the survival function are expressed as infinite mixtures of P-G family. The quantile function, asymptotes, shapes, stochastic ordering, moment generating function, order statistics, probability weighted moments and Rényi entropy are derived. Maximum likelihood estimation with large sample properties is presented. A Monte Carlo simulation is used to examine the pattern of the bias and the mean square error of the maximum likelihood estimators. An illustration of comparison with some of the important sub models of the family in modeling a real data reveals the utility of the proposed family.

Read more
Statistics Theory

On absolute central moments of Poisson distribution

A recurrence formula for absolute central moments of Poisson distribution is suggested.

Read more
Statistics Theory

On admissible estimation of a mean vector when the scale is unknown

We consider admissibility of generalized Bayes estimators of the mean of a multivariate normal distribution when the scale is unknown under quadratic loss. The priors considered put the improper invariant prior on the scale while the prior on the mean has a hierarchical normal structure conditional on the scale. This conditional hierarchical prior is essentially that of Maruyama and Strawderman (2021, Biometrika) (MS21) which is indexed by a hyperparameter a . In that paper a is chosen so this conditional prior is proper which corresponds to a>?? . This paper extends MS21 by considering improper conditional priors with a in the closed interval [??,??] , and establishing admissibility for such a . The authors, in Maruyama and Strawderman (2017, JMVA), have earlier shown that such conditional priors with a<?? lead to inadmissible estimators. This paper therefore completes the determination of admissibility/inadmissibility for this class of priors. It establishes the the boundary as a=?? , with admissibility holding for a?��?2 and inadmissibility for a<?? . This boundary corresponds exactly to that in the known scale case for these conditional priors, and which follows from Brown (1971, AOMS). As a notable benefit of this enlargement of the class of admissible generalized Bayes estimators, we give admissible and minimax estimators in all dimensions greater than 2 as opposed to MS21 which required the dimension to be greater than 4 . In one particularly interesting special case, we establish that the joint Stein prior for the unknown scale case leads to a minimax admissible estimator in all dimensions greater than 2 .

Read more
Statistics Theory

On attainability of Kendall's tau matrices and concordance signatures

The concordance signature of a multivariate continuous distribution is the vector of concordance probabilities for margins of all orders; it underlies the bivariate and multivariate Kendall's tau measure of concordance. It is shown that every attainable concordance signature is equal to the concordance signature of a unique mixture of the extremal copulas, that is the copulas with extremal correlation matrices consisting exclusively of 1's and -1's. This result establishes that the set of attainable Kendall rank correlation matrices of multivariate continuous distributions in arbitrary dimension is the set of convex combinations of extremal correlation matrices, a set known as the cut polytope. A methodology for testing the attainability of concordance signatures using linear optimization and convex analysis is provided. The elliptical copulas are shown to yield a strict subset of the attainable concordance signatures as well as a strict subset of the attainable Kendall rank correlation matrices; the Student t copula is seen to converge to a mixture of extremal copulas sharing its concordance signature with all elliptical distributions that have the same correlation matrix. A method of estimating an attainable concordance signature from data is derived and shown to correspond to using standard estimates of Kendall's tau in the absence of ties. The methodology has application to Monte Carlo simulations of dependent random variables as well as expert elicitation of consistent systems of Kendall's tau dependence measures.

Read more
Statistics Theory

On characteristic rank for matrix and tensor completion

In this lecture note, we discuss a fundamental concept, referred to as the {\it characteristic rank}, which suggests a general framework for characterizing the basic properties of various low-dimensional models used in signal processing. Below, we illustrate this framework using two examples: matrix and three-way tensor completion problems, and consider basic properties include identifiability of a matrix or tensor, given partial observations. In this note, we consider cases without observation noise to illustrate the principle.

Read more

Ready to get started?

Join us today