Alexander Goldenshluger
University of Haifa
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Alexander Goldenshluger.
Annals of Statistics | 2011
Alexander Goldenshluger; Oleg Lepski
We address the problem of density estimation with L s-loss by selection of kernel estimators. We develop a selection procedure and derive corresponding L s-risk oracle inequalities. It is shown that the proposed selection rule leads to the estimator being minimax adaptive over a scale of the anisotropic Nikolskii classes. The main technical tools used in our derivations are uniform bounds on the L s-norms of empirical processes developed recently by Goldenshluger and Lepski [Ann. Probab. (2011), to appear].
Annals of Statistics | 2004
Alexander Goldenshluger; Assaf Zeevi
This paper pursues a statistical study of the Hough transform; the celebrated computer vision algorithm used to detect the presence of lines in a noisy image. We first study asymptotic properties of the Hough transform estimator, whose objective is to find the line that “best” fits a set of planar points. In particular, we establish strong consistency, rates of convergence and characterize the limiting distribution of the Hough transform estimator. While the convergence rates are seen to be slower than those found in some standard regression methods, the Hough transform estimator is shown to be more robust as measured by its breakdown point. We next study the Hough transform in the context of the problem of detecting multiple lines. This is addressed via the framework of excess mass functionals and modality testing. Throughout, several numerical examples help illustrate various properties of the estimator. Relations between the Hough transform and more mainstream statistical paradigms and methods are discussed as well. Short Title: The Hough transform estimator
Bernoulli | 1999
Alexander Goldenshluger
We consider estimating an unknown function f from indirect white noise observations with particular emphasis on the problem of nonparametric deconvolution. Nonparametric estimators that can adapt to unknown smoothness of f are developed. The adaptive estimators are specified under two sets of assumptions on the kernel of the convolution transform. In particular, kernels having Fourier transform with polynomially and exponentially decaying tails are considered. It is shown that the proposed estimates possess, in a sense, the best possible abilities for pointwise adaptation.
Annals of Statistics | 2006
Alexander Goldenshluger; A. Tsybakov; Assaf Zeevi
We study nonparametric change-point estimation from indirect noisy observations. Focusing on the white noise convolution model, we consider two classes of functions that are smooth apart from the change-point. We establish lower bounds on the minimax risk in estimating the change-point and develop rate optimal estimation procedures. The results demonstrate that the best achievable rates of convergence are determined both by smoothness of the function away from the change-point and by the degree of ill-posedness of the convolution operator. Optimality is obtained by introducing a new technique that involves, as a key element, detection of zero crossings of an estimate of the properly smoothed second derivative of the underlying function.
Annals of Statistics | 2009
Alexander Goldenshluger
In this paper we study the aggregation problem that can be formulated as follows. Assume that we have a family of estimators F built on the basis of available observations. The goal is to construct a new estimator whose risk is as close as possible to that of the best estimator in the family. We propose a general aggregation scheme that is universal in the following sense: it applies for families of arbitrary estimators and a wide variety of models and global risk measures. The procedure is based on comparison of empirical estimates of certain linear functionals with estimates induced by the family F. We derive oracle inequalities and show that they are unimprovable in some sense. Numerical results demonstrate good practical behavior of the procedure.
Annals of Probability | 2011
Alexander Goldenshluger; Oleg Lepski
In this paper we develop a general machinery for finding explicit uniform probability and moment bounds on sub-additive positive functionals of random processes. Using the developed general technique we derive uniform bounds on the Ls–norms of empirical and regression–type processes. Usefulness of the obtained results is illustrated by application to the processes appearing in kernel density estimation and in nonparametric estimation of regression functions.
Stochastic Systems | 2013
Alexander Goldenshluger; Assaf Zeevi
We consider a two–armed bandit problem which involves sequential sampling from two non-homogeneous populations. The response in each is determined by a random covariate vector and a vector of parameters whose values are not known a priori. The goal is to maximize cumulative expected reward. We study this problem in a minimax setting, and develop rate-optimal polices that combine myopic action based on least squares estimates with a suitable “forced sampling” strategy. It is shown that the regret grows logarithmically in the time horizon n and no policy can achieve a slower growth rate over all feasible problem instances. In this setting of linear response bandits, the identity of the sub-optimal action changes with the values of the covariate vector, and the optimal policy is subject to sampling from the inferior population at a rate that grows like n.
Annals of Statistics | 2006
Alexander Goldenshluger; Assaf Zeevi
We consider the problem of estimating convex boundaries from blurred and noisy observations. In our model, the convolution of an intensity function f is observed with additive Gaussian white noise. The function f is assumed to have convex support G whose boundary is to be recovered. Rather than directly estimating the intensity function, we develop a procedure which is based on estimating the support function of the set G. This approach is closely related to the method of geometric hyperplane probing, a well–known technique in computer vision applications. We establish bounds that reveal how the estimation accuracy depends on the ill-posedness of the convolution operator and the behavior of the intensity function near the boundary. Short Title: Recovering convex boundaries
IEEE Transactions on Information Theory | 2006
Alexander Goldenshluger; Vladimir Spokoiny
We consider the problem of recovering edges of an image from noisy tomographic data. The original image is assumed to have a discontinuity jump (edge) along the boundary of a compact convex set. The Radon transform of the image is observed with noise, and the problem is to estimate the edge. We develop an estimation procedure which is based on recovering the support function of the edge. It is shown that the proposed estimator is nearly optimal in order in a minimax sense. Numerical examples illustrate reasonable practical behavior of the estimation procedure.
Annales De L Institut Henri Poincare-probabilites Et Statistiques | 2008
Alexander Goldenshluger; Anatoli Juditsky; Alexandre B. Tsybakov; Assaf Zeevi
We consider the problem of nonparametric estimation of signal singularities from indirect and noisy observations. Here by singularity, we mean a discontinuity (change-point) of the signal or of its derivative. The model of indirect observations we consider is that of a linear transform of the signal, observed in white noise. The estimation problem is analyzed in a minimax framework. We provide lower bounds for minimax risks and propose rate-optimal estimation procedures.