Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Mohamed Hebiri is active.

Publication


Featured researches published by Mohamed Hebiri.


Bernoulli | 2017

On the Prediction Performance of the Lasso

Arnak S. Dalalyan; Mohamed Hebiri; Johannes Lederer

Although the Lasso has been extensively studied, the relationship between its prediction performance and the correlations of the covariates is not fully understood. In this paper, we give new insights into this relationship in the context of multiple linear regression. We show, in particular, that the incorporation of a simple correlation measure into the tuning parameter can lead to a nearly optimal prediction performance of the Lasso even for highly correlated covariates. However, we also reveal that for moderately correlated covariates, the prediction performance of the Lasso can be mediocre irrespective of the choice of the tuning parameter. We finally show that our results also lead to near-optimal rates for the least-squares estimator with total variation penalty.


Mathematical Methods of Statistics | 2008

Some theoretical results on the Grouped Variables Lasso

Christophe Chesneau; Mohamed Hebiri

We consider the linear regression model with Gaussian error. We estimate the unknown parameters by a procedure inspired by the Group Lasso estimator introduced in [22]. We show that this estimator satisfies a sparsity inequality, i.e., a bound in terms of the number of non-zero components of the oracle regression vector. We prove that this bound is better, in some cases, than the one achieved by the Lasso and the Dantzig selector.


IEEE Transactions on Information Theory | 2013

How Correlations Influence Lasso Prediction

Mohamed Hebiri; Johannes Lederer

We study how correlations in the design matrix influence Lasso prediction. First, we argue that the higher the correlations, the smaller the optimal tuning parameter. This implies in particular that the standard tuning parameters, that do not depend on the design matrix, are not favorable. Furthermore, we argue that Lasso prediction works well for any degree of correlations if suitable tuning parameters are chosen. We study these two subjects theoretically as well as with simulations.


Physical Review A | 2013

Rank penalized estimation of a quantum system

Katia Meziani; Mohamed Hebiri; Cristina Butucea; Pierre Alquier

We introduce a new method to reconstruct the quantum matrix


Statistics and Computing | 2010

Sparse conformal predictors

Mohamed Hebiri

\bar{\rho}


International Symposium on Statistical Learning and Data Sciences | 2015

Confidence Sets for Classification

Christophe Denis; Mohamed Hebiri

of a system of


international conference on machine learning | 2013

Learning Heteroscedastic Models by Convex Programming under Group Sparsity

Arnak S. Dalalyan; Mohamed Hebiri; Katia Meziani; Joseph Salmon

n


Statistics & Probability Letters | 2011

Generalization of ℓ1 constraints for high dimensional regression problems

Pierre Alquier; Mohamed Hebiri

-qubits and estimate its rank


Journal of Statistical Planning and Inference | 2012

Transductive versions of the LASSO and the Dantzig Selector

Pierre Alquier; Mohamed Hebiri

d


arXiv: Statistics Theory | 2015

Consistency of plug-in confidence sets for classification in semi-supervised learning

Christophe Denis; Mohamed Hebiri

from data obtained by quantum state tomography measurements repeated

Collaboration


Dive into the Mohamed Hebiri's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Pierre Alquier

University College Dublin

View shared research outputs
Top Co-Authors

Avatar

Joseph Salmon

Institut Mines-Télécom

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Cristina Butucea

Centre national de la recherche scientifique

View shared research outputs
Researchain Logo
Decentralizing Knowledge