Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Jan Poland is active.

Publication


Featured researches published by Jan Poland.


algorithmic learning theory | 2004

Prediction with Expert Advice by Following the Perturbed Leader for General Weights

Marcus Hutter; Jan Poland

When applying aggregating strategies to Prediction with Expert Advice, the learning rate must be adaptively tuned. The natural choice of \(\sqrt{\mbox{complexity/current loss}}\) renders the analysis of Weighted Majority derivatives quite complicated. In particular, for arbitrary weights there have been no results proven so far. The analysis of the alternative “Follow the Perturbed Leader” (FPL) algorithm from [KV03] (based on Hannan’s algorithm) is easier. We derive loss bounds for adaptive learning rate and both finite expert classes with uniform weights and countable expert classes with arbitrary weights. For the former setup, our loss bounds match the best known results so far, while for the latter our results are new.


IEEE Transactions on Information Theory | 2005

Asymptotics of discrete MDL for online prediction

Jan Poland; Marcus Hutter

Minimum description length (MDL) is an important principle for induction and prediction, with strong relations to optimal Bayesian learning. This paper deals with learning processes which are independent and identically distributed (i.i.d.) by means of two-part MDL, where the underlying model class is countable. We consider the online learning framework, i.e., observations come in one by one, and the predictor is allowed to update its state of mind after each time step. We identify two ways of predicting by MDL for this setup, namely, a static and a dynamic one. (A third variant, hybrid MDL, will turn out inferior.) We will prove that under the only assumption that the data is generated by a distribution contained in the model class, the MDL predictions converge to the true values almost surely. This is accomplished by proving finite bounds on the quadratic, the Hellinger, and the Kullback-Leibler loss of the MDL learner, which are, however, exponentially worse than for Bayesian prediction. We demonstrate that these bounds are sharp, even for model classes containing only Bernoulli distributions. We show how these bounds imply regret bounds for arbitrary loss functions. Our results apply to a wide range of setups, namely, sequence prediction, pattern classification, regression, and universal induction in the sense of algorithmic information theory among others.


conference on learning theory | 2004

Convergence of Discrete MDL for Sequential Prediction.

Jan Poland; Marcus Hutter

We study the properties of the Minimum Description Length principle for sequence prediction, considering a two-part MDL estimator which is chosen from a countable class of models. This applies in particular to the important case of universal sequence prediction, where the model class corresponds to all algorithms for some fixed universal Turing machine (this correspondence is by enumerable semimeasures, hence the resulting models are stochastic). We prove convergence theorems similar to Solomonoff’s theorem of universal induction, which also holds for general Bayes mixtures. The bound characterizing the convergence speed for MDL predictions is exponentially larger as compared to Bayes mixtures. We observe that there are at least three different ways of using MDL for prediction. One of these has worse prediction properties, for which predictions only converge if the MDL estimator stabilizes. We establish sufficient conditions for this to occur. Finally, some immediate consequences for complexity relations and randomness criteria are proven.


Statistics and Computing | 2006

MDL convergence speed for Bernoulli sequences

Jan Poland; Marcus Hutter

The Minimum Description Length principle for online sequence estimation/prediction in a proper learning setup is studied. If the underlying model class is discrete, then the total expected square loss is a particularly interesting performance measure: (a) this quantity is finitely bounded, implying convergence with probability one, and (b) it additionally specifies the convergence speed. For MDL, in general one can only have loss bounds which are finite but exponentially larger than those for Bayes mixtures. We show that this is even the case if the model class contains only Bernoulli distributions. We derive a new upper bound on the prediction error for countable Bernoulli classes. This implies a small bound (comparable to the one for Bayes mixtures) for certain important model classes. We discuss the application to Machine Learning tasks such as classification and hypothesis testing, and generalization to countable classes of i.i.d. models.


international conference on stochastic algorithms: foundations and applications | 2005

FPL analysis for adaptive bandits

Jan Poland

A main problem of “Follow the Perturbed Leader” strategies for online decision problems is that regret bounds are typically proven against oblivious adversary. In partial observation cases, it was not clear how to obtain performance guarantees against adaptive adversary, without worsening the bounds. We propose a conceptually simple argument to resolve this problem. Using this, a regret bound of


algorithmic learning theory | 2004

On the convergence speed of MDL predictions for Bernoulli sequences

Jan Poland; Marcus Hutter

O(t^{\frac{2}{3}})


algorithmic learning theory | 2005

Defensive universal learning with experts

Jan Poland; Marcus Hutter

for FPL in the adversarial multi-armed bandit problem is shown. This bound holds for the common FPL variant using only the observations from designated exploration rounds. Using all observations allows for the stronger bound of


international conference on stochastic algorithms: foundations and applications | 2001

Evolutionary Search for Smooth Maps in Motor Control Unit Calibration

Jan Poland; Kosmas Knödler; Alexander Mitterer; Thomas Fleischhauer; Frank Zuber-Goos; Andreas Zell

O(\sqrt{t})


discovery science | 2006

Clustering pairwise distances with missing data: maximum cuts versus normalized cuts

Jan Poland; Thomas Zeugmann

, matching the best bound known so far (and essentially the known lower bound) for adversarial bandits. Surprisingly, this variant does not even need explicit exploration, it is self-stabilizing. However the sampling probabilities have to be either externally provided or approximated to sufficient accuracy, using O(t2log t) samples in each step.


MTZ - Motortechnische Zeitschrift | 2003

Modellbasierte Online-Optimierung moderner Verbrennungsmotoren

Kosmas Knödler; Jan Poland; Thomas Fleischhauer; Alexander Mitterer; Stephan Ullmann; Andreas Zell

We consider the Minimum Description Length principle for online sequence prediction. If the underlying model class is discrete, then the total expected square loss is a particularly interesting performance measure: (a) this quantity is bounded, implying convergence with probability one, and (b) it additionally specifies a rate of convergence. Generally, for MDL only exponential loss bounds hold, as opposed to the linear bounds for a Bayes mixture. We show that this is even the case if the model class contains only Bernoulli distributions. We derive a new upper bound on the prediction error for countable Bernoulli classes. This implies a small bound (comparable to the one for Bayes mixtures) for certain important model classes. The results apply to many Machine Learning tasks including classification and hypothesis testing. We provide arguments that our theorems generalize to countable classes of i.i.d. models.

Collaboration


Dive into the Jan Poland's collaboration.

Top Co-Authors

Avatar

Andreas Zell

University of Tübingen

View shared research outputs
Top Co-Authors

Avatar

Marcus Hutter

Australian National University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Rainer Nagel

University of Tübingen

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge