Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Daniel Polani is active.

Publication


Featured researches published by Daniel Polani.


Archive | 2011

Information Theory of Decisions and Actions

Naftali Tishby; Daniel Polani

The perception–action cycle is often defined as “the circular flow of information between an organism and its environment in the course of a sensory guided sequence of actions towards a goal” (Fuster, Neuron 30:319–333, 2001; International Journal of Psychophysiology 60(2):125–132, 2006). The question we address in this chapter is in what sense this “flow of information” can be described by Shannon’s measures of information introduced in his mathematical theory of communication. We provide an affirmative answer to this question using an intriguing analogy between Shannon’s classical model of communication and the perception–action cycle. In particular, decision and action sequences turn out to be directly analogous to codes in communication, and their complexity – the minimal number of (binary) decisions required for reaching a goal – directly bounded by information measures, as in communication. This analogy allows us to extend the standard reinforcement learning framework. The latter considers the future expected reward in the course of a behaviour sequence towards a goal (value-to-go). Here, we additionally incorporate a measure of information associated with this sequence: the cumulated information processing cost or bandwidth required to specify the future decision and action sequence (information-to-go). Using a graphical model, we derive a recursive Bellman optimality equation for information measures, in analogy to reinforcement learning; from this, we obtain new algorithms for calculating the optimal trade-off between the value-to-go and the required information-to-go, unifying the ideas behind the Bellman and the Blahut–Arimoto iterations. This trade-off between value-to-go and information-to-go provides a complete analogy with the compression–distortion trade-off in source coding. The present new formulation connects seemingly unrelated optimization problems. The algorithm is demonstrated on grid world examples.


congress on evolutionary computation | 2005

Empowerment: a universal agent-centric measure of control

Daniel Polani; Chrystopher L. Nehaniv

The classical approach to using utility functions suffers from the drawback of having to design and tweak the functions on a case by case basis. Inspired by examples from the animal kingdom, social sciences and games we propose empowerment, a rather universal function, defined as the information-theoretic capacity of an agents actuation channel. The concept applies to any sensorimotor apparatus. Empowerment as a measure reflects the properties of the apparatus as long as they are observable due to the coupling of sensors and actuators via the environment. Using two simple experiments we also demonstrate how empowerment influences sensor-actuator evolution


Connection Science | 2006

From unknown sensors and actuators to actions grounded in sensorimotor perceptions

Lars Olsson; Chrystopher L. Nehaniv; Daniel Polani

This article describes a developmental system based on information theory implemented on a real robot that learns a model of its own sensory and actuator apparatus. There is no innate knowledge regarding the modalities or representation of the sensory input and the actuators, and the system relies on generic properties of the robot’s world, such as piecewise smooth effects of movement on sensory changes. The robot develops the model of its sensorimotor system by first performing random movements to create an informational map of the sensors. Using this map, the robot then learns what effects the different possible actions have on the sensors. After this developmental process, the robot can perform basic visually guided movement.


Physical Review E | 2013

Bivariate Measure of Redundant Information

Malte Harder; Christoph Salge; Daniel Polani

We define a measure of redundant information based on projections in the space of probability distributions. Redundant information between random variables is information that is shared between those variables. But, in contrast to mutual information, redundant information denotes information that is shared about the outcome of a third variable. Formalizing this concept, and being able to measure it, is required for the non-negative decomposition of mutual information into redundant and synergistic information. Previous attempts to formalize redundant or synergistic information struggle to capture some desired properties. We introduce a new formalism for redundant information and prove that it satisfies all the properties necessary outlined in earlier work, as well as an additional criterion that we propose to be necessary to capture redundancy. We also demonstrate the behavior of this new measure for several examples, compare it to previous measures, and apply it to the decomposition of transfer entropy.


european conference on artificial life | 2005

All else being equal be empowered

Daniel Polani; Chrystopher L. Nehaniv

The classical approach to using utility functions suffers from the drawback of having to design and tweak the functions on a case by case basis. Inspired by examples from the animal kingdom, social sciences and games we propose empowerment, a rather universal function, defined as the information-theoretic capacity of an agent’s actuation channel. The concept applies to any sensorimotoric apparatus. Empowerment as a measure reflects the properties of the apparatus as long as they are observable due to the coupling of sensors and actuators via the environment.


Hfsp Journal | 2009

Information: currency of life?

Daniel Polani

In biology, the exception is mostly the rule, and the rule is mostly the exception. However, recent results indicate that known universal concepts in biology such as the genetic code or the utilization of ATP as a source of energy may be complemented by a large class of principles based on Shannons concept of information. The present position paper discusses various promising pathways toward the formulation of such generic informational principles and their relevance for the realm of biology.


Neural Computation | 2007

Representations of Space and Time in the Maximization of Information Flow in the Perception-Action Loop

Daniel Polani; Chrystopher L. Nehaniv

Sensor evolution in nature aims at improving the acquisition of information from the environment and is intimately related with selection pressure toward adaptivity and robustness. Our work in the area indicates that information theory can be applied to the perception-action loop. This letter studies the perception-action loop of agents, which is modeled as a causal Bayesian network. Finite state automata are evolved as agent controllers in a simple virtual world to maximize information flow through the perception-action loop. The information flow maximization organizes the agents behavior as well as its information processing. To gain more insight into the results, the evolved implicit representations of space and time are analyzed in an information-theoretic manner, which paves the way toward a principled and general understanding of the mechanisms guiding the evolution of sensors in nature and provides insights into the design of mechanisms for artificial sensor evolution.


nasa dod conference on evolvable hardware | 2004

Organization of the information flow in the perception-action loop of evolved agents

Daniel Polani; Chrystopher L. Nehaniv

Sensor evolution in nature aims at improving the acquisition of information from the environment and is intimately related with selection pressure towards adaptivity and robustness. Recent work in the area aims at studying the perception-action loop in a formalized information-theoretic manner. This paves the way towards a principled and general understanding of the mechanisms guiding the evolution of sensors in nature and provides insights into the design of mechanisms of artificial sensor evolution. In our paper we study the perception-action loop of agents. We evolve finite-state automata as agent controllers to solve an information acquisition task in a simple virtual world and study how the information flow is organized by evolution. Our analysis of the evolved automata and the information flow provides insight into how evolution organizes sensoric information acquisition, memory, processing and action selection. In addition, the results are compared to ideal information extraction schemes following from the Information Bottleneck principle.


PLOS ONE | 2008

Keep Your Options Open : An Information-Based Driving Principle for Sensorimotor Systems

Daniel Polani; Chrystopher L. Nehaniv

The central resource processed by the sensorimotor system of an organism is information. We propose an information-based quantity that allows one to characterize the efficiency of the perception-action loop of an abstract organism model. It measures the potential of the organism to imprint information on the environment via its actuators in a way that can be recaptured by its sensors, essentially quantifying the options available and visible to the organism. Various scenarios suggest that such a quantity could identify the preferred direction of evolution or adaptation of the sensorimotor loop of organisms.


Journal of Economic Dynamics and Control | 2003

Learning competitive pricing strategies by multi-agent reinforcement learning

Erich Kutschinski; Thomas Uthmann; Daniel Polani

In electronic marketplaces automated and dynamic pricing is becoming increasingly popular. Agents that perform this task can improve themselves by learning from past observations, possibly using reinforcement learning techniques. Co-learning of several adaptive agents against each other may lead to unforeseen results and increasingly dynamic behavior of the market. In this article we shed some light on price developments arising from a simple price adaptation strategy. Furthermore, we examine several adaptive pricing strategies and their learning behavior in a co-learning scenario with different levels of competition. Q-learning manages to learn best-reply strategies well, but is expensive to train.

Collaboration


Dive into the Daniel Polani's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Christoph Salge

University of Hertfordshire

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Lars Olsson

University of Hertfordshire

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Kerstin Dautenhahn

University of Hertfordshire

View shared research outputs
Top Co-Authors

Avatar

Cornelius Glackin

University of Hertfordshire

View shared research outputs
Top Co-Authors

Avatar

Martin Biehl

University of Hertfordshire

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Philippe Capdepuy

University of Hertfordshire

View shared research outputs
Researchain Logo
Decentralizing Knowledge