Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Andrey Kolobov is active.

Publication


Featured researches published by Andrey Kolobov.


international joint conference on artificial intelligence | 2011

Towards scalable mdp algorithms

Andrey Kolobov

Since the emergence of Artificial Intelligence as a field, planning under uncertainty has been viewed as one of its crucial subareas. Accordingly, the current lack of scalability of probabilistic planning techniques is a major reason why the grand vision of AI has not been fulfilled yet. Besides being an obstacle to advances in AI, scalability issues also hamper the applicability of probabilistic planning to real-world problems. A powerful framework for describing probabilistic planning problems is Markov Decision Processes (MDPs). Informally, an MDP specifies the objectives the agent is trying to achieve, the actions the agent can perform, and the states in which it can end up while working towards the objective. Solving an MDP means finding a policy, i.e. an assignment of actions to states, that allows the agent to achieve the objective. Optimal solution methods, those that look for the “best” policy according to some criterion, typically try to analyze all possible states or a large fraction of them. Since state space sizes of realistic scenarios can be astronomical, these algorithms quickly run out of memory. Fortunately, the mathematical structure of some classes of MDPs has allowed for inventing more efficient algorithms. This is the case for Stochastic Shortest Path (SSP) problems, whose mathematical properties gave rise to a family of algorithms called Find-and-Revise. When used in combination with a heuristic, the members of this family can find a near-optimal, or even optimal, policy while avoiding the analysis of many states. Nonetheless, the sheer number of states in real-world problems forces algorithms based on state-level analysis to exhaust their capabilities much too early, calling for a fundamentally different approach. Moreover, several expressive and potentially useful MDP classes are currently not known to have a mathematical structure for creating efficient approximation techniques. This dissertation advances the state of the art in probabilistic planning in three complementary ways. For SSP MDPs, it derives a novel class of approximate algorithms based on generalizing state analysis information across many states. This information is stored in the form of basis functions that are generated automatically and the number of which is much smaller than the number of states. As a result, the proposed algorithms have a very compact memory footprint and arrive at high-quality solutions faster than their state-based counterparts. In a parallel effort, the dissertation also builds up mathematical apparatus for classes of MDPs that previously had no applicable approximation algorithms. The developed theory enables the extension of the powerful Find-and-Revise paradigm to these MDP classes as well, providing the first memory-efficient algorithms for solving them. Last but not least, the dissertation will apply the proposed theoretical techniques to a large-scale real-world problem, urban traffic routing being one of the candidates.


international joint conference on artificial intelligence | 2005

BLOG: probabilistic models with unknown objects

Brian Milch; Bhaskara Marthi; Stuart J. Russell; David Sontag; Daniel L. Ong; Andrey Kolobov


Archive | 2012

Planning with Markov Decision Processes: An AI Perspective

Andrey Kolobov


international conference on automated planning and scheduling | 2011

Heuristic search for generalized stochastic shortest path MDPs

Andrey Kolobov; Daniel S. Weld; Hector Geffner


national conference on artificial intelligence | 2012

LRTDP versus UCT for online probabilistic planning

Andrey Kolobov; Mausam Mausam; Daniel S. Weld


international conference on artificial intelligence and statistics | 2005

Approximate inference for infinite contingent Bayesian networks

Brian Milch; Bhaskara Marthi; David Sontag; Stuart J. Russell; Daniel L. Ong; Andrey Kolobov


uncertainty in artificial intelligence | 2012

A theory of goal-oriented MDPs with dead ends

Andrey Kolobov; Daniel S. Weld


international joint conference on artificial intelligence | 2009

ReTrASE: integrating paradigms for approximate probabilistic planning

Andrey Kolobov; Daniel S. Weld


international conference on automated planning and scheduling | 2012

Reverse iterative deepening for finite-horizon MDPs with large branching factors

Andrey Kolobov; Peng Dai; Daniel S. Weld


national conference on artificial intelligence | 2010

SixthSense: fast and reliable recognition of dead ends in MDPs

Andrey Kolobov; Daniel S. Weld

Collaboration


Dive into the Andrey Kolobov's collaboration.

Top Co-Authors

Avatar

Daniel S. Weld

University of Washington

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Brian Milch

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Daniel L. Ong

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge