Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Kaushik Subramanian is active.

Publication


Featured researches published by Kaushik Subramanian.


Artificial Intelligence | 2014

Abstraction from demonstration for efficient reinforcement learning in high-dimensional domains

Luis C. Cobo; Kaushik Subramanian; Charles Lee Isbell; Aaron D. Lanterman; Andrea Lockerd Thomaz

Reinforcement learning (RL) and learning from demonstration (LfD) are two popular families of algorithms for learning policies for sequential decision problems, but they are often ineffective in high-dimensional domains unless provided with either a great deal of problem-specific domain information or a carefully crafted representation of the state and dynamics of the world. We introduce new approaches inspired by these two techniques, which we broadly call abstraction from demonstration. Our first algorithm, state abstraction from demonstration (AfD), uses a small set of human demonstrations of the task the agent must learn to determine a state-space abstraction. Our second algorithm, abstraction and decomposition from demonstration (ADA), is additionally able to determine a task decomposition from the demonstrations. These abstractions allow RL to scale up to higher-complexity domains, and offer much better performance than LfD with orders of magnitude fewer demonstrations. Using a set of videogame-like domains, we demonstrate that using abstraction from demonstration can obtain up to exponential speed-ups in table-based representations, and polynomial speed-ups when compared with function approximation-based RL algorithms such as fitted Q-learning and LSPI.


Archive | 2010

HELP---Human assisted Efficient Learning Protocols

Kaushik Subramanian

OF THE THESIS HELP Human assisted Efficient Learning Protocols by Kaushik Subramanian Thesis Director: Prof. Zoran Gajic In recent years, there has been a growing attention towards the development of artificial agents that can naturally communicate and interact with humans. The focus has primarily been on creating systems that have the ability to unify advanced learning algorithms along with various natural forms of human interaction (like providing advice, guidance, motivation, punishment, etc). However, despite the progress made, interactive systems are still directed towards researchers and scientists and consequently the everyday human is unable to exploit the potential of these systems. Another undesirable component is that in most cases, the interacting human is required to communicate with the artificial agent a large number of times, making the human often fatigued. In order to improve these systems, this thesis extends prior work and introduces novel approaches via Human-assisted Efficient Learning Protocols (HELP). Three case studies are presented that detail distinct aspects of HELP a) representation of the task to be learned and its associated constraints, b) the efficiency of the learning algorithm used by the artificial agent and c) the unexplored “natural” modes of human interaction. The case studies will show how an artificial agent is able to efficiently learn and perform complex tasks using only a limited number of interactions with a human. Each of these studies involves humans subjects interacting with a real robot and/or simulated agent to learn a particular task. The focus of HELP is to show


neural information processing systems | 2013

Policy Shaping: Integrating Human Feedback with Reinforcement Learning

Shane Griffith; Kaushik Subramanian; Jonathan Scholz; Charles Lee Isbell; Andrea Lockerd Thomaz


international conference on machine learning | 2011

Apprenticeship Learning About Multiple Intentions

Monica Babes; Vukosi Ntsakisi Marivate; Kaushik Subramanian; Michael L. Littman


international conference on machine learning | 2010

Generalizing Apprenticeship Learning across Hypothesis Classes

Thomas J. Walsh; Kaushik Subramanian; Michael L. Littman; Carlos Diuk


adaptive agents and multi-agents systems | 2016

Exploration from Demonstration for Interactive Reinforcement Learning

Kaushik Subramanian; Charles Lee Isbell; Andrea Lockerd Thomaz


international conference on robotics and automation | 2018

Navigating Occluded Intersections with Autonomous Vehicles Using Deep Reinforcement Learning

David Isele; Reza Rahimi; Akansel Cosgun; Kaushik Subramanian; Kikuo Fujimura


national conference on artificial intelligence | 2012

Novel Interaction Strategies for Learning from Teleoperation

Baris Akgun; Kaushik Subramanian; Andrea Lockerd Thomaz


national conference on artificial intelligence | 2011

Learning tasks and skills together from a human teacher

Baris Akgun; Kaushik Subramanian; Jaeeun Shim; Andrea Lockerd Thomaz


national conference on artificial intelligence | 2010

Task space behavior learning for humanoid robots using gaussian mixture models

Kaushik Subramanian

Collaboration


Dive into the Kaushik Subramanian's collaboration.

Top Co-Authors

Avatar

Andrea Lockerd Thomaz

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar

Charles Lee Isbell

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Baris Akgun

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Aaron D. Lanterman

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Brent Harrison

North Carolina State University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

David Isele

University of Pennsylvania

View shared research outputs
Top Co-Authors

Avatar

Himanshu Sahni

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Jonathan Scholz

Georgia Institute of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge