Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Richard Liaw is active.

Publication


Featured researches published by Richard Liaw.


arXiv: Distributed, Parallel, and Cluster Computing | 2017

Real-Time Machine Learning: The Missing Pieces

Robert Nishihara; Philipp Moritz; Stephanie Wang; Alexey Tumanov; William Paul; Johann Schleier-Smith; Richard Liaw; Mehrdad Niknami; Michael I. Jordan; Ion Stoica

Machine learning applications are increasingly deployed not only to serve predictions using static models, but also as tightly-integrated components of feedback loops involving dynamic, real-time decision making. These applications pose a new set of requirements, none of which are difficult to achieve in isolation, but the combination of which creates a challenge for existing distributed execution frameworks: computation with millisecond latency at high throughput, adaptive construction of arbitrary task graphs, and execution of heterogeneous kernels over diverse sets of resources. We assert that a new distributed execution framework is needed for such ML applications and propose a candidate approach with a proof-of-concept architecture that achieves a 63x performance improvement over a state-of-the-art execution framework for a representative application.


The International Journal of Robotics Research | 2018

SWIRL: A sequential windowed inverse reinforcement learning algorithm for robot tasks with delayed rewards

Sanjay Krishnan; Animesh Garg; Richard Liaw; Brijen Thananjeyan; Lauren Miller; Florian T. Pokorny; Ken Goldberg

We present sequential windowed inverse reinforcement learning (SWIRL), a policy search algorithm that is a hybrid of exploration and demonstration paradigms for robot learning. We apply unsupervised learning to a small number of initial expert demonstrations to structure future autonomous exploration. SWIRL approximates a long time horizon task as a sequence of local reward functions and subtask transition conditions. Over this approximation, SWIRL applies Q-learning to compute a policy that maximizes rewards. Experiments suggest that SWIRL requires significantly fewer rollouts than pure reinforcement learning and fewer expert demonstrations than behavioral cloning to learn a policy. We evaluate SWIRL in two simulated control tasks, parallel parking and a two-link pendulum. On the parallel parking task, SWIRL achieves the maximum reward on the task with 85% fewer rollouts than Q-learning, and one-eight of demonstrations needed by behavioral cloning. We also consider physical experiments on surgical tensioning and cutting deformable sheets using a da Vinci surgical robot. On the deformable tensioning task, SWIRL achieves a 36% relative improvement in reward compared with a baseline of behavioral cloning with segmentation.


arXiv: Distributed, Parallel, and Cluster Computing | 2017

Ray: A Distributed Framework for Emerging AI Applications.

Philipp Moritz; Robert Nishihara; Stephanie Wang; Alexey Tumanov; Richard Liaw; Eric Liang; William Paul; Michael I. Jordan; Ion Stoica


arXiv: Robotics | 2016

HIRL: Hierarchical Inverse Reinforcement Learning for Long-Horizon Tasks with Delayed Rewards.

Sanjay Krishnan; Animesh Garg; Richard Liaw; Lauren Miller; Florian T. Pokorny; Ken Goldberg


arXiv: Artificial Intelligence | 2017

Composing Meta-Policies for Autonomous Driving Using Hierarchical Deep Reinforcement Learning.

Richard Liaw; Sanjay Krishnan; Animesh Garg; Daniel Crankshaw; Joseph E. Gonzalez; Ken Goldberg


arXiv: Artificial Intelligence | 2017

Ray RLLib: A Composable and Scalable Reinforcement Learning Library.

Eric Liang; Richard Liaw; Robert Nishihara; Philipp Moritz; Roy Fox; Joseph E. Gonzalez; Ken Goldberg; Ion Stoica


arXiv: Artificial Intelligence | 2017

Ray RLlib: A Framework for Distributed Reinforcement Learning

Eric Liang; Richard Liaw; Philipp Moritz; Robert Nishihara; Roy Fox; Ken Goldberg; Joseph E. Gonzalez; Michael I. Jordan; Ion Stoica


Archive | 2017

Iterative Noise Injection for Scalable Imitation Learning.

Michael Laskey; Jonathan Lee; Wesley Yu-Shu Hsieh; Richard Liaw; Jeffrey Mahler; Roy Fox; Ken Goldberg


international conference on machine learning | 2018

RLlib: Abstractions for Distributed Reinforcement Learning

Eric Liang; Richard Liaw; Robert Nishihara; Philipp Moritz; Roy Fox; Ken Goldberg; Joseph E. Gonzalez; Michael I. Jordan; Ion Stoica


international conference on machine learning | 2018

AnonLib: Enabling Composition in Distributed Reinforcement Learning

Eric Liang; Richard Liaw; Robert Nishihara; Philipp Moritz; Roy Fox; Ken Goldberg; Joseph E. Gonzalez; Michael I. Jordan; Ion Stoica

Collaboration


Dive into the Richard Liaw's collaboration.

Top Co-Authors

Avatar

Ken Goldberg

University of California

View shared research outputs
Top Co-Authors

Avatar

Philipp Moritz

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Eric Liang

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Ion Stoica

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Roy Fox

Hebrew University of Jerusalem

View shared research outputs
Top Co-Authors

Avatar

Animesh Garg

University of California

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge