2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) | 2019

Bounded-Error LQR-Trees

 
 

Abstract


We present a feedback motion planning algorithm, Bounded-Error LQR-Trees, that leverages reinforcement learning theory to find a policy with a bounded amount of error. The algorithm composes locally valid linear-quadratic regulators (LQR) into a nonlinear controller, similar to how LQR-Trees constructs its policy, but minimizes the cost of the constructed policy by minimizing the Bellman Residual, which is estimated in the overlapping regions of LQR controllers. We prove a sample-based upper bound on the true Bellman Residual, and demonstrate a five-fold reduction in cost over previous methods on a simple underactuated nonlinear system.

Volume None
Pages 144-150
DOI 10.1109/IROS40897.2019.8967750
Language English
Journal 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)

Full Text