Arun Venkatraman
Carnegie Mellon University
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Arun Venkatraman.
Autonomous Robots | 2014
Dov Katz; Arun Venkatraman; Moslem Kazemi; J. Andrew Bagnell; Anthony Stentz
Autonomous manipulation in unstructured environments will enable a large variety of exciting and important applications. Despite its promise, autonomous manipulation remains largely unsolved. Even the most rudimentary manipulation task—such as removing objects from a pile—remains challenging for robots. We identify three major challenges that must be addressed to enable autonomous manipulation: object segmentation, action selection, and motion generation. These challenges become more pronounced when unknown man-made or natural objects are cluttered together in a pile. We present a system capable of manipulating unknown objects in such an environment. Our robot is tasked with clearing a table by removing objects from a pile and placing them into a bin. To that end, we address the three aforementioned challenges. Our robot perceives the environment with an RGB-D sensor, segmenting the pile into object hypotheses using non-parametric surface models. Our system then computes the affordances of each object, and selects the best affordance and its associated action to execute. Finally, our robot instantiates the proper compliant motion primitive to safely execute the desired action. For efficient and reliable action selection, we developed a framework for supervised learning of manipulation expertise. To verify the performance of our system, we conducted dozens of trials and report on several hours of experiments involving more than 1,500 interactions. The results show that our learning-based approach for pile manipulation outperforms a common sense heuristic as well as a random strategy, and is on par with human action selection.
robotics: science and systems | 2015
Katharina Mülling; Arun Venkatraman; Jean-Sebastien Valois; John E. Downey; Jeffrey A. Weiss; Shervin Javdani; Martial Hebert; Andrew B. Schwartz; Jennifer L. Collinger; J. Andrew Bagnell
Robot teleoperation systems face a common set of challenges including latency, low-dimensional user commands, and asymmetric control inputs. User control with BrainComputer Interfaces (BCIs) exacerbates these problems through especially noisy and erratic low-dimensional motion commands due to the difficulty in decoding neural activity. We introduce a general framework to address these challenges through a combination of computer vision, user intent inference, and arbitration between the human input and autonomous control schemes. Adjustable levels of assistance allow the system to balance the operator’s capabilities and feelings of comfort and control while compensating for a task’s difficulty. We present experimental results demonstrating significant performance improvement using the shared-control assistance framework on adapted rehabilitation benchmarks with two subjects implanted with intracortical brain-computer interfaces controlling a seven degree-of-freedom robotic manipulator as a prosthetic. Our results further indicate that shared assistance mitigates perceived user difficulty and even enables successful performance on previously infeasible tasks. We showcase the extensibility of our architecture with applications to quality-of-life tasks such as opening a door, pouring liquids from containers, and manipulation with novel objects in densely cluttered environments.
international symposium on experimental robotics | 2016
Arun Venkatraman; Roberto Capobianco; Lerrel Pinto; Martial Hebert; Daniele Nardi; J. Andrew Bagnell
Model-based reinforcement learning (MBRL) plays an important role in developing control strategies for robotic systems. However, when dealing with complex platforms, it is difficult to model systems dynamics with analytic models. While data-driven tools offer an alternative to tackle this problem, collecting data on physical systems is non-trivial. Hence, smart solutions are required to effectively learn dynamics models with small amount of examples. In this paper we present an extension to Data As Demonstrator for handling controlled dynamics in order to improve the multiple-step prediction capabilities of the learned dynamics models. Results show the efficacy of our algorithm in developing LQR, iLQR, and open-loop trajectory-based control strategies on simulated benchmarks as well as physical robot platforms.
national conference on artificial intelligence | 2015
Arun Venkatraman; Martial Hebert; J. Andrew Bagnell
robotics: science and systems | 2013
Dov Katz; Arun Venkatraman; Moslem Kazemi; Drew Bagnell; Anthony Stentz
international conference on machine learning | 2017
Wen Sun; Arun Venkatraman; Geoffrey J. Gordon; Byron Boots; J. Andrew Bagnell
Journal of Neuroengineering and Rehabilitation | 2016
John E. Downey; Jeffrey M. Weiss; Katharina Muelling; Arun Venkatraman; Jean-Sebastien Valois; Martial Hebert; J. Andrew Bagnell; Andrew B. Schwartz; Jennifer L. Collinger
neural information processing systems | 2017
Arun Venkatraman; Nicholas Rhinehart; Wen Sun; Lerrel Pinto; Martial Hebert; Byron Boots; Kris M. Kitani; James Andrew Bagnell
international conference on artificial intelligence and statistics | 2017
Hanzhang Hu; Wen Sun; Arun Venkatraman; Martial Hebert; J. Andrew Bagnell
Autonomous Robots | 2017
Katharina Muelling; Arun Venkatraman; Jean-Sebastien Valois; John E. Downey; Jeffrey A. Weiss; Shervin Javdani; Martial Hebert; Andrew B. Schwartz; Jennifer L. Collinger; J. Andrew Bagnell