Guilherme Maeda
University of Sydney
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Guilherme Maeda.
ieee-ras international conference on humanoid robots | 2014
Guilherme Maeda; Marco Ewerton; Rudolf Lioutikov; Heni Ben Amor; Jan Peters; Gerhard Neumann
This paper proposes a probabilistic framework based on movement primitives for robots that work in collaboration with a human coworker. Since the human coworker can execute a variety of unforeseen tasks a requirement of our system is that the robot assistant must be able to adapt and learn new skills on-demand, without the need of an expert programmer. Thus, this paper leverages on the framework of imitation learning and its application to human-robot interaction using the concept of Interaction Primitives (IPs). We introduce the use of Probabilistic Movement Primitives (ProMPs) to devise an interaction method that both recognizes the action of a human and generates the appropriate movement primitive of the robot assistant. We evaluate our method on experiments using a lightweight arm interacting with a human partner and also using motion capture trajectories of two humans assembling a box. The advantages of ProMPs in relation to the original formulation for interaction are exposed and compared.
Autonomous Robots | 2017
Guilherme Maeda; Gerhard Neumann; Marco Ewerton; Rudolf Lioutikov; Oliver Kroemer; Jan Peters
This paper proposes an interaction learning method for collaborative and assistive robots based on movement primitives. The method allows for both action recognition and human–robot movement coordination. It uses imitation learning to construct a mixture model of human–robot interaction primitives. This probabilistic model allows the assistive trajectory of the robot to be inferred from human observations. The method is scalable in relation to the number of tasks and can learn nonlinear correlations between the trajectories that describe the human–robot interaction. We evaluated the method experimentally with a lightweight robot arm in a variety of assistive scenarios, including the coordinated handover of a bottle to a human, and the collaborative assembly of a toolbox. Potential applications of the method are personal caregiver robots, control of intelligent prosthetic devices, and robot coworkers in factories.
ieee-ras international conference on humanoid robots | 2015
Rudolf Lioutikov; Gerhard Neumann; Guilherme Maeda; Jan Peters
Movement primitives are a well established approach for encoding and executing robot movements. While the primitives themselves have been extensively researched, the concept of movement primitive libraries has not received as much attention. Libraries of movement primitives represent the skill set of an agent and can be queried and sequenced in order to solve specific tasks. The goal of this work is to segment unlabeled demonstrations into an optimal set of skills. Our novel approach segments the demonstrations while learning a probabilistic representation of movement primitives. The method differs from current approaches by taking advantage of the often neglected, mutual dependencies between the segments contained in the demonstrations and the primitives to be encoded. Therefore, improving the combined quality of both segmentation and skill learning. Furthermore, our method allows incorporating domain specific insights using heuristics, which are subsequently evaluated and assessed through probabilistic inference methods. We demonstrate our method on a real robot application, where the robot segments demonstrations of a chair assembly task into a skill library. The library is subsequently used to assemble the chair in an order not present in the demonstrations.
IEEE Transactions on Control Systems and Technology | 2015
Guilherme Maeda; Ian R. Manchester; David C. Rye
This paper proposes a new control structure for tasks where explicit disturbance compensation is not only critical for overcoming poor feedback performance but is also challenging due to the complexity and nonrepetitive nature of the interaction between the plant and the environment. The approach proposed uses a particular form of iterative learning control (ILC) to estimate the previous disturbances, which are used as a preview of the disturbance in the next iteration. A disturbance observer is used to compensate for the difference between the ILC prediction and the true disturbance. The controller is evaluated and compared with a proportional controller, with ILC, and with an observer-based controller in extensive field trials using an automated excavator.
field and service robotics | 2014
Guilherme Maeda; David C. Rye; Surya P. N. Singh
This paper introduces a Cartesian impedance control framework in which reaction forces exceeding control authority directly reshape bucket motion during successive excavation passes. This novel approach to excavation results in an iterative process that does not require explicit prediction of terrain forces. This is in contrast to most excavation control approaches that are based on the generation, tracking and re-planning of single-pass tasks where the performance is limited by the accuracy of the prediction. In this view, a final trench profile is achieved iteratively, provided that the forces generated by the excavator are capable of removing some minimum amount of soil, maintaining convergence towards the goal. Field experiments show that a disturbance compensated controller is able to maintain convergence, and that a 2-DOF feedforward controller based on free motion inverse dynamics may not converge due to limited feedback gains.
ISRR (2) | 2018
Guilherme Maeda; Gerhard Neumann; Marco Ewerton; Rudolf Lioutikov; Jan Peters
This paper proposes an interaction learning method suited for semi-autonomous robots that work with or assist a human partner. The method aims at generating a collaborative trajectory of the robot as a function of the current action of the human. The trajectory generation is based on action recognition and prediction of the human movement given intermittent observations of his/her positions under unknown speeds of execution; a problem typically found when using motion capture systems in occluded scenarios. Of particular interest, the ability to predict the human movement while observing the initial part of the trajectory, allows for faster robot reactions. The method is based on probabilistically modelling the coupling between human-robot movement primitives and eliminates the need of time-alignment of the training data while being scalable in relation to the number of tasks. We evaluated the method using a 7-DoF lightweight robot arm equipped with a 5-finger hand in a multi-task collaborative assembly experiment, also comparing results with our previous method based on time-aligned trajectories.
The International Journal of Robotics Research | 2017
Guilherme Maeda; Marco Ewerton; Gerhard Neumann; Rudolf Lioutikov; Jan Peters
This paper proposes a method to achieve fast and fluid human–robot interaction by estimating the progress of the movement of the human. The method allows the progress, also referred to as the phase of the movement, to be estimated even when observations of the human are partial and occluded; a problem typically found when using motion capture systems in cluttered environments. By leveraging on the framework of Interaction Probabilistic Movement Primitives, phase estimation makes it possible to classify the human action, and to generate a corresponding robot trajectory before the human finishes his/her movement. The method is therefore suited for semi-autonomous robots acting as assistants and coworkers. Since observations may be sparse, our method is based on computing the probability of different phase candidates to find the phase that best aligns the Interaction Probabilistic Movement Primitives with the current observations. The method is fundamentally different from approaches based on Dynamic Time Warping that must rely on a consistent stream of measurements at runtime. The resulting framework can achieve phase estimation, action recognition and robot trajectory coordination using a single probabilistic representation. We evaluated the method using a seven-degree-of-freedom lightweight robot arm equipped with a five-finger hand in single and multi-task collaborative experiments. We compare the accuracy achieved by phase estimation with our previous method based on dynamic time warping.
ieee-ras international conference on humanoid robots | 2016
Dorothea Koert; Guilherme Maeda; Rudolf Lioutikov; Gerhard Neumann; Jan Peters
Learning motions from human demonstrations provides an intuitive way for non-expert users to teach tasks to robots. In particular, intelligent robotic co-workers should not only mimic human demonstrations but should also be able to adapt them to varying application scenarios. As such, robots must have the ability to generalize motions to different workspaces, e.g. to avoid obstacles not present during original demonstrations. Towards this goal our work proposes a unified method to (1) generalize robot motions to different workspaces, using a novel formulation of trajectory optimization that explicitly incorporates human demonstrations, and (2) to locally adapt and reuse the optimized solution in the form of a distribution of trajectories. This optimized distribution can be used, online, to quickly satisfy via-points and goals of a specific task. We validate the method using a 7 degrees of freedom (DoF) lightweight arm that grasps and places a ball into different boxes while avoiding obstacles that were not present during the original human demonstrations.
IAS | 2016
Rudolf Lioutikov; Oliver Kroemer; Guilherme Maeda; Jan Peters
Learning to perform complex tasks out of a sequence of simple small demonstrations is a key ability for more flexible robots. In this paper, we present a system that allows for the acquisition of such task executions based on dynamical movement primitives (DMPs). DMPs are a successful approach to encode and generalize robot movements. However, current applications involving DMPs mainly explore movements that, although challenging in terms of dexterity and dimensionality, usually comprise a single continuous movement. This article describes the implementation of a novel system that allows sequencing of simple demonstrations, each one encoded by its own DMP, to achieve a bimanual manipulation task that is too complex to be demonstrated with a single teaching action. As the experimental results show, the resulting system can successfully accomplish a sequenced task of grasping, placing and cutting a vegetable using a setup of a bimanual robot.
intelligent robots and systems | 2011
Guilherme Maeda; Surya P. N. Singh; David C. Rye
Operational space control has a number of desirable characteristics but is sensitive to model accuracy. For heavy machines the dynamics are difficult to model due to their friction and dynamic coupling, thus making full compensation imprecise. This work presents an approach in which a simplified model gives partial compensation via an open-loop feedforward input, pre-calculated in forward simulation. In this way, effects that are difficult to compensate for can be partially corrected without causing instability. Since the reference trajectory is known a priori, dynamic model parameters are tuned in its neighbourhood, reducing the burden of global modelling. The feasibility and performance of this approach is shown experimentally via improved free motion tracking of an excavator arm. This framework further supports efforts for direct impedance control between bucket tip and soil.