Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Austin Reiter is active.

Publication


Featured researches published by Austin Reiter.


medical image computing and computer assisted intervention | 2012

Feature classification for tracking articulated surgical tools

Austin Reiter; Peter K. Allen; Tao Zhao

Tool tracking is an accepted capability for computer-aided surgical intervention which has numerous applications, both in robotic and manual minimally-invasive procedures. In this paper, we describe a tracking system which learns visual feature descriptors as class-specific landmarks on an articulated tool. The features are localized in 3D using stereo vision and are fused with the robot kinematics to track all of the joints of the dexterous manipulator. Experiments are performed using previously-collected porcine data from a surgical robot.


computer vision and pattern recognition | 2017

Temporal Convolutional Networks for Action Segmentation and Detection

Colin Lea; Michael D. Flynn; René Vidal; Austin Reiter; Gregory D. Hager

The ability to identify and temporally segment fine-grained human actions throughout a video is crucial for robotics, surveillance, education, and beyond. Typical approaches decouple this problem by first extracting local spatiotemporal features from video frames and then feeding them into a temporal classifier that captures high-level temporal patterns. We describe a class of temporal models, which we call Temporal Convolutional Networks (TCNs), that use a hierarchy of temporal convolutions to perform fine-grained action segmentation or detection. Our Encoder-Decoder TCN uses pooling and upsampling to efficiently capture long-range temporal patterns whereas our Dilated TCN uses dilated convolutions. We show that TCNs are capable of capturing action compositions, segment durations, and long-range dependencies, and are over a magnitude faster to train than competing LSTM-based Recurrent Neural Networks. We apply these models to three challenging fine-grained datasets and show large improvements over the state of the art.


Journal of Robotic Surgery | 2013

Lessons learned using the insertable robotic effector platform (IREP) for single port access surgery

Nabil Simaan; Andrea Bajo; Austin Reiter; Long Wang; Peter K. Allen; Dennis L. Fowler

This paper presents the preliminary evaluation of a robotic system for single port access surgery. This system may be deployed through a 15-mm incision. It deploys two surgical arms and a third arm manipulating a stereo-vision module that tracks instrument location. The paper presents the design of the robot along with experiments demonstrating the capabilities of this robot. The evaluation includes use of tasks from fundamentals of laparoscopic surgery, evaluation of telemanipulation accuracy, knot tying, and vision tracking of tools.


european conference on computer vision | 2016

Segmental Spatiotemporal CNNs for Fine-Grained Action Segmentation

Colin Lea; Austin Reiter; René Vidal; Gregory D. Hager

Joint segmentation and classification of fine-grained actions is important for applications of human-robot interaction, video surveillance, and human skill evaluation. However, despite substantial recent progress in large-scale action classification, the performance of state-of-the-art fine-grained action recognition approaches remains low. We propose a model for action segmentation which combines low-level spatiotemporal features with a high-level segmental classifier. Our spatiotemporal CNN is comprised of a spatial component that represents relationships between objects and a temporal component that uses large 1D convolutional filters to capture how object relationships change across time. These features are used in tandem with a semi-Markov model that captures transitions from one action to another. We introduce an efficient constrained segmental inference algorithm for this model that is orders of magnitude faster than the current approach. We highlight the effectiveness of our Segmental Spatiotemporal CNN on cooking and surgical action datasets for which we observe substantially improved performance relative to recent baseline methods.


computer vision and pattern recognition | 2017

Interpretable 3D Human Action Analysis with Temporal Convolutional Networks

Tae Soo Kim; Austin Reiter

The discriminative power of modern deep learning models for 3D human action recognition is growing ever so potent. In conjunction with the recent resurgence of 3D human action representation with 3D skeletons, the quality and the pace of recent progress have been significant. However, the inner workings of state-of-the-art learning based methods in 3D human action recognition still remain mostly black-box. In this work, we propose to use a new class of models known as Temporal Convolutional Neural Networks (TCN) for 3D human action recognition. TCN provides us a way to explicitly learn readily interpretable spatio-temporal representations for 3D human action recognition. Through this work, we wish to take a step towards a spatio-temporal model that is easier to understand, explain and interpret. The resulting model, Res-TCN, achieves state-of-the-art results on the largest 3D human action recognition dataset, NTU-RGBD.


The International Journal of Robotics Research | 2014

Appearance learning for 3D tracking of robotic surgical tools

Austin Reiter; Peter K. Allen; Tao Zhao

In this paper, we present an appearance learning approach which is used to detect and track surgical robotic tools in laparoscopic sequences. By training a robust visual feature descriptor on low-level landmark features, we build a framework for fusing robot kinematics and 3D visual observations to track surgical tools over long periods of time across various types of environment. We demonstrate 3D tracking on multiple types of tool (with different overall appearances) as well as multiple tools simultaneously. We present experimental results using the da Vinci® surgical robot using a combination of both ex-vivo and in-vivo environments.


intelligent robots and systems | 2011

A learning algorithm for visual pose estimation of continuum robots

Austin Reiter; Roger E. Goldman; Andrea Bajo; Konstantinos Iliopoulos; Nabil Simaan; Peter K. Allen

Continuum robots offer significant advantages for surgical intervention due to their down-scalability, dexterity, and structural flexibility. While structural compliance offers a passive way to guard against trauma, it necessitates robust methods for online estimation of the robot configuration in order to enable precise position and manipulation control. In this paper, we address the pose estimation problem by applying a novel mapping of the robot configuration to a feature descriptor space using stereo vision. We generate a mapping of known features through a supervised learning algorithm that relates the feature descriptor to known ground truth. Features are represented in a reduced sub-space, which we call eigen-features. The descriptor provides some robustness to occlusions, which are inherent to surgical environments, and the methodology that we describe can be applied to multi-segment continuum robots for closed-loop control. Experimental validation on a single-segment continuum robot demonstrates the robustness and efficacy of the algorithm for configuration estimation. Results show that the errors are in the range of 1°.


intelligent robots and systems | 2010

An online learning approach to in-vivo tracking using synergistic features

Austin Reiter; Peter K. Allen

In this paper we present an online algorithm for robustly tracking surgical tools in dynamic environments that can assist a surgeon during in-vivo robotic surgery procedures. The next generation of in-vivo robotic surgical devices includes integrated imaging and effector platforms that need to be controlled through real-time visual feedback. Our tracking algorithm learns the appearance of the tool online to account for appearance and perspective changes. In addition, the tracker uses multiple features working together to model the object and discover new areas of the tool as it moves quickly, exits and re-enters the scene, or becomes occluded and requires recovery. The algorithm can persist through changes in lighting and pose by using a memory database, which is built online, using a series of features working together to exploit different aspects of the object being tracked. We present results using real in-vivo imaging data from a human partial nephrectomy.


european conference on computer vision | 2016

Temporal Convolutional Networks: A Unified Approach to Action Segmentation

Colin Lea; René Vidal; Austin Reiter; Gregory D. Hager

The dominant paradigm for video-based action segmentation is composed of two steps: first, compute low-level features for each frame using Dense Trajectories or a Convolutional Neural Network to encode local spatiotemporal information, and second, input these features into a classifier such as a Recurrent Neural Network (RNN) that captures high-level temporal relationships. While often effective, this decoupling requires specifying two separate models, each with their own complexities, and prevents capturing more nuanced long-range spatiotemporal relationships. We propose a unified approach, as demonstrated by our Temporal Convolutional Network (TCN), that hierarchically captures relationships at low-, intermediate-, and high-level time-scales. Our model achieves superior or competitive performance using video or sensor data on three public action segmentation datasets and can be trained in a fraction of the time it takes to train an RNN.


computer vision and pattern recognition | 2015

Beyond spatial pooling: Fine-grained representation learning in multiple domains

Chi Li; Austin Reiter; Gregory D. Hager

Object recognition systems have shown great progress over recent years. However, creating object representations that are robust to changes in viewpoint while capturing local visual details continues to be a challenge. In particular, recent convolutional architectures employ spatial pooling to achieve scale and shift invariances, but they are still sensitive to out-of-plane rotations. In this paper, we formulate a probabilistic framework for analyzing the performance of pooling. This framework suggests two directions for improvement. First, we apply multiple scales of filters coupled with different pooling granularities, and second we make use of color as an additional pooling domain, thereby reducing the sensitivity to spatial deformations. We evaluate our algorithm on the object instance recognition task using two independent publicly available RGB-D datasets, and demonstrate significant improvements over the current state-of-the-art. In addition, we present a new dataset for industrial objects to further validate the effectiveness of our approach versus other state-of-the-art approaches for object recognition using RGB-D data.

Collaboration


Dive into the Austin Reiter's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Masaru Ishii

Johns Hopkins University

View shared research outputs
Top Co-Authors

Avatar

Ayushi Sinha

Johns Hopkins University

View shared research outputs
Top Co-Authors

Avatar

Simon Leonard

Johns Hopkins University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Colin Lea

Johns Hopkins University

View shared research outputs
Top Co-Authors

Avatar

René Vidal

Johns Hopkins University

View shared research outputs
Top Co-Authors

Avatar

Derek Allman

Johns Hopkins University

View shared research outputs
Researchain Logo
Decentralizing Knowledge