Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Tianmin Shu is active.

Publication


Featured researches published by Tianmin Shu.


computer vision and pattern recognition | 2015

Joint inference of groups, events and human roles in aerial videos

Tianmin Shu; Dan Xie; Brandon Rothrock; Sinisa Todorovic; Song-Chun Zhu

With the advent of drones, aerial video analysis becomes increasingly important; yet, it has received scant attention in the literature. This paper addresses a new problem of parsing low-resolution aerial videos of large spatial areas, in terms of 1) grouping, 2) recognizing events and 3) assigning roles to people engaged in events. We propose a novel framework aimed at conducting joint inference of the above tasks, as reasoning about each in isolation typically fails in our setting. Given noisy tracklets of people and detections of large objects and scene surfaces (e.g., building, grass), we use a spatiotemporal AND-OR graph to drive our joint inference, using Markov Chain Monte Carlo and dynamic programming. We also introduce a new formalism of spatiotemporal templates characterizing latent sub-events. For evaluation, we have collected and released a new aerial videos dataset using a hex-rotor flying over picnic areas rich with group events. Our results demonstrate that we successfully address above inference tasks under challenging conditions.


computer vision and pattern recognition | 2017

CERN: Confidence-Energy Recurrent Network for Group Activity Recognition

Tianmin Shu; Sinisa Todorovic; Song-Chun Zhu

This work is about recognizing human activities occurring in videos at distinct semantic levels, including individual actions, interactions, and group activities. The recognition is realized using a two-level hierarchy of Long Short-Term Memory (LSTM) networks, forming a feed-forward deep architecture, which can be trained end-to-end. In comparison with existing architectures of LSTMs, we make two key contributions giving the name to our approach as Confidence-Energy Recurrent Network – CERN. First, instead of using the common softmax layer for prediction, we specify a novel energy layer (EL) for estimating the energy of our predictions. Second, rather than finding the common minimum-energy class assignment, which may be numerically unstable under uncertainty, we specify that the EL additionally computes the p-values of the solutions, and in this way estimates the most confident energy minimum. The evaluation on the Collective Activity and Volleyball datasets demonstrates: (i) advantages of our two contributions relative to the common softmax and energy-minimization formulations and (ii) a superior performance relative to the state-of-the-art approaches.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2018

Learning and Inferring “Dark Matter” and Predicting Human Intents and Trajectories in Videos

Dan Xie; Tianmin Shu; Sinisa Todorovic; Song-Chun Zhu

This paper presents a method for localizing functional objects and predicting human intents and trajectories in surveillance videos of public spaces, under no supervision in training. People in public spaces are expected to intentionally take shortest paths (subject to obstacles) toward certain objects (e.g., vending machine, picnic table, dumpster etc.) where they can satisfy certain needs (e.g., quench thirst). Since these objects are typically very small or heavily occluded, they cannot be inferred by their visual appearance but indirectly by their influence on peoples trajectories. Therefore, we call them “dark matter”, by analogy to cosmology, since their presence can only be observed as attractive or repulsive “fields” in the public space. A person in the scene is modeled as an intelligent agent engaged in one of the “fields” selected depending his/her intent. An agents trajectory is derived from an Agent-based Lagrangian Mechanics. The agents can change their intents in the middle of motion and thus alter the trajectory. For evaluation, we compiled and annotated a new dataset. The results demonstrate our effectiveness in predicting human intent behaviors and trajectories, and localizing and discovering distinct types of “dark matter” in wide public spaces.


international conference on robotics and automation | 2017

Learning social affordance grammar from videos: Transferring human interactions to human-robot interactions

Tianmin Shu; Xiaofeng Gao; Michael S. Ryoo; Song-Chun Zhu

In this paper, we present a general framework for learning social affordance grammar as a spatiotemporal AND-OR graph (ST-AOG) from RGB-D videos of human interactions, and transfer the grammar to humanoids to enable a real-time motion inference for human-robot interaction (HRI). Based on Gibbs sampling, our weakly supervised grammar learning can automatically construct a hierarchical representation of an interaction with long-term joint sub-tasks of both agents and short term atomic actions of individual agents. Based on a new RGB-D video dataset with rich instances of human interactions, our experiments of Baxter simulation, human evaluation, and real Baxter test demonstrate that the model learned from limited training data successfully generates human-like behaviors in unseen scenarios and outperforms both baselines.


Topics in Cognitive Science | 2018

Perception of Human Interaction Based on Motion Trajectories: From Aerial Videos to Decontextualized Animations

Tianmin Shu; Yujia Peng; Lifeng Fan; Hongjing Lu; Song-Chun Zhu

People are adept at perceiving interactions from movements of simple shapes, but the underlying mechanism remains unknown. Previous studies have often used object movements defined by experimenters. The present study used aerial videos recorded by drones in a real-life environment to generate decontextualized motion stimuli. Motion trajectories of displayed elements were the only visual input. We measured human judgments of interactiveness between two moving elements and the dynamic change in such judgments over time. A hierarchical model was developed to account for human performance in this task. The model represents interactivity using latent variables and learns the distribution of critical movement features that signal potential interactivity. The model provides a good fit to human judgments and can also be generalized to the original Heider-Simmel animations (1944). The model can also synthesize decontextualized animations with a controlled degree of interactiveness, providing a viable tool for studying animacy and social perception.


international joint conference on artificial intelligence | 2016

Learning social affordance for human-robot interaction

Tianmin Shu; Michael S. Ryoo; Song-Chun Zhu


Cognitive Science | 2016

Critical Features of Joint Actions that Signal Human Interaction.

Tianmin Shu; Steven M. Thurman; Dawn Chen; Song-Chun Zhu; Hongjing Lu


international conference on learning representations | 2018

Hierarchical and Interpretable Skill Acquisition in Multi-task Reinforcement Learning

Tianmin Shu; Caiming Xiong; Richard Socher


computer vision and pattern recognition | 2018

Where and Why Are They Looking? Jointly Inferring Human Attention and Intentions in Complex Tasks

Ping Wei; Yang Liu; Tianmin Shu; Nanning Zheng; Song-Chun Zhu


arXiv: Artificial Intelligence | 2018

Interactive Agent Modeling by Learning to Probe

Tianmin Shu; Caiming Xiong; Ying Nian Wu; Song-Chun Zhu

Collaboration


Dive into the Tianmin Shu's collaboration.

Top Co-Authors

Avatar

Song-Chun Zhu

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Dan Xie

University of California

View shared research outputs
Top Co-Authors

Avatar

Hongjing Lu

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Lifeng Fan

University of California

View shared research outputs
Top Co-Authors

Avatar

Michael S. Ryoo

Indiana University Bloomington

View shared research outputs
Top Co-Authors

Avatar

Yujia Peng

University of California

View shared research outputs
Top Co-Authors

Avatar

Brandon Rothrock

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Dawn Chen

University of California

View shared research outputs
Researchain Logo
Decentralizing Knowledge