Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Yonglong Tian is active.

Publication


Featured researches published by Yonglong Tian.


computer vision and pattern recognition | 2015

DeepID-Net: Deformable deep convolutional neural networks for object detection

Wanli Ouyang; Xiaogang Wang; Xingyu Zeng; Shi Qiu; Ping Luo; Yonglong Tian; Hongsheng Li; Shuo Yang; Zhe Wang; Chen Change Loy; Xiaoou Tang

In this paper, we propose deformable deep convolutional neural networks for generic object detection. This new deep learning object detection framework has innovations in multiple aspects. In the proposed new deep architecture, a new deformation constrained pooling (def-pooling) layer models the deformation of object parts with geometric constraint and penalty. A new pre-training strategy is proposed to learn feature representations more suitable for the object detection task and with good generalization capability. By changing the net structures, training strategies, adding and removing some key components in the detection pipeline, a set of models with large diversity are obtained, which significantly improves the effectiveness of model averaging. The proposed approach improves the mean averaged precision obtained by RCNN [14], which was the state-of-the-art, from 31% to 50.3% on the ILSVRC2014 detection test set. It also outperforms the winner of ILSVRC2014, GoogLeNet, by 6.1%. Detailed component-wise analysis is also provided through extensive experimental evaluation, which provide a global view for people to understand the deep learning object detection pipeline.


computer vision and pattern recognition | 2014

Switchable Deep Network for Pedestrian Detection

Ping Luo; Yonglong Tian; Xiaogang Wang; Xiaoou Tang

In this paper, we propose a Switchable Deep Network (SDN) for pedestrian detection. The SDN automatically learns hierarchical features, salience maps, and mixture representations of different body parts. Pedestrian detection faces the challenges of background clutter and large variations of pedestrian appearance due to pose and viewpoint changes and other factors. One of our key contributions is to propose a Switchable Restricted Boltzmann Machine (SRBM) to explicitly model the complex mixture of visual variations at multiple levels. At the feature levels, it automatically estimates saliency maps for each test sample in order to separate background clutters from discriminative regions for pedestrian detection. At the part and body levels, it is able to infer the most appropriate template for the mixture models of each part and the whole body. We have devised a new generative algorithm to effectively pretrain the SDN and then fine-tune it with back-propagation. Our approach is evaluated on the Caltech and ETH datasets and achieves the state-of-the-art detection performance.


international conference on computer vision | 2015

Deep Learning Strong Parts for Pedestrian Detection

Yonglong Tian; Ping Luo; Xiaogang Wang; Xiaoou Tang

Recent advances in pedestrian detection are attained by transferring the learned features of Convolutional Neural Network (ConvNet) to pedestrians. This ConvNet is typically pre-trained with massive general object categories (e.g. ImageNet). Although these features are able to handle variations such as poses, viewpoints, and lightings, they may fail when pedestrian images with complex occlusions are present. Occlusion handling is one of the most important problem in pedestrian detection. Unlike previous deep models that directly learned a single detector for pedestrian detection, we propose DeepParts, which consists of extensive part detectors. DeepParts has several appealing properties. First, DeepParts can be trained on weakly labeled data, i.e. only pedestrian bounding boxes without part annotations are provided. Second, DeepParts is able to handle low IoU positive proposals that shift away from ground truth. Third, each part detector in DeepParts is a strong detector that can detect pedestrian by observing only a part of a proposal. Extensive experiments in Caltech dataset demonstrate the effectiveness of DeepParts, which yields a new state-of-the-art miss rate of 11:89%, outperforming the second best method by 10%.


acm special interest group on data communication | 2018

RF-based 3D skeletons

Mingmin Zhao; Yonglong Tian; Hang Zhao; Mohammad Abu Alsheikh; Tianhong Li; Rumen Hristov; Zachary Kabelac; Dina Katabi; Antonio Torralba

This paper introduces RF-Pose3D, the first system that infers 3D human skeletons from RF signals. It requires no sensors on the body, and works with multiple people and across walls and occlusions. Further, it generates dynamic skeletons that follow the people as they move, walk or sit. As such, RF-Pose3D provides a significant leap in RF-based sensing and enables new applications in gaming, healthcare, and smart homes. RF-Pose3D is based on a novel convolutional neural network (CNN) architecture that performs high-dimensional convolutions by decomposing them into low-dimensional operations. This property allows the network to efficiently condense the spatio-temporal information in RF signals. The network first zooms in on the individuals in the scene, and crops the RF signals reflected off each person. For each individual, it localizes and tracks their body parts - head, shoulders, arms, wrists, hip, knees, and feet. Our evaluation results show that RF-Pose3D tracks each keypoint on the human body with an average error of 4.2 cm, 4.0 cm, and 4.9 cm along the X, Y, and Z axes respectively. It maintains this accuracy even in the presence of multiple people, and in new environments that it has not seen in the training set. Demo videos are available at our website: http://rfpose3d.csail.mit.edu.


Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies | 2018

RF-Based Fall Monitoring Using Convolutional Neural Networks

Yonglong Tian; Guang-He Lee; Hao He; Chen-Yu Hsu; Dina Katabi

Falls are the top reason for fatal and non-fatal injuries among seniors. Existing solutions are based on wearable fall-alert sensors, but medical research has shown that they are ineffective, mostly because seniors do not wear them. These revelations have led to new passive sensors that infer falls by analyzing Radio Frequency (RF) signals in homes. Seniors can go about their lives as usual without the need to wear any device. While passive monitoring has made major advances, current approaches still cannot deal with the complexities of real-world scenarios. They typically train and test their classifiers on the same people in the same environments, and cannot generalize to new people or new environments. Further, they cannot separate motions from different people and can easily miss a fall in the presence of other motions. To overcome these limitations, we introduce Aryokee, an RF-based fall detection system that uses convolutional neural networks governed by a state machine. Aryokee works with new people and environments unseen in the training set. It also separates different sources of motion to increase robustness. Results from testing Aryokee with over 140 people performing 40 types of activities in 57 different environments show a recall of 94% and a precision of 92% in detecting falls.


computer vision and pattern recognition | 2015

Pedestrian detection aided by deep learning semantic tasks

Yonglong Tian; Ping Luo; Xiaogang Wang; Xiaoou Tang


arXiv: Computer Vision and Pattern Recognition | 2014

DeepID-Net: multi-stage and deformable deep convolutional neural networks for object detection

Wanli Ouyang; Ping Luo; Xingyu Zeng; Shi Qiu; Yonglong Tian; Hongsheng Li; Shuo Yang; Zhe Wang; Yuanjun Xiong; Chen Qian; Zhenyao Zhu; Ruohui Wang; Chen Change Loy; Xiaogang Wang; Xiaoou Tang


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2017

DeepID-Net: Object Detection with Deformable Part Based Convolutional Neural Networks

Wanli Ouyang; Xingyu Zeng; Xiaogang Wang; Shi Qiu; Ping Luo; Yonglong Tian; Hongsheng Li; Shuo Yang; Zhe Wang; Hongyang Li; Kun Wang; Junjie Yan; Chen Change Loy; Xiaoou Tang


computer vision and pattern recognition | 2018

Through-Wall Human Pose Estimation Using Radio Signals

Mingmin Zhao; Tianhong Li; Mohammad Abu Alsheikh; Yonglong Tian; Hang Zhao; Antonio Torralba; Dina Katabi


international conference on machine learning | 2018

Representation Learning on Graphs with Jumping Knowledge Networks

Keyulu Xu; Chengtao Li; Yonglong Tian; Tomohiro Sonobe; Ken-ichi Kawarabayashi; Stefanie Jegelka

Collaboration


Dive into the Yonglong Tian's collaboration.

Top Co-Authors

Avatar

Ping Luo

The Chinese University of Hong Kong

View shared research outputs
Top Co-Authors

Avatar

Xiaogang Wang

The Chinese University of Hong Kong

View shared research outputs
Top Co-Authors

Avatar

Xiaoou Tang

The Chinese University of Hong Kong

View shared research outputs
Top Co-Authors

Avatar

Dina Katabi

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Chen Change Loy

The Chinese University of Hong Kong

View shared research outputs
Top Co-Authors

Avatar

Hongsheng Li

The Chinese University of Hong Kong

View shared research outputs
Top Co-Authors

Avatar

Shi Qiu

The Chinese University of Hong Kong

View shared research outputs
Top Co-Authors

Avatar

Shuo Yang

The Chinese University of Hong Kong

View shared research outputs
Top Co-Authors

Avatar

Xingyu Zeng

The Chinese University of Hong Kong

View shared research outputs
Top Co-Authors

Avatar

Zhe Wang

The Chinese University of Hong Kong

View shared research outputs
Researchain Logo
Decentralizing Knowledge