Yemin Shi
Peking University
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Yemin Shi.
IEEE Transactions on Multimedia | 2017
Yemin Shi; Yonghong Tian; Yaowei Wang; Tiejun Huang
Learning the spatial-temporal representation of motion information is crucial to human action recognition. Nevertheless, most of the existing features or descriptors cannot capture motion information effectively, especially for long-term motion. To address this problem, this paper proposes a long-term motion descriptor called sequential deep trajectory descriptor (sDTD). Specifically, we project dense trajectories into two-dimensional planes, and subsequently a CNN-RNN network is employed to learn an effective representation for long-term motion. Unlike the popular two-stream ConvNets, the sDTD stream is introduced into a three-stream framework so as to identify actions from a video sequence. Consequently, this three-stream framework can simultaneously capture static spatial features, short-term motion, and long-term motion in the video. Extensive experiments were conducted on three challenging datasets: KTH, HMDB51, and UCF101. Experimental results show that our method achieves state-of-the-art performance on the KTH and UCF101 datasets, and is comparable to the state-of-the-art methods on the HMDB51 dataset.
international conference on multimedia and expo | 2015
Yemin Shi; Wei Zeng; Tiejun Huang; Yaowei Wang
Human action recognition is widely recognized as a challenging task due to the difficulty of effectively characterizing human action in a complex scene. Recent studies have shown that the dense-trajectory-based methods can achieve state-of-the-art recognition results on some challenging datasets. However, in these methods, each dense trajectory is often represented as a vector of coordinates, consequently losing the structural relationship between different trajectories. To address the problem, this paper proposes a novel Deep Trajectory Descriptor (DTD) for action recognition. First, we extract dense trajectories from multiple consecutive frames and then project them onto a canvas. This will result in a “trajectory texture” image which can effectively characterize the relative motion in these frames. Based on these trajectory texture images, a deep neural network (DNN) is utilized to learn a more compact and powerful representation of dense trajectories. In the action recognition system, the DTD descriptor, together with other non-trajectory features such as HOG, HOF and MBH, can provide an effective way to characterize human action from various aspects. Experimental results show that our system can statistically outperform several state-of-the-art approaches, with an average accuracy of 95:6% on KTH and an accuracy of 92.14% on UCF50.
international conference on computer vision | 2017
Yemin Shi; Yonghong Tian; Yaowei Wang; Wei Zeng; Tiejun Huang
arXiv: Computer Vision and Pattern Recognition | 2016
Yemin Shi; Yonghong Tian; Yaowei Wang; Tiejun Huang
arXiv: Computer Vision and Pattern Recognition | 2016
Yemin Shi; Yonghong Tian; Yaowei Wang; Tiejun Huang
international conference on multimedia and expo | 2018
Zongxian Li; Yemin Shi; Yonghong Tian; Wei Zeng; Yaowei Wang
international conference on multimedia and expo | 2018
Yemin Shi; Yonghong Tian; Tiejun Huang; Yaowei Wang
international conference on multimedia and expo | 2018
Yixiong Zou; Yemin Shi; Yaowei Wang; Yu Shu; Qingsheng Yuan; Yonghong Tian
international conference on multimedia and expo | 2018
Yemin Shi; Yaowei Wang; Yixiong Zou; Qingsheng Yuan; Yonghong Tian; Yu Shu
2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR) | 2018
Jia Li; Yunpeng Zhai; Yaowei Wang; Yemin Shi; Yonghong Tian