Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Jian-Fang Hu is active.

Publication


Featured researches published by Jian-Fang Hu.


computer vision and pattern recognition | 2015

Jointly learning heterogeneous features for RGB-D activity recognition

Jian-Fang Hu; Wei-Shi Zheng; Jian-Huang Lai; Jianguo Zhang

In this paper, we focus on heterogeneous features learning for RGB-D activity recognition. We find that features from different channels (RGB, depth) could share some similar hidden structures, and then propose a joint learning model to simultaneously explore the shared and feature-specific components as an instance of heterogeneous multi-task learning. The proposed model formed in a unified framework is capable of: 1) jointly mining a set of subspaces with the same dimensionality to exploit latent shared features across different feature channels, 2) meanwhile, quantifying the shared and feature-specific components of features in the subspaces, and 3) transferring feature-specific intermediate transforms (i-transforms) for learning fusion of heterogeneous features across datasets. To efficiently train the joint model, a three-step iterative optimization algorithm is proposed, followed by a simple inference model. Extensive experimental results on four activity datasets have demonstrated the efficacy of the proposed method. A new RGB-D activity dataset focusing on human-object interaction is further contributed, which presents more challenges for RGB-D activity benchmarking.


european conference on computer vision | 2016

Real-Time RGB-D Activity Prediction by Soft Regression

Jian-Fang Hu; Wei-Shi Zheng; Lianyang Ma; Gang Wang; Jian-Huang Lai

In this paper, we propose a novel approach for predicting ongoing activities captured by a low-cost depth camera. Our approach avoids a usual assumption in existing activity prediction systems that the progress level of ongoing sequence is given. We overcome this limitation by learning a soft label for each subsequence and develop a soft regression framework for activity prediction to learn both predictor and soft labels jointly. In order to make activity prediction work in a real-time manner, we introduce a new RGB-D feature called “local accumulative frame feature (LAFF)”, which can be computed efficiently by constructing an integral feature map. Our experiments on two RGB-D benchmark datasets demonstrate that the proposed regression-based activity prediction model outperforms existing models significantly and also show that the activity prediction on RGB-D sequence is more accurate than that on RGB channel.


international conference on computer vision | 2013

Recognising Human-Object Interaction via Exemplar Based Modelling

Jian-Fang Hu; Wei-Shi Zheng; Jian-Huang Lai; Shaogang Gong; Tao Xiang

Human action can be recognised from a single still image by modelling Human-object interaction (HOI), which infers the mutual spatial structure information between human and object as well as their appearance. Existing approaches rely heavily on accurate detection of human and object, and estimation of human pose. They are thus sensitive to large variations of human poses, occlusion and unsatisfactory detection of small size objects. To overcome this limitation, a novel exemplar based approach is proposed in this work. Our approach learns a set of spatial pose-object interaction exemplars, which are density functions describing how a person is interacting with a manipulated object for different activities spatially in a probabilistic way. A representation based on our HOI exemplar thus has great potential for being robust to the errors in human/object detection and pose estimation. A new framework consists of a proposed exemplar based HOI descriptor and an activity specific matching model that learns the parameters is formulated for robust human activity recognition. Experiments on two benchmark activity datasets demonstrate that the proposed approach obtains state-of-the-art performance.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2017

Jointly Learning Heterogeneous Features for RGB-D Activity Recognition

Jian-Fang Hu; Wei-Shi Zheng; Jian-Huang Lai; Jianguo Zhang

In this paper, we focus on heterogeneous features learning for RGB-D activity recognition. We find that features from different channels (RGB, depth) could share some similar hidden structures, and then propose a joint learning model to simultaneously explore the shared and feature-specific components as an instance of heterogeneous multi-task learning. The proposed model formed in a unified framework is capable of: 1) jointly mining a set of subspaces with the same dimensionality to exploit latent shared features across different feature channels, 2) meanwhile, quantifying the shared and feature-specific components of features in the subspaces, and 3) transferring feature-specific intermediate transforms (i-transforms) for learning fusion of heterogeneous features across datasets. To efficiently train the joint model, a three-step iterative optimization algorithm is proposed, followed by a simple inference model. Extensive experimental results on four activity datasets have demonstrated the efficacy of the proposed method. A new RGB-D activity dataset focusing on human-object interaction is further contributed, which presents more challenges for RGB-D activity benchmarking.


IEEE Transactions on Circuits and Systems for Video Technology | 2016

Exemplar-Based Recognition of Human–Object Interactions

Jian-Fang Hu; Wei-Shi Zheng; Jian-Huang Lai; Shaogang Gong; Tao Xiang

Human action can be recognized from a single still image by modeling human-object interactions (HOIs), which infers the mutual spatial structure information between human and the manipulated object as well as their appearance. Existing approaches rely heavily on accurate detection of human and object and estimation of human pose; they are thus sensitive to large variations of human poses, occlusion, and unsatisfactory detection of small size objects. To overcome this limitation, a novel exemplar-based approach is proposed in this paper. Our approach learns a set of spatial pose-object interaction exemplars, which are probabilistic density functions describing spatially how a person is interacting with a manipulated object for different activities. Specifically, a new framework consisting of an exemplar-based HOI descriptor and an associated matching model is formulated for robust human action recognition in still images. In addition, the framework is extended to perform HOI recognition in videos, where the proposed exemplar representation is used for implicit frame selection to negate irrelevant or noisy frames by temporal structured HOI modeling. Extensive experiments are carried out on two image action datasets and two video action datasets. The results demonstrate the effectiveness of our proposed methods and show that our approach is able to achieve state-of-the-art performance, compared with several recently proposed competitors.


Pattern Recognition | 2017

Multi-task mid-level feature learning for micro-expression recognition

Jiachi He; Jian-Fang Hu; Xi Lu; Wei-Shi Zheng

Due to the short duration and low intensity of micro-expressions, the recognition of micro-expression is still a challenging problem. In this paper, we develop a novel multi-task mid-level feature learning method to enhance the discrimination ability of extracted low-level features by learning a set of class-specific feature mappings, which would be used for generating our mid-level feature representation. Moreover, two weighting schemes are employed to concatenate different mid-level features. We also construct a new mobile micro-expression set to evaluate the performance of the proposed mid-level feature learning framework. The experimental results on two widely used non-mobile micro-expression datasets and one mobile micro-expression set demonstrate that the proposed method can generally improve the performance of the low-level features, and achieve comparable results with the state-of-the-art methods. HighlightsA multi-task mid-level feature learning framework was proposed to improve the discrimination ability of low-level features.A new micro-expression database captured by mobile devices was collected.Extensive experiments are conducted to illustrate that our method can improve the performance of existing low-level features.


Pattern Recognition | 2016

One-pass online learning

Zhaoze Zhou; Wei-Shi Zheng; Jian-Fang Hu; Yong Xu; Jane You

Online learning is very important for processing sequential data and helps alleviate the computation burden on large scale data as well. Especially, one-pass online learning is to predict a new coming samples label and update the model based on the prediction, where each coming sample is used only once and never stored. So far, existing one-pass online learning methods are globally modeled and do not take the local structure of the data distribution into consideration, which is a significant factor of handling the nonlinear data separation case. In this work, we propose a local online learning (LOL) method, a multiple hyperplane Passive Aggressive algorithm integrated with online clustering, so that all local hyperplanes are learned jointly and working cooperatively. This is achieved by formulating a common component as information traffic among multiple hyperplanes in LOL. A joint optimization algorithm is proposed and theoretical analysis on the cumulative error is also provided. Extensive experiments on 11 datasets show that LOL can learn a nonlinear decision boundary, overall achieving notably better performance without using any kernel modeling and second order modeling. HighlightsPropose an one-pass local online learning algorithm (LOL).LOL learns multiple hyperplanes jointly.LOL makes non-linear online learning more effective and accurate.Provide theoretical analysis on the cumulative error of LOL.Experimentally show the effectiveness of the proposed method.


IEEE Transactions on Image Processing | 2017

Global-Local Temporal Saliency Action Prediction

Shaofan Lai; Wei-Shi Zheng; Jian-Fang Hu; Jianguo Zhang

Action prediction on a partially observed action sequence is a very challenging task. To address this challenge, we first design a global-local distance model, where a global-temporal distance compares subsequences as a whole and local-temporal distance focuses on individual segment. Our distance model introduces temporal saliency for each segment to adapt its contribution. Finally, a global-local temporal action prediction model is formulated in order to jointly learn and fuse these two types of distances. Such a prediction model is capable of recognizing action of: 1) an on-going sequence and 2) a sequence with arbitrarily frames missing between the beginning and end (known as gap-filling). Our proposed model is tested and compared with related action prediction models on BIT, UCF11, and HMDB data sets. The results demonstrated the effectiveness of our proposal. In particular, we showed the benefit of our proposed model on predicting unseen action types and the advantage on addressing the gapfilling problem as compared with recently developed action prediction models.


Pattern Recognition | 2017

Sparse transfer for facial shape-from-shading

Jian-Fang Hu; Wei-Shi Zheng; Xiaohua Xie; Jian-Huang Lai

A sparse transfer model was proposed to fuse a set of source face shapes in a selective way in order to assist the shape reconstruction of target face.A non-Lambertian reflectance model was formulated to model the interaction between light and the surface of human face.Extensive experiments were conducted to illustrate that our method can improve the performance of face shape reconstruction, especially when only a small number of target images are available. We present an image-based 3D face shape reconstruction method which transfers shape cues inferred from source face images to guide the reconstruction of the target face. Specifically, a sparse face shape adaption mechanism is used to generate a target-specific reference shape by adaptively and selectively combining source face shapes. This reference shape can also facilitate the reconstruction optimization for the target shape. As an off-line process, each source shape has been derived from a set of given sufficient source images (more than 9) based on a non-Lambertian reflectance model. Such a process allows for the existence of cast shadow and specularity, and more accurately infers the source shape. Guided by the target-specific reference shape, the shape of a target face can be estimated using a small number of images (even only one). The proposed reconstruction method refers to a lighting estimation and an albedo estimation for the target face. No standard 3D shape (such as the high-precision scanned 3D face) is required in the reconstruction process. Compared to the state-of-the-arts including the Photometric Stereo, Tensor Spline, the single reference based method, and the GEM algorithm, the proposed sparse transfer model can produce visually better facial details and obtain smaller reconstruction errors.


international symposium on neural networks | 2016

Facial skin beautification via sparse representation over learned layer dictionary

Xi Lu; Xiaobin Chang; Xiaohua Xie; Jian-Fang Hu; Wei-Shi Zheng

In this paper, we propose a facial skin beautification framework to remove facial spots based on layer dictionary learning and sparse representation. More precisely, we first decompose the face image into three layers: lighting layer, detail layer and color layer. The corresponding detail layer dictionary are learned by using 60 thousands beauty images collected from the Internet. Thereafter, the detail layer of the image is reconstructed by using sparse representation. Moreover, a binary mask obtained from the learned layer is used to transform detail information from original detail layer to the learned one. The experiment results demonstrate that the proposed method is more effective in eliminating moles, flaws and wrinkles in face image compared with representative commercial systems like PicTreat, Portrait+, Portraitrue and MeituPic.

Collaboration


Dive into the Jian-Fang Hu's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Xi Lu

National University of Defense Technology

View shared research outputs
Top Co-Authors

Avatar

Xiaohua Xie

Sun Yat-sen University

View shared research outputs
Top Co-Authors

Avatar

Yongyi Tang

Sun Yat-sen University

View shared research outputs
Top Co-Authors

Avatar

Shaogang Gong

Queen Mary University of London

View shared research outputs
Top Co-Authors

Avatar

Tao Xiang

Queen Mary University of London

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Jiachi He

Sun Yat-sen University

View shared research outputs
Researchain Logo
Decentralizing Knowledge