Meysam Madadi
University of Barcelona
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Meysam Madadi.
european conference on computer vision | 2014
Sergio Escalera; Xavier Baró; Jordi Gonzàlez; Miguel Ángel Bautista; Meysam Madadi; Miguel Reyes; Víctor Ponce-López; Hugo Jair Escalante; Jamie Shotton; Isabelle Guyon
This paper summarizes the ChaLearn Looking at People 2014 challenge data and the results obtained by the participants. The competition was split into three independent tracks: human pose recovery from RGB data, action and interaction recognition from RGB data sequences, and multi-modal gesture recognition from RGB-Depth sequences. For all the tracks, the goal was to perform user-independent recognition in sequences of continuous images using the overlapping Jaccard index as the evaluation measure. In this edition of the ChaLearn challenge, two large novel data sets were made publicly available and the Microsoft Codalab platform were used to manage the competition. Outstanding results were achieved in the three challenge tracks, with accuracy results of 0.20, 0.50, and 0.85 for pose recovery, action/interaction recognition, and multi-modal gesture recognition, respectively.
Pattern Recognition Letters | 2015
Meysam Madadi; Sergio Escalera; Jordi Gonzàlez; F. Xavier Roca; Felipe Lumbreras
A system for RGB-Depth human body segmentation and description is presented.Body clusters are automatically computed and a multi-class classifier is trained.3D alignment is performed within an iterative 3D shape context fitting approach.We show robust biometry measurements by applying orthogonal plates to body hull.Results on a novel data set improve segmentation accuracy in relation to RF. This paper presents a novel method extracting biometric measures using depth sensors. Given a multi-part labeled training data, a new subject is aligned to the best model of the dataset, and soft biometrics such as lengths or circumference sizes of limbs and body are computed. The process is performed by training relevant pose clusters, defining a representative model, and fitting a 3D shape context descriptor within an iterative matching procedure. We show robust measures by applying orthogonal plates to body hull. We test our approach in a novel full-body RGB-Depth data set, showing accurate estimation of soft biometrics and better segmentation accuracy in comparison with random forest approach without requiring large training data.
ieee international conference on automatic face gesture recognition | 2017
Meysam Madadi; Sergio Escalera; Alex Carruesco; Carlos Andujar; Xavier Baró; Jordi Gonzàlez
State-of-the-art approaches on hand pose estimation from depth images have reported promising results under quite controlled considerations. In this paper we propose a two-step pipeline for recovering the hand pose from a sequence of depth images. The pipeline has been designed to deal with images taken from any viewpoint and exhibiting a high degree of finger occlusion. In a first step we initialize the hand pose using a part-based model, fitting a set of hand components in the depth images. In a second step we consider temporal data and estimate the parameters of a trained bilinear model consisting of shape and trajectory bases. Results on a synthetic, highly-occluded dataset demonstrate that the proposed method outperforms most recent pose recovering approaches, including those based on CNNs.
Image and Vision Computing | 2018
Meysam Madadi; Sergio Escalera; Alex Carruesco; Carlos Andujar; Xavier Baró; Jordi Gonzàlez
Abstract State-of-the-art approaches on hand pose estimation from depth images have reported promising results under quite controlled considerations. In this paper we propose a two-step pipeline for recovering the hand pose from a sequence of depth images. The pipeline has been designed to deal with images taken from any viewpoint and exhibiting a high degree of finger occlusion. In a first step we initialize the hand pose using a part-based model, fitting a set of hand components in the depth images. In a second step we consider temporal data and estimate the parameters of a trained bilinear model consisting of shape and trajectory bases. We evaluate our approach on a new created synthetic hand dataset along with NYU and MSRA real datasets. Results demonstrate that the proposed method outperforms the most recent pose recovering approaches, including those based on CNNs.
international conference on computer vision | 2017
Jun Wan; Sergio Escalera; Gholamreza Anbarjafari; Hugo Jair Escalante; Xavier Baró; Isabelle Guyon; Meysam Madadi; Jüri Allik; Jelena Gorbova; Chi Lin; Yiliang Xie
computer vision and pattern recognition | 2018
Shanxin Yuan; Guillermo Garcia-Hernando; Björn Stenger; Gyeongsik Moon; Ju Yong Chang; Kyoung Mu Lee; Pavlo Molchanov; Jan Kautz; Sina Honari; Liuhao Ge; Junsong Yuan; Xinghao Chen; Guijin Wang; Fan Yang; Kai Akiyama; Yang Wu; Qingfu Wan; Meysam Madadi; Sergio Escalera; Shile Li; Dongheui Lee; Iason Oikonomidis; Antonis A. Argyros; Tae-Kyun Kim
arXiv: Computer Vision and Pattern Recognition | 2017
Meysam Madadi; Sergio Escalera; Xavier Baró; Jordi Gonzàlez
arXiv: Computer Vision and Pattern Recognition | 2017
Umut Güçlü; Yağmur Güçlütürk; Meysam Madadi; Sergio Escalera; Xavier Baró; Jordi Gonzàlez; Rob van Lier; Marcel A. J. van Gerven
international conference on computer vision | 2017
Yağmur Güçlütürk; Umut Güçlü; Marc Pérez; Hugo Jair Escalante; Xavier Baró; Carlos Andujar; Isabelle Guyon; Julio C. S. Jacques Junior; Meysam Madadi; Sergio Escalera; Marcel A. J. van Gerven; Rob van Lier
international symposium on neural networks | 2017
Hugo Jair Escalante; Isabelle Guyon; Sergio Escalera; Julio Cezar Silveira Jacques; Meysam Madadi; Xavier Baró; Stéphane Ayache; Evelyne Viegas; Yağmur Güçlütürk; Umut Güçlü; Marcel A. J. van Gerven; Rob van Lier