Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Jordi Gonzàlez is active.

Publication


Featured researches published by Jordi Gonzàlez.


computer vision and pattern recognition | 2010

Harmony potentials for joint classification and segmentation

Josep M. Gonfaus; Xavier Boix; Joost van de Weijer; Andrew D. Bagdanov; Joan Serrat; Jordi Gonzàlez

Hierarchical conditional random fields have been successfully applied to object segmentation. One reason is their ability to incorporate contextual information at different scales. However, these models do not allow multiple labels to be assigned to a single node. At higher scales in the image, this yields an oversimplified model, since multiple classes can be reasonable expected to appear within one region. This simplified model especially limits the impact that observations at larger scales may have on the CRF model. Neglecting the information at larger scales is undesirable since class-label estimates based on these scales are more reliable than at smaller, noisier scales. To address this problem, we propose a new potential, called harmony potential, which can encode any possible combination of class labels. We propose an effective sampling strategy that renders tractable the underlying optimization problem. Results show that our approach obtains state-of-the-art results on two challenging datasets: Pascal VOC 2009 and MSRC-21.


computer vision and pattern recognition | 2011

A coarse-to-fine approach for fast deformable object detection

Marco Pedersoli; Andrea Vedaldi; Jordi Gonzàlez

We present a method that can dramatically accelerate object detection with part based models. The method is based on the observation that the cost of detection is likely to be dominated by the cost of matching each part to the image, and not by the cost of computing the optimal configuration of the parts as commonly assumed. Therefore accelerating detection requires minimizing the number of part-to-image comparisons. To this end we propose a multiple-resolutions hierarchical part based model and a corresponding coarse-to-fine inference procedure that recursively eliminates from the search space unpromising part placements. The method yields a ten-fold speedup over the standard dynamic programming approach and is complementary to the cascade-of-parts approach of [9]. Compared to the latter, our method does not have parameters to be determined empirically, which simplifies its use during the training of the model. Most importantly, the two techniques can be combined to obtain a very significant speedup, of two orders of magnitude in some cases. We evaluate our method extensively on the PASCAL VOC and INRIA datasets, demonstrating a very high increase in the detection speed with little degradation of the accuracy.


european conference on computer vision | 2014

ChaLearn Looking at People Challenge 2014: Dataset and Results

Sergio Escalera; Xavier Baró; Jordi Gonzàlez; Miguel Ángel Bautista; Meysam Madadi; Miguel Reyes; Víctor Ponce-López; Hugo Jair Escalante; Jamie Shotton; Isabelle Guyon

This paper summarizes the ChaLearn Looking at People 2014 challenge data and the results obtained by the participants. The competition was split into three independent tracks: human pose recovery from RGB data, action and interaction recognition from RGB data sequences, and multi-modal gesture recognition from RGB-Depth sequences. For all the tracks, the goal was to perform user-independent recognition in sequences of continuous images using the overlapping Jaccard index as the evaluation measure. In this edition of the ChaLearn challenge, two large novel data sets were made publicly available and the Microsoft Codalab platform were used to manage the competition. Outstanding results were achieved in the three challenge tracks, with accuracy results of 0.20, 0.50, and 0.85 for pose recovery, action/interaction recognition, and multi-modal gesture recognition, respectively.


International Journal of Computer Vision | 2012

Harmony Potentials

Xavier Boix; Josep M. Gonfaus; Joost van de Weijer; Andrew D. Bagdanov; Joan Serrat; Jordi Gonzàlez

The Hierarchical Conditional Random Field (HCRF) model have been successfully applied to a number of image labeling problems, including image segmentation. However, existing HCRF models of image segmentation do not allow multiple classes to be assigned to a single region, which limits their ability to incorporate contextual information across multiple scales. At higher scales in the image, this representation yields an oversimplified model since multiple classes can be reasonably expected to appear within large regions. This simplified model particularly limits the impact of information at higher scales. Since class-label information at these scales is usually more reliable than at lower, noisier scales, neglecting this information is undesirable. To address these issues, we propose a new consistency potential for image labeling problems, which we call the harmony potential. It can encode any possible combination of labels, penalizing only unlikely combinations of classes. We also propose an effective sampling strategy over this expanded label set that renders tractable the underlying optimization problem. Our approach obtains state-of-the-art results on two challenging, standard benchmark datasets for semantic image segmentation: PASCAL VOC 2010, and MSRC-21.


computer vision and pattern recognition | 2012

On partial least squares in head pose estimation: How to simultaneously deal with misalignment

Murad Al Haj; Jordi Gonzàlez; Larry S. Davis

Head pose estimation is a critical problem in many computer vision applications. These include human computer interaction, video surveillance, face and expression recognition. In most prior work on heads pose estimation, the positions of the faces on which the pose is to be estimated are specified manually. Therefore, the results are reported without studying the effect of misalignment. We propose a method based on partial least squares (PLS) regression to estimate pose and solve the alignment problem simultaneously. The contributions of this paper are two-fold: 1) we show that the kernel version of PLS (kPLS) achieves better than state-of-the-art results on the estimation problem and 2) we develop a technique to reduce misalignment based on the learned PLS factors.


international conference on computer vision | 2009

Detection and removal of chromatic moving shadows in surveillance scenarios

Ivan Huerta; Michael Boelstoft Holte; Thomas B. Moeslund; Jordi Gonzàlez

Segmentation in the surveillance domain has to deal with shadows to avoid distortions when detecting moving objects. Most segmentation approaches dealing with shadow detection are typically restricted to penumbra shadows. Therefore, such techniques cannot cope well with umbra shadows. Consequently, umbra shadows are usually detected as part of moving objects. In this paper we present a novel technique based on gradient and colour models for separating chromatic moving cast shadows from detected moving objects. Firstly, both a chromatic invariant colour cone model and an invariant gradient model are built to perform automatic segmentation while detecting potential shadows. In a second step, regions corresponding to potential shadows are grouped by considering “a bluish effect” and an edge partitioning. Lastly, (i) temporal similarities between textures and (ii) spatial similarities between chrominance angle and brightness distortions are analysed for all potential shadow regions in order to finally identify umbra shadows. Unlike other approaches, our method does not make any a-priori assumptions about camera location, surface geometries, surface textures, shapes and types of shadows, objects, and background. Experimental results show the performance and accuracy of our approach in different shadowed materials and illumination conditions.


IEEE Journal of Selected Topics in Signal Processing | 2012

A Local 3-D Motion Descriptor for Multi-View Human Action Recognition from 4-D Spatio-Temporal Interest Points

Michael Boelstoft Holte; Bhaskar Chakraborty; Jordi Gonzàlez; Thomas B. Moeslund

In this paper, we address the problem of human action recognition in reconstructed 3-D data acquired by multi-camera systems. We contribute to this field by introducing a novel 3-D action recognition approach based on detection of 4-D (3-D space


international conference on computer vision | 2015

ChaLearn Looking at People 2015: Apparent Age and Cultural Event Recognition Datasets and Results

Sergio Escalera; Junior Fabian; Pablo Pardo; Xavier Baró; Jordi Gonzàlez; Hugo Jair Escalante; Dusan Misevic; Ulrich K. Steiner; Isabelle Guyon

+


IEEE Transactions on Image Processing | 2011

Accurate Moving Cast Shadow Suppression Based on Local Color Constancy Detection

Ariel Amato; Mikhail Mozerov; Andrew D. Bagdanov; Jordi Gonzàlez

time) spatio-temporal interest points (STIPs) and local description of 3-D motion features. STIPs are detected in multi-view images and extended to 4-D using 3-D reconstructions of the actors and pixel-to-vertex correspondences of the multi-camera setup. Local 3-D motion descriptors, histogram of optical 3-D flow (HOF3D), are extracted from estimated 3-D optical flow in the neighborhood of each 4-D STIP and made view-invariant. The local HOF3D descriptors are divided using 3-D spatial pyramids to capture and improve the discrimination between arm- and leg-based actions. Based on these pyramids of HOF3D descriptors we build a bag-of-words (BoW) vocabulary of human actions, which is compressed and classified using agglomerative information bottleneck (AIB) and support vector machines (SVMs), respectively. Experiments on the publicly available i3DPost and IXMAS datasets show promising state-of-the-art results and validate the performance and view-invariance of the approach.


international conference on distributed smart cameras | 2009

A distributed camera system for multi-resolution surveillance

Nicola Bellotto; Eric Sommerlade; Ben Benfold; Charles Bibby; Ian D. Reid; Daniel Roth; Charles Fernandez; Luc Van Gool; Jordi Gonzàlez

Following previous series on Looking at People (LAP) competitions [14, 13, 11, 12, 2], in 2015 ChaLearn ran two new competitions within the field of Looking at People: (1) age estimation, and (2) cultural event recognition, both in still images. We developed a crowd-sourcing application to collect and label data about the apparent age of people (as opposed to the real age). In terms of cultural event recognition, one hundred categories had to be recognized. These tasks involved scene understanding and human body analysis. This paper summarizes both challenges and data, as well as the results achieved by the participants of the competition. Details of the ChaLearn LAP competitions can be found at http://gesture.chalearn.org/.

Collaboration


Dive into the Jordi Gonzàlez's collaboration.

Top Co-Authors

Avatar

F. Xavier Roca

Autonomous University of Barcelona

View shared research outputs
Top Co-Authors

Avatar

Juan José Villanueva

Autonomous University of Barcelona

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Mikhail Mozerov

Autonomous University of Barcelona

View shared research outputs
Top Co-Authors

Avatar

Daniel Rowe

Autonomous University of Barcelona

View shared research outputs
Top Co-Authors

Avatar

Ignasi Rius

Autonomous University of Barcelona

View shared research outputs
Top Co-Authors

Avatar

Ariel Amato

Autonomous University of Barcelona

View shared research outputs
Top Co-Authors

Avatar

Carles Fernández

Autonomous University of Barcelona

View shared research outputs
Top Co-Authors

Avatar

Pau Baiget

Autonomous University of Barcelona

View shared research outputs
Top Co-Authors

Avatar

Xavier Baró

Open University of Catalonia

View shared research outputs
Researchain Logo
Decentralizing Knowledge