Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Adriana Romero is active.

Publication


Featured researches published by Adriana Romero.


IEEE Transactions on Geoscience and Remote Sensing | 2016

Unsupervised Deep Feature Extraction for Remote Sensing Image Classification

Adriana Romero; Carlo Gatta; Gustau Camps-Valls

This paper introduces the use of single-layer and deep convolutional networks for remote sensing data analysis. Direct application to multi- and hyperspectral imagery of supervised (shallow or deep) convolutional networks is very challenging given the high input data dimensionality and the relatively small amount of available labeled data. Therefore, we propose the use of greedy layerwise unsupervised pretraining coupled with a highly efficient algorithm for unsupervised learning of sparse features. The algorithm is rooted on sparse representations and enforces both population and lifetime sparsity of the extracted features, simultaneously. We successfully illustrate the expressive power of the extracted representations in several scenarios: classification of aerial scenes, as well as land-use classification in very high resolution or land-cover classification from multi- and hyperspectral images. The proposed algorithm clearly outperforms standard principal component analysis (PCA) and its kernel counterpart (kPCA), as well as current state-of-the-art algorithms of aerial classification, while being extremely computationally efficient at learning representations of data. Results show that single-layer convolutional networks can extract powerful discriminative features only when the receptive field accounts for neighboring pixels and are preferred when the classification requires high resolution and detailed results. However, deep architectures significantly outperform single-layer variants, capturing increasing levels of abstraction and complexity throughout the feature hierarchy.


computer vision and pattern recognition | 2017

The One Hundred Layers Tiramisu: Fully Convolutional DenseNets for Semantic Segmentation

Simon Jégou; Michal Drozdzal; David Vázquez; Adriana Romero; Yoshua Bengio

State-of-the-art approaches for semantic image segmentation are built on Convolutional Neural Networks (CNNs). The typical segmentation architecture is composed of (a) a downsampling path responsible for extracting coarse semantic features, followed by (b) an upsampling path trained to recover the input image resolution at the output of the model and, optionally, (c) a post-processing module (e.g. Conditional Random Fields) to refine the model predictions.,,,,,, Recently, a new CNN architecture, Densely Connected Convolutional Networks (DenseNets), has shown excellent results on image classification tasks. The idea of DenseNets is based on the observation that if each layer is directly connected to every other layer in a feed-forward fashion then the network will be more accurate and easier to train.,,,,,, In this paper, we extend DenseNets to deal with the problem of semantic segmentation. We achieve state-of-the-art results on urban scene benchmark datasets such as CamVid and Gatech, without any further post-processing module nor pretraining. Moreover, due to smart construction of the model, our approach has much less parameters than currently published best entries for these datasets.


international conference on computer vision | 2011

Simultaneous correspondence and non-rigid 3D reconstruction of the coronary tree from single X-ray images

Eduard Serradell; Adriana Romero; Ruben Leta; Carlo Gatta; Francesc Moreno-Noguer

We present a novel approach to simultaneously reconstruct the 3D structure of a non-rigid coronary tree and estimate point correspondences between an input X-ray image and a reference 3D shape. At the core of our approach lies an optimization scheme that iteratively fits a generative 3D model of increasing complexity and guides the matching process. As a result, and in contrast to existing approaches that assume rigidity or quasi-rigidity of the structure, our method is able to retrieve large non-linear deformations even when the input data is corrupted by the presence of noise and partial occlusions. We extensively evaluate our approach under synthetic and real data and demonstrate a remarkable improvement compared to state-of-the-art.


workshop on hyperspectral image and signal processing evolution in remote sensing | 2014

Unsupervised deep feature extraction of hyperspectral images

Adriana Romero; Carlo Gatta; Gustavo Camps-Valls

This paper presents an effective unsupervised sparse feature learning algorithm to train deep convolutional networks on hyperspectral images. Deep convolutional hierarchical representations are learned and then used for pixel classification. Features in lower layers present less abstract representations of data, while higher layers represent more abstract and complex characteristics. We successfully illustrate the performance of the extracted representations in a challenging AVIRIS hyperspectral image classification problem, compared to standard dimensionality reduction methods like principal component analysis (PCA) and its kernel counterpart (kPCA). The proposed method largely outperforms the previous state-of-the-art results on the same experimental setting. Results show that single layer networks can extract powerful discriminative features only when the receptive field accounts for neighboring pixels. Regarding the deep architecture, we can conclude that: (1) additional layers in a deep architecture significantly improve the performance w.r.t. single layer variants; (2) the max-pooling step in each layer is mandatory to achieve satisfactory results; and (3) the performance gain w.r.t. the number of layers is upper bounded, since the spatial resolution is reduced at each pooling, resulting in too spatially coarse output features.


computer vision and pattern recognition | 2014

Unrolling Loopy Top-Down Semantic Feedback in Convolutional Deep Networks

Carlo Gatta; Adriana Romero; Joost van de Veijer

In this paper, we propose a novel way to perform top-down semantic feedback in convolutional deep networks for efficient and accurate image parsing. We also show how to add global appearance/semantic features, which have shown to improve image parsing performance in state-of-the-art methods, and was not present in previous convolutional approaches. The proposed method is characterised by an efficient training and a sufficiently fast testing. We use the well known SIFTflow dataset to numerically show the advantages provided by our contributions, and to compare with state-of-the-art image parsing convolutional based approaches.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2015

Meta-Parameter Free Unsupervised Sparse Feature Learning

Adriana Romero; Petia Radeva; Carlo Gatta

We propose a meta-parameter free, off-the-shelf, simple and fast unsupervised feature learning algorithm, which exploits a new way of optimizing for sparsity. Experiments on CIFAR-10, STL-10 and UCMerced show that the method achieves the state-of-the-art performance, providing discriminative features that generalize well.


Medical Image Analysis | 2018

Learning normalized inputs for iterative estimation in medical image segmentation

Michal Drozdzal; Gabriel Chartrand; Eugene Vorontsov; Mahsa Shakeri; Lisa Di Jorio; An Tang; Adriana Romero; Yoshua Bengio; Chris Pal; Samuel Kadoury

HighlightsImage segmentation pipeline based on Fully Convolutional Networks (FCN) and ResNets is proposed.FCN can serve as a pre‐processor to normalize medical imaging input data.A trainable FCN is an alternative to hand‐designed, modality specific pre‐processing steps.Our pipeline obtains or matches state‐of‐the‐art performance on 3 segmentation datasets. Graphical abstract Figure. No Caption available. Abstract In this paper, we introduce a simple, yet powerful pipeline for medical image segmentation that combines Fully Convolutional Networks (FCNs) with Fully Convolutional Residual Networks (FC‐ResNets). We propose and examine a design that takes particular advantage of recent advances in the understanding of both Convolutional Neural Networks as well as ResNets. Our approach focuses upon the importance of a trainable pre‐processing when using FC‐ResNets and we show that a low‐capacity FCN model can serve as a pre‐processor to normalize medical input data. In our image segmentation pipeline, we use FCNs to obtain normalized images, which are then iteratively refined by means of a FC‐ResNet to generate a segmentation prediction. As in other fully convolutional approaches, our pipeline can be used off‐the‐shelf on different image modalities. We show that using this pipeline, we exhibit state‐of‐the‐art performance on the challenging Electron Microscopy benchmark, when compared to other 2D methods. We improve segmentation results on CT images of liver lesions, when contrasting with standard FCN methods. Moreover, when applying our 2D pipeline on a challenging 3D MRI prostate segmentation challenge we reach results that are competitive even when compared to 3D methods. The obtained results illustrate the strong potential and versatility of the pipeline by achieving accurate segmentations on a variety of image modalities and different anatomical regions.


international geoscience and remote sensing symposium | 2015

Shared feature representations of LiDAR and optical images: Trading sparsity for semantic discrimination

Manuel Campos-Taberner; Adriana Romero; Carlo Gatta; Gustau Camps-Valls

This paper studies the level of complementary information conveyed by extremely high resolution LiDAR and optical images. We pursue this goal following an indirect approach via unsupervised spatial-spectral feature extraction. We used a recently presented unsupervised convolutional neural network trained to enforce both population and lifetime spar-sity in the feature representation. We derived independent and joint feature representations, and analyzed the sparsity scores and the discriminative power. Interestingly, the obtained results revealed that the RGB+LiDAR representation is no longer sparse, and the derived basis functions merge color and elevation yielding a set of more expressive colored edge filters. The joint feature representation is also more discriminative when used for clustering and topological data visualization.


iberian conference on pattern recognition and image analysis | 2013

Do We Really Need All These Neurons

Adriana Romero; Carlo Gatta

Restricted Boltzmann Machines (RBMs) are generative neural networks that have received much attention recently. In particular, choosing the appropriate number of hidden units is important as it might hinder their representative power. According to the literature, RBM require numerous hidden units to approximate any distribution properly. In this paper, we present an experiment to determine whether such amount of hidden units is required in a classification context. We then propose an incremental algorithm that trains RBM reusing the previously trained parameters using a trade-off measure to determine the appropriate number of hidden units. Results on the MNIST and OCR letters databases show that using a number of hidden units, which is one order of magnitude smaller than the literature estimate, suffices to achieve similar performance. Moreover, the proposed algorithm allows to estimate the required number of hidden units without the need of training many RBM from scratch.


Journal of Healthcare Engineering | 2017

A Benchmark for Endoluminal Scene Segmentation of Colonoscopy Images

David Vázquez; Jorge Bernal; F. Javier Sánchez; Gloria Fernández-Esparrach; Antonio M. López; Adriana Romero; Michal Drozdzal; Aaron C. Courville

Colorectal cancer (CRC) is the third cause of cancer death worldwide. Currently, the standard approach to reduce CRC-related mortality is to perform regular screening in search for polyps and colonoscopy is the screening tool of choice. The main limitations of this screening procedure are polyp miss rate and the inability to perform visual assessment of polyp malignancy. These drawbacks can be reduced by designing decision support systems (DSS) aiming to help clinicians in the different stages of the procedure by providing endoluminal scene segmentation. Thus, in this paper, we introduce an extended benchmark of colonoscopy image segmentation, with the hope of establishing a new strong benchmark for colonoscopy image analysis research. The proposed dataset consists of 4 relevant classes to inspect the endoluminal scene, targeting different clinical needs. Together with the dataset and taking advantage of advances in semantic segmentation literature, we provide new baselines by training standard fully convolutional networks (FCNs). We perform a comparative study to show that FCNs significantly outperform, without any further postprocessing, prior results in endoluminal scene segmentation, especially with respect to polyp segmentation and localization.

Collaboration


Dive into the Adriana Romero's collaboration.

Top Co-Authors

Avatar

Carlo Gatta

Autonomous University of Barcelona

View shared research outputs
Top Co-Authors

Avatar

Yoshua Bengio

Université de Montréal

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Petia Radeva

University of Barcelona

View shared research outputs
Top Co-Authors

Avatar

David Vázquez

Autonomous University of Barcelona

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Antonio M. López

Autonomous University of Barcelona

View shared research outputs
Top Co-Authors

Avatar

Eduard Serradell

Spanish National Research Council

View shared research outputs
Top Co-Authors

Avatar

F. Javier Sánchez

Autonomous University of Barcelona

View shared research outputs
Top Co-Authors

Avatar

Francesc Moreno-Noguer

Spanish National Research Council

View shared research outputs
Researchain Logo
Decentralizing Knowledge