Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Emanuele Coviello is active.

Publication


Featured researches published by Emanuele Coviello.


acm multimedia | 2010

A new approach to cross-modal multimedia retrieval

Nikhil Rasiwasia; Jose Costa Pereira; Emanuele Coviello; Gabriel Doyle; Gert R. G. Lanckriet; Roger Levy; Nuno Vasconcelos

The problem of joint modeling the text and image components of multimedia documents is studied. The text component is represented as a sample from a hidden topic model, learned with latent Dirichlet allocation, and images are represented as bags of visual (SIFT) features. Two hypotheses are investigated: that 1) there is a benefit to explicitly modeling correlations between the two components, and 2) this modeling is more effective in feature spaces with higher levels of abstraction. Correlations between the two components are learned with canonical correlation analysis. Abstraction is achieved by representing text and images at a more general, semantic level. The two hypotheses are studied in the context of the task of cross-modal document retrieval. This includes retrieving the text that most closely matches a query image, or retrieving the images that most closely match a query text. It is shown that accounting for cross-modal correlations and semantic abstraction both improve retrieval accuracy. The cross-modal model is also shown to outperform state-of-the-art image retrieval systems on a unimodal retrieval task.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2014

On the Role of Correlation and Abstraction in Cross-Modal Multimedia Retrieval

Jose Costa Pereira; Emanuele Coviello; Gabriel Doyle; Nikhil Rasiwasia; Gert R. G. Lanckriet; Roger Levy; Nuno Vasconcelos

The problem of cross-modal retrieval from multimedia repositories is considered. This problem addresses the design of retrieval systems that support queries across content modalities, for example, using an image to search for texts. A mathematical formulation is proposed, equating the design of cross-modal retrieval systems to that of isomorphic feature spaces for different content modalities. Two hypotheses are then investigated regarding the fundamental attributes of these spaces. The first is that low-level cross-modal correlations should be accounted for. The second is that the space should enable semantic abstraction. Three new solutions to the cross-modal retrieval problem are then derived from these hypotheses: correlation matching (CM), an unsupervised method which models cross-modal correlations, semantic matching (SM), a supervised technique that relies on semantic representation, and semantic correlation matching (SCM), which combines both. An extensive evaluation of retrieval performance is conducted to test the validity of the hypotheses. All approaches are shown successful for text retrieval in response to image queries and vice versa. It is concluded that both hypotheses hold, in a complementary form, although evidence in favor of the abstraction hypothesis is stronger than that for correlation.


IEEE Transactions on Audio, Speech, and Language Processing | 2011

Time Series Models for Semantic Music Annotation

Emanuele Coviello; Antoni B. Chan; Gert R. G. Lanckriet

Many state-of-the-art systems for automatic music tagging model music based on bag-of-features representations which give little or no account of temporal dynamics, a key characteristic of the audio signal. We describe a novel approach to automatic music annotation and retrieval that captures temporal (e.g., rhythmical) aspects as well as timbral content. The proposed approach leverages a recently proposed song model that is based on a generative time series model of the musical content-the dynamic texture mixture (DTM) model-that treats fragments of audio as the output of a linear dynamical system. To model characteristic temporal dynamics and timbral content at the tag level, a novel, efficient, and hierarchical expectation-maximization (EM) algorithm for DTM (HEM-DTM) is used to summarize the common information shared by DTMs modeling individual songs associated with a tag. Experiments show learning the semantics of music benefits from modeling temporal dynamics.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2013

Clustering Dynamic Textures with the Hierarchical EM Algorithm for Modeling Video

Adeel Mumtaz; Emanuele Coviello; Gert R. G. Lanckriet; Antoni B. Chan

The dynamic texture (DT) is a probabilistic generative model, defined over space and time, that represents a video as the output of a linear dynamical system (LDS). The DT model has been applied to a wide variety of computer vision problems, such as motion segmentation, motion classification, and video registration. In this paper, we derive a new algorithm for clustering DT models that is based on the hierarchical EM algorithm. The proposed clustering algorithm is capable of both clustering DTs and learning novel DT cluster centers that are representative of the cluster members, in a manner that is consistent with the underlying generative probabilistic model of the DT. We then demonstrate the efficacy of the clustering algorithm on several applications in motion analysis, including hierarchical motion clustering, semantic motion annotation, and bag-of-systems codebook generation.


IEEE Transactions on Audio, Speech, and Language Processing | 2013

A Bag of Systems Representation for Music Auto-Tagging

Katherine Ellis; Emanuele Coviello; Antoni B. Chan; Gert R. G. Lanckriet

We present a content-based automatic tagging system for music that relies on a high-level, concise “Bag of Systems” (BoS) representation of the characteristics of a musical piece. The BoS representation leverages a rich dictionary of musical codewords, where each codeword is a generative model that captures timbral and temporal characteristics of music. Songs are represented as a BoS histogram over codewords, which allows for the use of traditional algorithms for text document retrieval to perform auto-tagging. Compared to estimating a single generative model to directly capture the musical characteristics of songs associated with a tag, the BoS approach offers the flexibility to combine different generative models at various time resolutions through the selection of the BoS codewords. Additionally, decoupling the modeling of audio characteristics from the modeling of tag-specific patterns makes BoS a more robust and rich representation of music. Experiments show that this leads to superior auto-tagging performance.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2015

A Scalable and Accurate Descriptor for Dynamic Textures Using Bag of System Trees

Adeel Mumtaz; Emanuele Coviello; Gert R. G. Lanckriet; Antoni B. Chan

The bag-of-systems (BoS) representation is a descriptor of motion in a video, where dynamic texture (DT) codewords represent the typical motion patterns in spatio-temporal patches extracted from the video. The efficacy of the BoS descriptor depends on the richness of the codebook, which depends on the number of codewords in the codebook. However, for even modest sized codebooks, mapping videos onto the codebook results in a heavy computational load. In this paper we propose the BoS Tree, which constructs a bottom-up hierarchy of codewords that enables efficient mapping of videos to the BoS codebook. By leveraging the tree structure to efficiently index the codewords, the BoS Tree allows for fast look-ups in the codebook and enables the practical use of larger, richer codebooks. We demonstrate the effectiveness of BoS Trees on classification of four video datasets, as well as on annotation of a video dataset and a music dataset. Finally, we show that, although the fast look-ups of BoS Tree result in different descriptors than BoS for the same video, the overall distance (and kernel) matrices are highly correlated resulting in similar classification performance.


computer vision and pattern recognition | 2012

Growing a bag of systems tree for fast and accurate classification

Emanuele Coviello; Adeel Mumtaz; Antoni B. Chan; Gert R. G. Lanckriet

The bag-of-systems (BoS) representation is a descriptor of motion in a video, where dynamic texture (DT) codewords represent the typical motion patterns in spatio-temporal patches extracted from the video. The efficacy of the BoS descriptor depends on the richness of the codebook, which directly depends on the number of codewords in the codebook. However, for even modest sized codebooks, mapping videos onto the codebook results in a heavy computational load. In this paper we propose the BoS Tree, which constructs a bottom-up hierarchy of codewords that enables efficient mapping of videos to the BoS codebook. By leveraging the tree structure to efficiently index the codewords, the BoS Tree allows for fast look-ups in the codebook and enables the practical use of larger, richer codebooks. We demonstrate the effectiveness of BoS Trees on classification of three video datasets, as well as on annotation of a music dataset.


computer vision and pattern recognition | 2010

Clustering dynamic textures with the hierarchical EM algorithm

Antoni B. Chan; Emanuele Coviello; Gert R. G. Lanckriet


Journal of Machine Learning Research | 2014

Clustering hidden Markov models with variational HEM

Emanuele Coviello; Antoni B. Chan; Gert R. G. Lanckriet


international symposium/conference on music information retrieval | 2011

SEMANTIC ANNOTATION AND RETRIEVAL OF MUSIC USING A BAG OF SYSTEMS REPRESENTATION

Katherine Ellis; Emanuele Coviello; Gert R. G. Lanckriet

Collaboration


Dive into the Emanuele Coviello's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Antoni B. Chan

City University of Hong Kong

View shared research outputs
Top Co-Authors

Avatar

Adeel Mumtaz

City University of Hong Kong

View shared research outputs
Top Co-Authors

Avatar

Gabriel Doyle

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Roger Levy

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Tim Chuk

University of Hong Kong

View shared research outputs
Researchain Logo
Decentralizing Knowledge