Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Oriol Nieto is active.

Publication


Featured researches published by Oriol Nieto.


international conference on acoustics, speech, and signal processing | 2013

Convex non-negative matrix factorization for automatic music structure identification

Oriol Nieto; Tristan Jehan

We propose a novel and fast approach to discover structure in western popular music by using a specific type of matrix factorization that adds a convex constrain to obtain a decomposition that can be interpreted as a set of weighted cluster centroids. We show that these centroids capture the different sections of a musical piece (e.g. verse, chorus) in a more consistent and efficient way than classic non-negative matrix factorization. This technique is capable of identifying the boundaries of the sections and then grouping them into different clusters. Additionally, we evaluate this method on two different datasets and show that it is competitive compared to other music segmentation techniques, outperforming other matrix factorization methods.


international conference on acoustics, speech, and signal processing | 2014

Music segment similarity using 2D-Fourier Magnitude Coefficients

Oriol Nieto; Juan Pablo Bello

Music segmentation is the task of automatically identifying the different segments of a piece. In this work we present a novel approach to cluster the musical segments based on their acoustic similarity by using 2D-Fourier Magnitude Coefficients (2D-FMCs). These coefficients, computed from a chroma representation, significantly simplify the problem of clustering the different segments since they are key transposition and phase shift invariant. We explore various strategies to obtain the 2D-FMC patches that represent entire segments and apply k-means to label them. Finally, we discuss possible ways of estimating k and compare our competitive results with the current state of the art.


conference on recommender systems | 2017

A Deep Multimodal Approach for Cold-start Music Recommendation

Sergio Oramas; Oriol Nieto; Mohamed Sordo; Xavier Serra

An increasing amount of digital music is being published daily. Music streaming services often ingest all available music, but this poses a challenge: how to recommend new artists for which prior knowledge is scarce? In this work we aim to address this so-called cold-start problem by combining text and audio information with user feedback data using deep network architectures. Our method is divided into three steps. First, artist embeddings are learned from biographies by combining semantics, text features, and aggregated usage data. Second, track embeddings are learned from the audio signal and available feedback data. Finally, artist and track embeddings are combined in a multimodal network. Results suggest that both splitting the recommendation problem between feature levels (i.e., artist metadata and audio track), and merging feature embeddings in a multimodal approach improve the accuracy of the recommendations.


web search and data mining | 2018

Predicting Audio Advertisement Quality

Samaneh Ebrahimi; Hossein Vahabi; Matthew Prockup; Oriol Nieto

Online audio advertising is a particular form of advertising used abundantly in online music streaming services. In these platforms, which tend to host tens of thousands of unique audio advertisements (ads), providing high quality ads ensures a better user experience and results in longer user engagement. Therefore, the automatic assessment of these ads is an important step toward audio ads ranking and better audio ads creation. In this paper we propose one way to measure the quality of the audio ads using a proxy metric called Long Click Rate (LCR), which is defined by the amount of time a user engages with the follow-up display ad (that is shown while the audio ad is playing) divided by the impressions. We later focus on predicting the audio ad quality using only acoustic features such as harmony, rhythm, and timbre of the audio, extracted from the raw waveform. We discuss how the characteristics of the sound can be connected to concepts such as the clarity of the audio ad message, its trustworthiness, etc. Finally, we propose a new deep learning model for audio ad quality prediction, which outperforms the other discussed models trained on hand-crafted features. To the best of our knowledge, this is the first large-scale audio ad quality prediction study.


Frontiers in Psychology | 2017

Evaluating Hierarchical Structure in Music Annotations

Brian McFee; Oriol Nieto; Morwaread Farbood; Juan Pablo Bello

Music exhibits structure at multiple scales, ranging from motifs to large-scale functional components. When inferring the structure of a piece, different listeners may attend to different temporal scales, which can result in disagreements when they describe the same piece. In the field of music informatics research (MIR), it is common to use corpora annotated with structural boundaries at different levels. By quantifying disagreements between multiple annotators, previous research has yielded several insights relevant to the study of music cognition. First, annotators tend to agree when structural boundaries are ambiguous. Second, this ambiguity seems to depend on musical features, time scale, and genre. Furthermore, it is possible to tune current annotation evaluation metrics to better align with these perceptual differences. However, previous work has not directly analyzed the effects of hierarchical structure because the existing methods for comparing structural annotations are designed for “flat” descriptions, and do not readily generalize to hierarchical annotations. In this paper, we extend and generalize previous work on the evaluation of hierarchical descriptions of musical structure. We derive an evaluation metric which can compare hierarchical annotations holistically across multiple levels. sing this metric, we investigate inter-annotator agreement on the multilevel annotations of two different music corpora, investigate the influence of acoustic properties on hierarchical annotations, and evaluate existing hierarchical segmentation algorithms against the distribution of inter-annotator agreement.


Proceedings of the 14th Python in Science Conference | 2015

librosa: Audio and Music Signal Analysis in Python

Brian McFeek; Colin Raffel; Dawen Liang; Matt McVicar; Eric Battenberg; Oriol Nieto


international symposium/conference on music information retrieval | 2014

JAMS: A JSON Annotated Music Specification for Reproducible MIR Research.

Eric J. Humphrey; Justin Salamon; Oriol Nieto; Jon Forsyth; Rachel M. Bittner; Juan Pablo Bello


international symposium/conference on music information retrieval | 2014

MIR_EVAL: A Transparent Implementation of Common MIR Metrics.

Colin Raffel; Brian McFee; Eric J. Humphrey; Justin Salamon; Oriol Nieto; Dawen Liang; Daniel P. W. Ellis


international symposium/conference on music information retrieval | 2013

Data Driven and Discriminative Projections for Large-Scale Cover Song Identification.

Eric J. Humphrey; Oriol Nieto; Juan Pablo Bello


international symposium/conference on music information retrieval | 2014

Identifying polyphonic musical patterns from audio recordings using music segmentation techniques

Oriol Nieto; Morwaread Farbood

Collaboration


Dive into the Oriol Nieto's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Brian McFee

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Xavier Serra

Pompeu Fabra University

View shared research outputs
Top Co-Authors

Avatar

Matt McVicar

National Institute of Advanced Industrial Science and Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge