Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Cynthia C. S. Liem is active.

Publication


Featured researches published by Cynthia C. S. Liem.


Proceedings of the 1st international ACM workshop on Music information retrieval with user-centered and multimodal strategies | 2011

The need for music information retrieval with user-centered and multimodal strategies

Cynthia C. S. Liem; Meinard Müller; Douglas Eck; George Tzanetakis; Alan Hanjalic

Music is a widely enjoyed content type, existing in many multifaceted representations. With the digital information age, a lot of digitized music information has theoretically become available at the users fingertips. However, the abundance of information is too large-scaled and too diverse to annotate, oversee and present in a consistent and human manner, motivating the development of automated Music Information Retrieval (Music-IR) techniques. In this paper, we encourage to consider music content beyond a monomodal audio signal and argue that Music-IR approaches with multimodal and user-centered strategies are necessary to serve real-life usage patterns and maintain and improve accessibility of digital music data. After discussing relevant existing work in these directions, we show that the field of Music-IR faces similar challenges as neighboring fields, and thus suggest opportunities for joint collaboration and mutual inspiration.


acm sigmm conference on multimedia systems | 2013

A professionally annotated and enriched multimodal data set on popular music

Markus Schedl; Nicola Orio; Cynthia C. S. Liem; Geoffroy Peeters

This paper presents the MusiClef data set, a multimodal data set of professionally annotated music. It includes editorial metadata about songs, albums, and artists, as well as MusicBrainz identifiers to facilitate linking to other data sets. In addition, several state-of-the-art audio features are provided. Different sets of annotations and music context data -- collaboratively generated user tags, web pages about artists and albums, and the annotation labels provided by music experts -- are included too. Versions of this data set were used in the MusiClef evaluation campaigns in 2011 and 2012 for auto-tagging tasks. We report on the motivation for the data set, on its composition, on related sets, and on the evaluation campaigns in which versions of the set were already used. These campaigns likewise represent one use case, i.e. music auto-tagging, of the data set. The complete data set is publicly available for download at http://www.cp.jku.at/musiclef.


International Journal of Multimedia Information Retrieval | 2013

When music makes a scene

Cynthia C. S. Liem; Martha Larson; Alan Hanjalic

Music frequently occurs as an important reinforcing and meaning-creating element in multimodal human experiences. This way, cross-modal connotative associations are established, which are actively exploited in professional multimedia productions. A lay user who wants to use music in a similar way may have a result in mind, but may lack the right musical vocabulary to express the corresponding information need. However, if the connotative associations between music and visual narrative are strong enough, characterizations of music in terms of a narrative multimedia context can be envisioned. In this article, we present the outcomes of a user study considering this problem. Through a survey for which respondents were recruited via crowdsourcing methods, we solicited descriptions of cinematic situations for which fragments of royalty-free production music would be suitable soundtracks. As we will show, these descriptions can reliably be recognized by other respondents as belonging to the music fragments that triggered them. We do not fix any description vocabulary beforehand, but rather give respondents a lot of freedom to express their associations. From these free descriptions, common narrative elements emerge that can be generalized in terms of event structure. The insights gained this way can be used to inform new conceptual foundations for supervised methods, and to provide new perspectives on meaningful and multimedia context-aware querying, retrieval and analysis.


international conference on multimedia and expo | 2015

PHENICX: Innovating the classical music experience

Cynthia C. S. Liem; Emilia Gómez; Markus Schedl

PHENICX (“Performances as Highly Enriched aNd Interactive Concert eXperiences”) is an EU FP7 project that lasts from February 2013 to January 2016. It focuses on creating novel digital concert experiences, improving the accessibility of classical music concert performances by enhancing and enriching them in novel multimodal ways. This requires a usercentered approach throughout the project. After introducing the project, we discuss its goals, the technological challenges it offers, and current scientific and technological outcomes. Subsequently, we discuss how integrated prototypes combine several technological advances in the project into coherent user-ready interfaces, offering novel ways to experience the timeline of a concert, and rediscover and re-experience it afterwards. Finally, we discuss how PHENICX outcomes have been demonstrated live in concert halls.


Multimodal Music Processing (M. Müller and M. Goto, eds.), Seminar 11041 of Dagstuhl Follow-Ups, Schloss Dagstuhl, Germany | 2012

Music Information Retrieval: An Inspirational Guide to Transfer from Related Disciplines

Felix Weninger; Björn W. Schuller; Cynthia C. S. Liem; Frank Kurth; Alan Hanjalic

The emerging field of Music Information Retrieval (MIR) has been influenced by neighboring domains in signal processing and machine learning, including automatic speech recognition, image processing and text information retrieval. In this contribution, we start with concrete examples for methodology transfer between speech and music processing, oriented on the building blocks of pattern recognition: preprocessing, feature extraction, and classification/decoding. We then assume a higher level viewpoint when describing sources of mutual inspiration derived from text and image information retrieval. We conclude that dealing with the peculiarities of music in MIR research has contributed to advancing the state-of-the-art in other fields, and that many future challenges in MIR are strikingly similar to those that other research areas have been facing.


Computer Vision and Image Understanding | 2016

On detecting the playing/non-playing activity of musicians in symphonic music videos

Alessio Bazzica; Cynthia C. S. Liem; Alan Hanjalic

We propose a semi-automatic annotation system for large symphonic orchestras videos.We leverage video redundancy, image clustering, and human annotation.Our method successfully deals with several intra-class variability issues.Human annotation effort reduced while maintaining high level of output quality.Comprehensive analysis of the impact of different modules on the overall performance. Information on whether a musician in a large symphonic orchestra plays her instrument at a given time stamp or not is valuable for a wide variety of applications aiming at mimicking and enriching the classical music concert experience on modern multimedia platforms. In this work, we propose a novel method for generating playing/non-playing labels per musician over time by efficiently and effectively combining an automatic analysis of the video recording of a symphonic concert and human annotation. In this way, we address the inherent deficiencies of traditional audio-only approaches in the case of large ensembles, as well as those of standard human action recognition methods based on visual models. The potential of our approach is demonstrated on two representative concert videos (about 7 hours of content) using a synchronized symbolic music score as ground truth. In order to identify the open challenges and the limitations of the proposed method, we carry out a detailed investigation of how different modules of the system affect the overall performance.


cross language evaluation forum | 2012

MusiClef: multimodal music tagging task

Nicola Orio; Cynthia C. S. Liem; Geoffroy Peeters; Markus Schedl

MusiClef is a multimodal music benchmarking initiative that will be running a MediaEval 2012 Brave New Task on Multimodal Music Tagging. This paper describes the setup of this task, showing how it complements existing benchmarking initiatives and fosters less explored methodological directions in Music Information Retrieval. MusiClef deals with a concrete use case, encourages multimodal approaches based on these, and strives for transparency of results as much as possible. Transparency is encouraged at several levels and stages, from the feature extraction procedure up to the evaluation phase, in which a dedicated categorization of ground truth tags will be used to deepen the understanding of the relation between the proposed approaches and experimental results.


arXiv: Learning | 2018

Transfer Learning of Artist Group Factors to Musical Genre Classification

Jaehun Kim; Minz Won; Xavier Serra; Cynthia C. S. Liem

The automated recognition of music genres from audio information is a challenging problem, as genre labels are subjective and noisy. Artist labels are less subjective and less noisy, while certain artists may relate more strongly to certain genres. At the same time, at prediction time, it is not guaranteed that artist labels are available for a given audio segment. Therefore, in this work, we propose to apply the transfer learning framework, learning artist-related information which will be used at inference time for genre classification. We consider different types of artist-related information, expressed through artist group factors, which will allow for more efficient learning and stronger robustness to potential label noise. Furthermore, we investigate how to achieve the highest validation accuracy on the given FMA dataset, by experimenting with various kinds of transfer methods, including single-task transfer, multi-task transfer and finally multi-task learning.


acm multimedia | 2012

MuseSync: standing on the shoulders of Hollywood

Cynthia C. S. Liem; Alessio Bazzica; Alan Hanjalic

In this extended abstract, we present a novel story-driven approach to soundtrack retrieval for user-generated videos. Cinematic knowledge on cross-modal associations is exploited through folksonomic story text retrieval from collaborative online metadata resources. Subsequently, audiovisual synchronization is applied based on high-level features described by users. The approach is demonstrated in the MuseSync prototype system.


Multimodal Music Processing | 2012

Music Information Technology and Professional Stakeholder Audiences: Mind the Adoption Gap

Cynthia C. S. Liem; Andreas Rauber; Thomas Lidy; Richard Lewis; Christopher Raphael; Joshua D. Reiss; Tim Crawford; Alan Hanjalic

The academic discipline focusing on the processing and organization of digital music information, commonly known as Music Information Retrieval (MIR), has multidisciplinary roots and interests. Thus, MIR technologies have the potential to have impact across disciplinary boundaries and to enhance the handling of music information in many different user communities. However, in practice, many MIR research agenda items appear to have a hard time leaving the lab in order to be widely adopted by their intended audiences. On one hand, this is because the MIR field still is relatively young, and technologies therefore need to mature. On the other hand, there may be deeper, more fundamental challenges with regard to the user audience. In this contribution, we discuss MIR technology adoption issues that were experienced with professional music stakeholders in audio mixing, performance, musicology and sales industry. Many of these stakeholders have mindsets and priorities that differ considerably from those of most MIR academics, influencing their reception of new MIR technology. We mention the major observed differences and their backgrounds, and argue that these are essential to be taken into account to allow for truly successful cross-disciplinary collaboration and technology adoption in MIR.

Collaboration


Dive into the Cynthia C. S. Liem's collaboration.

Top Co-Authors

Avatar

Alan Hanjalic

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Markus Schedl

Johannes Kepler University of Linz

View shared research outputs
Top Co-Authors

Avatar

Mark S. Melenhorst

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Martha Larson

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Marko Tkalcic

Free University of Bozen-Bolzano

View shared research outputs
Researchain Logo
Decentralizing Knowledge