Sven Ewan Shepstone
Bang & Olufsen
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Sven Ewan Shepstone.
IEEE Transactions on Consumer Electronics | 2013
Sven Ewan Shepstone; Zheng-Hua Tan; Søren Holdt Jensen
Recommending TV content to groups of viewers is best carried out when relevant information such as the demographics of the group is available. However, it can be difficult and time consuming to extract information for every user in the group. This paper shows how an audio analysis of the age and gender of a group of users watching the TV can be used for recommending a sequence of N short TV content items for the group. First, a state of the art audio-based classifier determines the age and gender of each user in an M-user group and creates a group profile. A genetic recommender algorithm then selects for each user in the profile, a single personalized multimedia item for viewing. When the number of items to be presented is different to the number of viewers in the group, i.e. M = N, a novel adaptation algorithm is proposed that first converts the M-user group profile to an N-slot content profile, thus ensuring that items are proportionally allocated to users with respect to their demographic categorization. The proposed system is compared to an ideal system where the group demographics are provided explicitly. Results using real speaker utterances show that, in spite of the inaccuracies of state-of-the-art age-and-gender detection systems, the proposed system has a significant ability to predict an item with a matching age and gender category. User studies were conducted where subjects were asked to rate a sequence of advertisements, where half of the advertisements were randomly selected, and the other half were selected using the audio-derived demographics. The recommended advertisements received a significant higher median rating of 7.75, as opposed to 4.25 for the randomly selected advertisements.
International Journal of Social Robotics | 2018
Zheng-Hua Tan; Nicolai Bæk Thomsen; Xiaodong Duan; Evgenios Vlachos; Sven Ewan Shepstone; Morten Højfeldt Rasmussen; Jesper Lisby Højvang
We present one way of constructing a social robot, such that it is able to interact with humans using multiple modalities. The robotic system is able to direct attention towards the dominant speaker using sound source localization and face detection, it is capable of identifying persons using face recognition and speaker identification and the system is able to communicate and engage in a dialog with humans by using speech recognition, speech synthesis and different facial expressions. The software is built upon the open-source robot operating system framework and our software is made publicly available. Furthermore, the electrical parts (sensors, laptop, base platform, etc.) are standard components, thus allowing for replicating the system. The design of the robot is unique and we justify why this design is suitable for our robot and the intended use. By making software, hardware and design accessible to everyone, we make research in social robotics available to a broader audience. To evaluate the properties and the appearance of the robot we invited users to interact with it in pairs (active interaction partner/observer) and collected their responses via an extended version of the Godspeed Questionnaire. Results suggest an overall positive impression of the robot and interaction experience, as well as significant differences in responses based on type of interaction and gender.
IEEE Transactions on Affective Computing | 2018
Sven Ewan Shepstone; Zheng-Hua Tan; Søren Holdt Jensen
This paper introduces a novel framework for combining the strengths of machine-based and human-based emotion classification. Peoples’ ability to tell similar emotions apart is known as emotional granularity, which can be high or low, and is measurable. This paper proposes granularity-adapted classification that can be used as a front-end to drive a recommender, based on emotions from speech. In this context, incorrectly predicted peoples’ emotions could lead to poor recommendations, reducing user satisfaction. Instead of identifying a single emotion class, an adapted class is proposed, and is an aggregate of underlying emotion classes chosen based on granularity. In the recommendation context, the adapted class maps to a larger region in valence-arousal space, from which a list of potentially more similar content items is drawn, and recommended to the user. To determine the effectiveness of adapted classes, we measured the emotional granularity of subjects, and for each subject, used their pairwise similarity judgments of emotion to compare the effectiveness of adapted classes versus single emotion classes taken from a baseline system. A customized Euclidean-based similarity metric is used to measure the relative proximity of emotion classes. Results show that granularity-adapted classification can improve the potential similarity by up to 9.6 percent.
IEEE Transactions on Multimedia | 2014
Sven Ewan Shepstone; Zheng-Hua Tan; Søren Holdt Jensen
IEEE Transactions on Audio, Speech, and Language Processing | 2016
Sven Ewan Shepstone; Kong Aik Lee; Haizhou Li; Zheng-Hua Tan; Søren Holdt Jensen
international conference on acoustics, speech, and signal processing | 2015
Sven Ewan Shepstone; Kong Aik Lee; Haizhou Li; Zheng-Hua Tan; Søren Holdt Jensen
international conference on user modeling adaptation and personalization | 2018
Miklas Strøm Kristoffersen; Sven Ewan Shepstone; Zheng-Hua Tan
arXiv: Information Retrieval | 2018
Miklas Strøm Kristoffersen; Sven Ewan Shepstone; Zheng-Hua Tan
IEEE Transactions on Consumer Electronics | 2018
Sven Ewan Shepstone; Zheng-Hua Tan; Miklas Strøm Kristoffersen
Archive | 2015
Sven Ewan Shepstone