Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Evangelos Sariyanidi is active.

Publication


Featured researches published by Evangelos Sariyanidi.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2015

Automatic Analysis of Facial Affect: A Survey of Registration, Representation, and Recognition

Evangelos Sariyanidi; Hatice Gunes; Andrea Cavallaro

Automatic affect analysis has attracted great interest in various contexts including the recognition of action units and basic or non-basic emotions. In spite of major efforts, there are several open questions on what the important cues to interpret facial expressions are and how to encode them. In this paper, we review the progress across a range of affect recognition applications to shed light on these fundamental questions. We analyse the state-of-the-art solutions by decomposing their pipelines into fundamental components, namely face registration, representation, dimensionality reduction and recognition. We discuss the role of these components and highlight the models and new trends that are followed in their design. Moreover, we provide a comprehensive analysis of facial representations by uncovering their advantages and limitations; we elaborate on the type of information they encode and discuss how they deal with the key challenges of illumination variations, registration errors, head-pose variations, occlusions, and identity bias. This survey allows us to identify open issues and to define future directions for designing real-world affect recognition systems.


british machine vision conference | 2013

Local Zernike Moment Representation for Facial Affect Recognition.

Evangelos Sariyanidi; Hatice Gunes; Muhittin Gökmen; Andrea Cavallaro

Local representations became popular for facial affect recognition as they efficiently capture the image discontinuities, which play an important role for interpreting facial actions. We propose to use Local Zernike Moments (ZMs) [4] due to their useful and compact description of the image discontinuities and texture. Their main advantage in comparison to well-established alternatives such as Local Binary Patterns (LBPs) [5], is their flexibility in terms of the size and level of detail of the local description. We introduce a local ZM-based representation which involves a non-linear encoding layer (quantisation). The functionality of this layer is mapping similar facial configurations together and increasing compactness. We demonstrate the use of the local ZM-based representation for posed and naturalistic affect recognition on standard datasets, and show its superiority to alternative approaches for both tasks. Contemporary representations are often designed as frameworks consisting of three layers [2]: (Local) feature extraction, non-linear encoding and pooling. Non-linear encoding aims at enhancing the relevance of local features by increasing their robustness against image noise. Pooling describes small spatial neighbourhoods as single entities, ignoring the precise location of the encoded features, and increasing the tolerance against small geometric inconsistencies. In what follows, we describe the proposed local ZM-based representation scheme in terms of this threelayered framework. Feature Extraction – Local Zernike Moments: The computation of (complex) ZMs can be considered equivalent to representing an image in an alternative space. As shown in Figure 1-a, an image is decomposed onto a set of basis matrices (ZM bases), which are useful for describing the variation at different directions and scales. ZM bases are orthogonal, therefore there is no overlap in the information conveyed by each feature (ZM coefficient). ZMs are usually computed for the entire image, however in this case, ZMs cannot capture the local variation due to ZM bases lacking localisation [3]. In contrary, when computed around local neighbourhoods across the image, they become an efficient tool for describing the image discontinuities which are essential to interpreting facial activity. Non-linear Encoding – Quantisation: We perform quantisation via converting local features into binary values. Such coarse quantisation increases compactness and allows us to code each local block only with a single integer. Figure 1-b illustrates the process of obtaining the Quantised Local ZM (QLZM) image. Firstly, local ZM coefficients are computed across the input image (LZM layer) — each image in the LZM layer (LZM image) contains the features that are extracted through a particular ZM basis. Next, each LZM image is converted into a binary image by quantising each pixel via the signum(·) function. Finally, the QLZM image is obtained by combining all of the binary images. Specifically, each pixel in a particular location of the QLZM image is an integer (QLZM integer), computed by concatenating all of the binary values in the corresponding location of all binary images. The QLZM image is similar to an LBP-transformed image, in the sense that it contains integers of a limited range. Yet, the physical meaning of the information encoded by each integer is quite different. LBP integers describe a circular block by considering only the values along the border, neglecting the pixels that remain inside the block. Therefore, the efficient operation scale of LBPs is usually limited to 3-5 pixels [1, 5]. QLZM integers, on the other hand, describe blocks as a whole, and provide flexibility in terms of operation scale without major loss of information. Pooling – Histograms: Our representation scheme pools encoded features over local histograms. Figure 1-c illustrates the overall pipeline of the proposed representation scheme. Firstly, the QLZM image is computed through the process that is illustrated in detail in Figure 1-b. Next, . . . . . . ... = ZM Coefficients (local features) ZM Bases


ieee international conference on automatic face gesture recognition | 2015

Let me tell you about your personality! † : Real-time personality prediction from nonverbal behavioural cues

Oya Celiktutan; Evangelos Sariyanidi; Hatice Gunes

Although automatic personality analysis has been studied extensively in recent years, it has not yet been adopted for real time applications and real life practices. To the best of our knowledge, this demonstration is a first attempt at predicting the widely used Big Five personality dimensions and a number of social dimensions from nonverbal behavioural cues in real-time. The proposed system aims to analyse the nonverbal behaviour of the person that interacts with a small humanoid robot through a live streaming camera, and delivers the predicted personality and social dimensions on the fly.


international conference on multimodal interfaces | 2014

MAPTRAITS 2014 - The First Audio/Visual Mapping Personality Traits Challenge - An Introduction: Perceived Personality and Social Dimensions

Oya Celiktutan; Florian Eyben; Evangelos Sariyanidi; Hatice Gunes; Björn W. Schuller

The Audio/Visual Mapping Personality Challenge and Workshop (MAPTRAITS) is a competition event that is organised to facilitate the development of signal processing and machine learning techniques for the automatic analysis of personality traits and social dimensions. MAPTRAITS includes two sub-challenges, the continuous space-time sub-challenge and the quantised space-time sub-challenge. The continuous sub-challenge evaluated how systems predict the variation of perceived personality traits and social dimensions in time, whereas the quantised challenge evaluated the ability of systems to predict the overall perceived traits and dimensions in shorter video clips. To analyse the effect of audio and visual modalities on personality perception, we compared systems under three different settings: visual-only, audio-only and audio-visual. With MAPTRAITS we aimed at improving the knowledge on the automatic analysis of personality traits and social dimensions by producing a benchmarking protocol and encouraging the participation of various research groups from different backgrounds.


Proceedings of the 2014 Workshop on Mapping Personality Traits Challenge and Workshop | 2014

MAPTRAITS 2014: The First Audio/Visual Mapping Personality Traits Challenge

Oya Celiktutan; Florian Eyben; Evangelos Sariyanidi; Hatice Gunes; Björn W. Schuller

The Audio/Visual Mapping Personality Challenge and Workshop (MAPTRAITS) is a competition event aimed at the comparison of signal processing and machine learning methods for automatic visual, vocal and/or audio-visual analysis of personality traits and social dimensions, namely, extroversion, agreeableness, conscientiousness, neuroticism, openness, engagement, facial attractiveness, vocal attractiveness, and likability. The MAPTRAITS Challenge aims to bring forth existing efforts and major accomplishments in modelling and analysis of personality and social dimensions in both quantised and continuous time and space. This paper provides the details of the two Sub-Challenges, their conditions, datasets, labels and baseline features that made available to the researchers who are interested in taking part in this challenge and workshop.


IEEE Transactions on Image Processing | 2017

Learning Bases of Activity for Facial Expression Recognition

Evangelos Sariyanidi; Hatice Gunes; Andrea Cavallaro

The extraction of descriptive features from the sequences of faces is a fundamental problem in facial expression analysis. Facial expressions are represented by psychologists as a combination of elementary movements known as action units: each movement is localised and its intensity is specified with a score that is small when the movement is subtle and large when the movement is pronounced. Inspired by this approach, we propose a novel data-driven feature extraction framework that represents facial expression variations as a linear combination of localised basis functions, whose coefficients are proportional to movement intensity. We show that the linear basis functions of the proposed framework can be obtained by training a sparse linear model with Gabor phase shifts computed from facial videos. The proposed framework addresses generalisation issues that are not tackled by existing learnt representations, and achieves, with the same learning parameters, state-of-the-art results in recognising both posed expressions and spontaneous micro-expressions. This performance is confirmed even when the data used to train the model differ from test data in terms of the intensity of facial movements and frame rate.


asian conference on computer vision | 2014

Probabilistic Subpixel Temporal Registration for Facial Expression Analysis

Evangelos Sariyanidi; Hatice Gunes; Andrea Cavallaro

Face images in a video sequence should be registered accurately before any analysis, otherwise registration errors may be interpreted as facial activity. Subpixel accuracy is crucial for the analysis of subtle actions. In this paper we present PSTR (Probabilistic Subpixel Temporal Registration), a framework that achieves high registration accuracy. Inspired by the human vision system, we develop a motion representation that measures registration errors among subsequent frames, a probabilistic model that learns the registration errors from the proposed motion representation, and an iterative registration scheme that identifies registration failures thus making PSTR aware of its errors. We evaluate PSTR’s temporal registration accuracy on facial action and expression datasets, and demonstrate its ability to generalise to naturalistic data even when trained with controlled data.


IEEE Transactions on Image Processing | 2017

Robust Registration of Dynamic Facial Sequences

Evangelos Sariyanidi; Hatice Gunes; Andrea Cavallaro

Accurate face registration is a key step for several image analysis applications. However, existing registration methods are prone to temporal drift errors or jitter among consecutive frames. In this paper, we propose an iterative rigid registration framework that estimates the misalignment with trained regressors. The input of the regressors is a robust motion representation that encodes the motion between a misaligned frame and the reference frame(s), and enables reliable performance under non-uniform illumination variations. Drift errors are reduced when the motion representation is computed from multiple reference frames. Furthermore, we use the


Proceedings of SPIE | 2013

Loop closure detection using local Zernike moment patterns

Evangelos Sariyanidi; Onur Sencan; Hakan Temeltas

L_{2}


IEEE Transactions on Image Processing | 2017

Biologically Inspired Motion Encoding for Robust Global Motion Estimation

Evangelos Sariyanidi; Hatice Gunes; Andrea Cavallaro

norm of the representation as a cue for performing coarse-to-fine registration efficiently. Importantly, the framework can identify registration failures and correct them. Experiments show that the proposed approach achieves significantly higher registration accuracy than the state-of-the-art techniques in challenging sequences.

Collaboration


Dive into the Evangelos Sariyanidi's collaboration.

Top Co-Authors

Avatar

Hatice Gunes

University of Cambridge

View shared research outputs
Top Co-Authors

Avatar

Andrea Cavallaro

Queen Mary University of London

View shared research outputs
Top Co-Authors

Avatar

Oya Celiktutan

Queen Mary University of London

View shared research outputs
Top Co-Authors

Avatar

Hakan Temeltas

Istanbul Technical University

View shared research outputs
Top Co-Authors

Avatar

Onur Sencan

Istanbul Technical University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Can Erhan

Istanbul Technical University

View shared research outputs
Top Co-Authors

Avatar

Muhittin Gökmen

Istanbul Technical University

View shared research outputs
Top Co-Authors

Avatar

Jan Ondras

University of Cambridge

View shared research outputs
Top Co-Authors

Avatar

Salih Cihan Tek

Istanbul Technical University

View shared research outputs
Researchain Logo
Decentralizing Knowledge