Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Ali Orkan Bayer is active.

Publication


Featured researches published by Ali Orkan Bayer.


conference on computational natural language learning | 2015

The UniTN Discourse Parser in CoNLL 2015 Shared Task: Token-level Sequence Labeling with Argument-specific Models

Evgeny A. Stepanov; Giuseppe Riccardi; Ali Orkan Bayer

Penn Discourse Treebank style discourse parsing is a composite task of identifying discourse relations (explicit or nonexplicit), their connective and argument spans, and assigning a sense to these relations from the hierarchy of senses. In this paper we describe University of Trento parser submitted to CoNLL 2015 Shared Task on Shallow Discourse Parsing. The span detection tasks for explicit relations are cast as token-level sequence labeling. The argument span decisions are conditioned on relations’ being intra- or intersentential. Non-explicit relation detection and sense assignment tasks are cast as classification. In the end-to-end closedtrack evaluation, the parser ranked second with a global F-measure of 0.2184


spoken language technology workshop | 2012

Joint language models for automatic speech recognition and understanding

Ali Orkan Bayer; Giuseppe Riccardi

Language models (LMs) are one of the main knowledge sources used by automatic speech recognition (ASR) and Spoken Language Understanding (SLU) systems. In ASR systems they are optimized to decode words from speech for a transcription task. In SLU systems they are optimized to map words into concept constructs or interpretation representations. Performance optimization is generally designed independently for ASR and SLU models in terms of word accuracy and concept accuracy respectively. However, the best word accuracy performance does not always yield the best understanding performance. In this paper we investigate how LMs originally trained to maximize word accuracy can be parametrized to account for speech understanding constraints and maximize concept accuracy. Incremental reduction in concept error rate is observed when a LM is trained on word-to-concept mappings. We show how to optimize the joint transcription and understanding task performance in the lexical-semantic relation space.


ieee automatic speech recognition and understanding workshop | 2013

Language style and domain adaptation for cross-language SLU porting

Evgeny A. Stepanov; Ilya Kashkarev; Ali Orkan Bayer; Giuseppe Riccardi; Arindam Ghosh

Automatic cross-language Spoken Language Understanding porting is plagued by two limitations. First, SLU are usually trained on limited domain corpora. Second, language pair resources (e.g. aligned corpora) are scarce or unmatched in style (e.g. news vs. conversation). We present experiments on automatic style adaptation of the input for the translation systems and their output for SLU. We approach the problem of scarce aligned data by adapting the available parallel data to the target domain using limited in-domain and larger web crawled close-to-domain corpora. SLU performance is optimized by reranking its output with Recurrent Neural Network-based joint language model. We evaluate end-to-end SLU porting on close and distant language pairs: Spanish - Italian and Turkish - Italian; and achieve significant improvements both in translation quality and SLU performance.


Computer Speech & Language | 2016

Semantic Language models with deep neural Networks

Ali Orkan Bayer; Giuseppe Riccardi

Abstract In this paper we explore the use of semantics in training language models for automatic speech recognition and spoken language understanding. Traditional language models (LMs) do not consider the semantic constraints and train models based on fixed-sized word histories. The theory of frame semantics analyzes word meanings and their constructs by using “semantic frames”. Semantic frames represent a linguistic scene with its relevant participants and their relations. They are triggered by target words and include slots which are filled by frame elements. We present semantic LMs (SELMs), which use recurrent neural network architectures and the linguistic scene of frame semantics as context. SELMs incorporate semantic features which are extracted from semantic frames and target words. In this way, long-range and “latent” dependencies, i.e. the implicit semantic dependencies between words, are incorporated into LMs. This is crucial especially when the main aim of spoken language systems is understanding what the user means. Semantic features consist of low-level features, where frame and target information is directly used; and deep semantic encodings, where deep autoencoders are used to extract semantic features. We evaluate the performance of SELMs on publicly available corpora: the Wall Street Journal read-speech corpus and the LUNA human–human conversational corpus. The encoding of semantic frames into SELMs improves the word recognition performance and especially the recognition performance of the target words, the meaning bearing elements of semantic frames. We assess the performance of SELMs for the understanding tasks and we show that SELMs yield better semantic frame identification performance compared to recurrent neural network LMs.


ieee automatic speech recognition and understanding workshop | 2013

On-line adaptation of semantic models for spoken language understanding

Ali Orkan Bayer; Giuseppe Riccardi

Spoken language understanding (SLU) systems extract semantic information from speech signals, which is usually mapped onto concept sequences. The distribution of concepts in dialogues are usually sparse. Therefore, general models may fail to model the concept distribution for a dialogue and semantic models can benefit from adaptation. In this paper, we present an instance-based approach for on-line adaptation of semantic models. We show that we can improve the performance of an SLU system on an utterance, by retrieving relevant instances from the training data and using them for on-line adapting the semantic models. The instance-based adaptation scheme uses two different similarity metrics edit distance and n-gram match score on three different to-kenizations; word-concept pairs, words, and concepts. We have achieved a significant improvement (6% relative) in the understanding performance by conducting rescoring experiments on the n-best lists that the SLU outputs. We have also applied a two-level adaptation scheme, where adaptation is first applied to the automatic speech recognizer (ASR) and then to the SLU.


language resources and evaluation | 2018

Cross-language transfer of semantic annotation via targeted crowdsourcing: task design and evaluation

Evgeny A. Stepanov; Shammur Absar Chowdhury; Ali Orkan Bayer; Arindam Ghosh; Ioannis Klasinas; Marcos Calvo; Emilio Sanchis; Giuseppe Riccardi

AbstractModern data-driven spoken language systems (SLS) require manual semantic annotation for training spoken language understanding parsers. Multilingual porting of SLS demands significant manual effort and language resources, as this manual annotation has to be replicated. Crowdsourcing is an accessible and cost-effective alternative to traditional methods of collecting and annotating data. The application of crowdsourcing to simple tasks has been well investigated. However, complex tasks, like cross-language semantic annotation transfer, may generate low judgment agreement and/or poor performance. The most serious issue in cross-language porting is the absence of reference annotations in the target language; thus, crowd quality control and the evaluation of the collected annotations is difficult. In this paper we investigate targeted crowdsourcing for semantic annotation transfer that delegates to crowds a complex task such as segmenting and labeling of concepts taken from a domain ontology; and evaluation using source language annotation. To test the applicability and effectiveness of the crowdsourced annotation transfer we have considered the case of close and distant language pairs: Italian–Spanish and Italian–Greek. The corpora annotated via crowdsourcing are evaluated against source and target language expert annotations. We demonstrate that the two evaluation references (source and target) highly correlate with each other; thus, drastically reduce the need for the target language reference annotations.


language resources and evaluation | 2014

The Development of the Multilingual LUNA Corpus for Spoken Language System Porting

Evgeny A. Stepanov; Giuseppe Riccardi; Ali Orkan Bayer


spoken language technology workshop | 2014

Semantic language models for Automatic Speech Recognition

Ali Orkan Bayer; Giuseppe Riccardi


conference of the international speech communication association | 2014

Cross-language transfer of semantic annotation via targeted crowdsourcing.

Shammur Absar Chowdhury; Arindam Ghosh; Evgeny A. Stepanov; Ali Orkan Bayer; Giuseppe Riccardi; Ioannis Klasinas


conference of the international speech communication association | 2013

Motivational feedback in crowdsourcing: a case study in speech transcription.

Giuseppe Riccardi; Arindam Ghosh; Shammur Absar Chowdhury; Ali Orkan Bayer

Collaboration


Dive into the Ali Orkan Bayer's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Marcos Calvo

Polytechnic University of Valencia

View shared research outputs
Top Co-Authors

Avatar

Ioannis Klasinas

Technical University of Crete

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Emilio Sanchis

Polytechnic University of Valencia

View shared research outputs
Top Co-Authors

Avatar

Fernando García

Polytechnic University of Valencia

View shared research outputs
Top Co-Authors

Avatar

Giuseppe Carenini

University of British Columbia

View shared research outputs
Researchain Logo
Decentralizing Knowledge