Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Varvara Logacheva is active.

Publication


Featured researches published by Varvara Logacheva.


workshop on statistical machine translation | 2015

Findings of the 2015 Workshop on Statistical Machine Translation

Ondrej Bojar; Rajen Chatterjee; Christian Federmann; Barry Haddow; Matthias Huck; Chris Hokamp; Philipp Koehn; Varvara Logacheva; Christof Monz; Matteo Negri; Matt Post; Carolina Scarton; Lucia Specia; Marco Turchi

This paper presents the results of the WMT15 shared tasks, which included a standard news translation task, a metrics task, a tuning task, a task for run-time estimation of machine translation quality, and an automatic post-editing task. This year, 68 machine translation systems from 24 institutions were submitted to the ten translation directions in the standard translation task. An additional 7 anonymized systems were included, and were then evaluated both automatically and manually. The quality estimation task had three subtasks, with a total of 10 teams, submitting 34 entries. The pilot automatic postediting task had a total of 4 teams, submitting 7 entries.


meeting of the association for computational linguistics | 2016

Findings of the 2016 Conference on Machine Translation.

Ondˇrej Bojar; Rajen Chatterjee; Christian Federmann; Yvette Graham; Barry Haddow; Matthias Huck; Antonio Jimeno Yepes; Philipp Koehn; Varvara Logacheva; Christof Monz; Matteo Negri; Aurélie Névéol; Mariana L. Neves; Martin Popel; Matt Post; Raphael Rubino; Carolina Scarton; Lucia Specia; Marco Turchi; Karin Verspoor; Marcos Zampieri

This paper presents the results of the WMT16 shared tasks, which included five machine translation (MT) tasks (standard news, IT-domain, biomedical, multimodal, pronoun), three evaluation tasks (metrics, tuning, run-time estimation of MT quality), and an automatic post-editing task and bilingual document alignment task. This year, 102 MT systems from 24 institutions (plus 36 anonymized online systems) were submitted to the 12 translation directions in the news translation task. The IT-domain task received 31 submissions from 12 institutions in 7 directions and the Biomedical task received 15 submissions systems from 5 institutions. Evaluation was both automatic and manual (relative ranking and 100-point scale assessments). The quality estimation task had three subtasks, with a total of 14 teams, submitting 39 entries. The automatic post-editing task had a total of 6 teams, submitting 11 entries.


workshop on statistical machine translation | 2015

SHEF-NN: Translation Quality Estimation with Neural Networks

Kashif Shah; Varvara Logacheva; Gustavo Paetzold; Frédéric Blain; Daniel Beck; Fethi Bougares; Lucia Specia

We describe our systems for Tasks 1 and 2 of the WMT15 Shared Task on Quality Estimation. Our submissions use (i) a continuous space language model to extract additional features for Task 1 (SHEFGP, SHEF-SVM), (ii) a continuous bagof-words model to produce word embeddings as features for Task 2 (SHEF-W2V) and (iii) a combination of features produced by QuEst++ and a feature produced with word embedding models (SHEFQuEst++). Our systems outperform the baseline as well as many other submissions. The results are especially encouraging for Task 2, where our best performing system (SHEF-W2V) only uses features learned in an unsupervised fashion.


workshop on statistical machine translation | 2015

Data enhancement and selection strategies for the word-level Quality Estimation

Varvara Logacheva; Chris Hokamp; Lucia Specia

This paper describes the DCU-SHEFF word-level Quality Estimation (QE) system submitted to the QE shared task at WMT15. Starting from a baseline set of features and a CRF algorithm to learn a sequence tagging model, we propose improvements in two ways: (i) by filtering out the training sentences containing too few errors, and (ii) by adding incomplete sequences to the training data to enrich the model with new information. We also experiment with considering the task as a classification problem, and report results using a subset of the features with Random Forest classifiers.


meeting of the association for computational linguistics | 2016

Metrics for Evaluation of Word-level Machine Translation Quality Estimation

Varvara Logacheva; Michal Lukasik; Lucia Specia

The aim of this paper is to investigate suitable evaluation strategies for the task of word-level quality estimation of machine translation. We suggest various metrics to replace F1-score for the “BAD” class, which is currently used as main metric. We compare the metrics’ performance on real system outputs and synthetically generated datasets and suggest a reliable alternative to the F1-BAD score — the multiplication of F1-scores for different classes. Other metrics have lower discriminative power and are biased by unfair labellings.


Archive | 2018

ConvAI Dataset of Topic-Oriented Human-to-Chatbot Dialogues

Varvara Logacheva; Mikhail Burtsev; Valentin Malykh; Vadim Polulyakh; Aleksandr Seliverstov

This paper contains the description and the analysis of the dataset collected during the Conversational Intelligence Challenge (ConvAI) which took place in 2017. During the evaluation round we collected over 4,000 dialogues from 10 chatbots and 1,000 volunteers. Here we provide the dataset statistics and outline some possible improvements for future data collection experiments.


Archive | 2018

The First Conversational Intelligence Challenge

Mikhail Burtsev; Varvara Logacheva; Valentin Malykh; Iulian Vlad Serban; Ryan Lowe; Shrimai Prabhumoye; Alan W. Black; Alexander I. Rudnicky; Yoshua Bengio

The first Conversational Intelligence Challenge was conducted over 2017 with finals at NIPS conference. The challenge IS aimed at evaluating the state of the art in non-goal-driven dialogue systems (chatbots) and collecting a large dataset of human-to-machine and human-to-human conversations manually labelled for quality. We established a task for formal human evaluation of chatbots that allows to test capabilities of chatbot in topic-oriented dialogue. Instead of traditional chit-chat, participating systems and humans were given a task to discuss a short text. Ten dialogue systems participated in the competition. The majority of them combined multiple conversational models such as question answering and chit-chat systems to make conversations more natural. The evaluation of chatbots was performed by human assessors. Almost 1,000 volunteers were attracted and over 4,000 dialogues were collected during the competition. Final score of the dialogue quality for the best bot was 2.7 compared to 3.8 for human. This demonstrates that current technology allows supporting dialogue on a given topic but with quality significantly lower than that of human. To close this gap we plan to continue the experiments by organising the next conversational intelligence competition. This future work will benefit from the data we collected and dialogue systems that we made available after the competition presented in the paper.


Proceedings of the First Conference on Machine Translation: Volume 2, Shared Task Papers | 2016

USFD's Phrase-level Quality Estimation Systems.

Varvara Logacheva; Frédéric Blain; Lucia Specia

We describe the submissions of the University of Sheffield (USFD) for the phraselevel Quality Estimation (QE) shared task of WMT16. We test two different approaches for phrase-level QE: (i) we enrich the provided set of baseline features with information about the context of the phrases, and (ii) we exploit predictions at other granularity levels (word and sentence). These approaches perform closely in terms of multiplication of F1-scores (primary evaluation metric), but are considerably different in terms of the F1scores for individual classes.


conference of the european chapter of the association for computational linguistics | 2014

Confidence-based Active Learning Methods for Machine Translation

Varvara Logacheva; Lucia Specia

The paper presents experiments with active learning methods for the acquisition of training data in the context of machine translation. We propose a confidencebased method which is superior to the state-of-the-art method both in terms of quality and complexity. Additionally, we discovered that oracle selection techniques that use real quality scores lead to poor results, making the effectiveness of confidence-driven methods of active learning for machine translation questionable.


Proceedings of the Second Conference on Machine Translation | 2017

Findings of the 2017 Conference on Machine Translation (WMT17)

Ondřej Bojar; Rajen Chatterjee; Christian Federmann; Yvette Graham; Barry Haddow; Shujian Huang; Matthias Huck; Philipp Koehn; Qun Liu; Varvara Logacheva; Christof Monz; Matteo Negri; Matt Post; Raphael Rubino; Lucia Specia; Marco Turchi

Collaboration


Dive into the Varvara Logacheva's collaboration.

Top Co-Authors

Avatar

Lucia Specia

University of Sheffield

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Mikhail Burtsev

Moscow Institute of Physics and Technology

View shared research outputs
Top Co-Authors

Avatar

Valentin Malykh

Moscow Institute of Physics and Technology

View shared research outputs
Top Co-Authors

Avatar

Rajen Chatterjee

Indian Institute of Technology Bombay

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Barry Haddow

University of Edinburgh

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge