Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Raj Dabre is active.

Publication


Featured researches published by Raj Dabre.


north american chapter of the association for computational linguistics | 2015

Leveraging Small Multilingual Corpora for SMT Using Many Pivot Languages

Raj Dabre; Fabien Cromieres; Sadao Kurohashi; Pushpak Bhattacharyya

We present our work on leveraging multilingual parallel corpora of small sizes for Statistical Machine Translation between Japanese and Hindi using multiple pivot languages. In our setting, the source and target part of the corpus remains the same, but we show that using several different pivot to extract phrase pairs from these source and target parts lead to large BLEU improvements. We focus on a variety of ways to exploit phrase tables generated using multiple pivots to support a direct source-target phrase table. Our main method uses the Multiple Decoding Paths (MDP) feature of Moses, which we empirically verify as the best compared to the other methods we used. We compare and contrast our various results to show that one can overcome the limitations of small corpora by using as many pivot languages as possible in a multilingual setting. Most importantly, we show that such pivoting aids in learning of additional phrase pairs which are not learned when the direct sourcetarget corpus is small. We obtained improvements of up to 3 BLEU points using multiple pivots for Japanese to Hindi translation compared to when only one pivot is used. To the best of our knowledge, this work is also the first of its kind to attempt the simultaneous utilization of 7 pivot languages at decoding time.


meeting of the association for computational linguistics | 2017

An Empirical Comparison of Domain Adaptation Methods for Neural Machine Translation

Chenhui Chu; Raj Dabre; Sadao Kurohashi

In this paper, we propose a novel domain adaptation method named “mixed fine tuning” for neural machine translation (NMT). We combine two existing approaches namely fine tuning and multi domain NMT. We first train an NMT model on an out-of-domain parallel corpus, and then fine tune it on a parallel corpus which is a mix of the in-domain and out-of-domain corpora. All corpora are augmented with artificial tags to indicate specific domains. We empirically compare our proposed method against fine tuning and multi domain methods and discuss its benefits and shortcomings.


Journal of Information Processing | 2018

A Comprehensive Empirical Comparison of Domain Adaptation Methods for Neural Machine Translation

Chenhui Chu; Raj Dabre; Sadao Kurohashi

Neural machine translation (NMT) has shown very promising results when there are large amounts of parallel corpora. However, for low resource domains, vanilla NMT cannot give satisfactory performance due to overfitting on the small size of parallel corpora. Two categories of domain adaptation approaches have been proposed for low resource NMT, i.e., adaptation using out-of-domain parallel corpora and in-domain monolingual corpora. In this paper, we conduct a comprehensive empirical comparison of the methods in both categories. For domain adaptation using out-of-domain parallel corpora, we further propose a novel domain adaptation method named mixed fine tuning, which combines two existing methods namely fine tuning and multi domain NMT. For domain adaptation using in-domain monolingual corpora, we compare two existing methods namely language model fusion and synthetic data generation. In addition, we propose a method that combines these two categories. We empirically compare all the methods and discuss their benefits and shortcomings. To the best of our knowledge, this is the first work on a comprehensive empirical comparison of domain adaptation methods for NMT.


Proceedings of the First Conference on Machine Translation: Volume 2, Shared Task Papers | 2016

The Kyoto University Cross-Lingual Pronoun Translation System.

Raj Dabre; Yevgeniy Puzikov; Fabien Cromieres; Sadao Kurohashi

In this paper we describe our system we designed and implemented for the crosslingual pronoun prediction task as a part of WMT 2016. The majority of the paper will be dedicated to the system whose outputs we submitted wherein we describe the simplified mathematical model, the details of the components and the working by means of an architecture diagram which also serves as a flowchart. We then discuss the results of the official scores and our observations on the same.


Journal of Information Processing | 2018

Exploiting Multilingual Corpora Simply and Efficiently in Neural Machine Translation

Raj Dabre; Fabien Cromieres; Sadao Kurohashi

In this paper, we explore a simple approach for “Multi-Source Neural Machine Translation” (MSNMT) which only relies on preprocessing a N-way multilingual corpus without modifying the Neural Machine Translation (NMT) architecture or training procedure. We simply concatenate the source sentences to form a single, long multisource input sentence while keeping the target side sentence as it is and train an NMT system using this preprocessed corpus. We evaluate our method in resource poor as well as resource rich settings and show its effectiveness (up to 4 BLEU using 2 source languages and up to 6 BLEU using 5 source languages) and compare them against existing approaches. We also provide some insights on how the NMT system leverages multilingual information in such a scenario by visualizing attention. We then show that this multi-source approach can be used for transfer learning to improve the translation quality for single-source systems without using any additional corpora thereby highlighting the importance of multilingual-multiway corpora in low resource scenarios. We also extract and evaluate a multilingual dictionary by a method that utilizes the multi-source attention and show that it works fairly well despite its simplicity.


international conference on computational linguistics | 2016

Kyoto University Participation to WAT 2016.

Fabien Cromieres; Raj Dabre; Toshiaki Nakazawa; Sadao Kurohashi


WAT | 2014

KyotoEBMT System Description for the 1st Workshop on Asian Translation.

John Richardson; Raj Dabre; Chenhui Chu; Fabien Cromieres; Toshiaki Nakazawa; Sadao Kurohashi


language resources and evaluation | 2016

Parallel Sentence Extraction from Comparable Corpora with Neural Network Features.

Chenhui Chu; Raj Dabre; Sadao Kurohashi


international conference on networks | 2015

Augmenting Pivot based SMT with word segmentation.

Rohit More; Anoop Kunchukuttan; Pushpak Bhattacharyya; Raj Dabre


international conference on computational linguistics | 2012

Morphological Analyzer for Affix Stacking Languages: A Case Study of Marathi

Raj Dabre; Archana Amberkar; Pushpak Bhattacharyya

Collaboration


Dive into the Raj Dabre's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Pushpak Bhattacharyya

Indian Institute of Technology Bombay

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Diptesh Kanojia

Indian Institute of Technology Bombay

View shared research outputs
Top Co-Authors

Avatar

Manish Shrivastava

Indian Institute of Technology Bombay

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Anoop Kunchukuttan

Indian Institute of Technology Bombay

View shared research outputs
Researchain Logo
Decentralizing Knowledge