Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Jiajun Zhang is active.

Publication


Featured researches published by Jiajun Zhang.


meeting of the association for computational linguistics | 2014

Bilingually-constrained Phrase Embeddings for Machine Translation

Jiajun Zhang; Shujie Liu; Mu Li; Ming Zhou; Chengqing Zong

We propose Bilingually-constrained Recursive Auto-encoders (BRAE) to learn semantic phrase embeddings (compact vector representations for phrases), which can distinguish the phrases with different semantic meanings. The BRAE is trained in a way that minimizes the semantic distance of translation equivalents and maximizes the semantic distance of nontranslation pairs simultaneously. After training, the model learns how to embed each phrase semantically in two languages and also learns how to transform semantic embedding space in one language to the other. We evaluate our proposed method on two end-to-end SMT tasks (phrase table pruning and decoding with phrasal semantic similarities) which need to measure semantic similarity between a source phrase and its translation candidates. Extensive experiments show that the BRAE is remarkably effective in these two tasks.


empirical methods in natural language processing | 2016

Exploiting Source-side Monolingual Data in Neural Machine Translation.

Jiajun Zhang; Chengqing Zong

Neural Machine Translation (NMT) based on the encoder-decoder architecture has recently become a new paradigm. Researchers have proven that the target-side monolingual data can greatly enhance the decoder model of NMT. However, the source-side monolingual data is not fully explored although it should be useful to strengthen the encoder model of NMT, especially when the parallel corpus is far from sufficient. In this paper, we propose two approaches to make full use of the sourceside monolingual data in NMT. The first approach employs the self-learning algorithm to generate the synthetic large-scale parallel data for NMT training. The second approach applies the multi-task learning framework using two NMTs to predict the translation and the reordered source-side monolingual sentences simultaneously. The extensive experiments demonstrate that the proposed methods obtain significant improvements over the strong attention-based NMT.


IEEE Intelligent Systems | 2015

Deep Neural Networks in Machine Translation: An Overview

Jiajun Zhang; Chengqing Zong

Deep neural networks (DNNs) are widely used in machine translation (MT). This article gives an overview of DNN applications in various aspects of MT.


Archive | 2012

Handling Unknown Words in Statistical Machine Translation from a New Perspective

Jiajun Zhang; Feifei Zhai; Chengqing Zong

Unknown words are one of the key factors which drastically impact the translation quality. Traditionally, nearly all the related research work focus on obtaining the translation of the unknown words in different ways. In this paper, we propose a new perspective to handle unknown words in statistical machine translation. Instead of trying great effort to find the translation of unknown words, this paper focuses on determining the semantic function the unknown words serve as in the test sentence and keeping the semantic function unchanged in the translation process. In this way, unknown words will help the phrase reordering and lexical selection of their surrounding words even though they still remain untranslated. In order to determine the semantic function of each unknown word, this paper employs the distributional semantic model and the bidirectional language model. Extensive experiments on Chinese-to-English translation show that our methods can substantially improve the translation quality.


meeting of the association for computational linguistics | 2017

Neural System Combination for Machine Translation.

Long Zhou; Wenpeng Hu; Jiajun Zhang; Chengqing Zong

Neural machine translation (NMT) becomes a new approach to machine translation and generates much more fluent results compared to statistical machine translation (SMT). However, SMT is usually better than NMT in translation adequacy. It is therefore a promising direction to combine the advantages of both NMT and SMT. In this paper, we propose a neural system combination framework leveraging multi-source NMT, which takes as input the outputs of NMT and SMT systems and produces the final translation. Extensive experiments on the Chinese-to-English translation task show that our model archives significant improvement by 5.3 BLEU points over the best single system output and 3.4 BLEU points over the state-of-the-art traditional system combination methods.


international conference on computational linguistics | 2008

Sentence Type Based Reordering Model for Statistical Machine Translation

Jiajun Zhang; Chengqing Zong; Shoushan Li

Many reordering approaches have been proposed for the statistical machine translation (SMT) system. However, the information about the type of source sentence is ignored in the previous works. In this paper, we propose a group of novel reordering models based on the source sentence type for Chinese-to-English translation. In our approach, an SVM-based classifier is employed to classify the given Chinese sentences into three types: special interrogative sentences, other interrogative sentences, and non-question sentences. The different reordering models are developed oriented to the different sentence types. Our experiments show that the novel reordering models have obtained an improvement of more than 2.65% in BLEU for a phrase-based spoken language translation system.


international joint conference on artificial intelligence | 2017

Learning Sentence Representation with Guidance of Human Attention.

Shaonan Wang; Jiajun Zhang; Chengqing Zong

Recently, much progress has been made in learning general-purpose sentence representations that can be used across domains. However, most of the existing models typically treat each word in a sentence equally. In contrast, extensive studies have proven that human read sentences efficiently by making a sequence of fixation and saccades. This motivates us to improve sentence representations by assigning different weights to the vectors of the component words, which can be treated as an attention mechanism on single sentences. To that end, we propose two novel attention models, in which the attention weights are derived using significant predictors of human reading time, i.e., Surprisal, POS tags and CCG supertags. The extensive experiments demonstrate that the proposed methods significantly improve upon the state-of-the-art sentence representation models.


NLPCC/ICCPOL | 2016

Character-Based LSTM-CRF with Radical-Level Features for Chinese Named Entity Recognition

Chuanhai Dong; Jiajun Zhang; Chengqing Zong; Masanori Hattori; Hui Di

State-of-the-art systems of Chinese Named Entity Recognition (CNER) require large amounts of hand-crafted features and domain-specific knowledge to achieve high performance. In this paper, we apply a bidirectional LSTM-CRF neural network that utilizes both character-level and radical-level representations. We are the first to use character-based BLSTM-CRF neural architecture for CNER. By contrasting the results of different variants of LSTM blocks, we find the most suitable LSTM block for CNER. We are also the first to investigate Chinese radical-level representations in BLSTM-CRF architecture and get better performance without carefully designed features. We evaluate our system on the third SIGHAN Bakeoff MSRA data set for simplfied CNER task and achieve state-of-the-art performance 90.95% F1.


IEEE Transactions on Audio, Speech, and Language Processing | 2016

Abstractive Cross-Language Summarization via Translation Model Enhanced Predicate Argument Structure Fusing

Jiajun Zhang; Yu Zhou; Chengqing Zong

Cross-language multidocument summarization is the task to generate a summary in a target language (e.g., Chinese) from a collection of documents in a different source language (e.g., English). Previous methods such as the extractive and compressive algorithms focus only on single sentence selection and compression, which cannot make full use of the similar sentences containing complementary information. Furthermore, the translation model knowledge is not fully explored in previous approaches. To address these two problems, we propose in this paper an abstractive cross-language summarization framework. First, the source language documents are translated into target language with a machine translation system. Then, the method constructs a pool of bilingual concepts and facts represented by the bilingual elements of the source-side predicate-argument structures (PAS) and their target-side counterparts. Finally, new summary sentences are produced by fusing bilingual PAS elements with the integer linear programming algorithm to maximize both of the salience and translation quality of the PAS elements. The experimental results on English-to-Chinese cross-language summarization demonstrate that our proposed method outperforms the state-of-the-art extractive systems in both automatic and manual evaluations.


Proceedings of the CoNLL-16 shared task | 2016

An End-to-End Chinese Discourse Parser with Adaptation to Explicit and Non-explicit Relation Recognition.

Xiaomian Kang; Haoran Li; Long Zhou; Jiajun Zhang; Chengqing Zong

This paper describes our end-to-end discourse parser in the CoNLL-2016 Shared Task on Chinese Shallow Discourse Parsing. To adapt to the characteristics of Chinese, we implement a uniform framework for both explicit and non-explicit relation parsing. In this framework, we are the first to utilize a seed-expansion approach for the argument extraction subtask. In the official evaluation, our system achieves an F1 score of 26.90% in overall performance on the blind test set.

Collaboration


Dive into the Jiajun Zhang's collaboration.

Top Co-Authors

Avatar

Chengqing Zong

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Yu Zhou

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Haoran Li

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Feifei Zhai

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Junnan Zhu

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Long Zhou

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Shaonan Wang

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Yang Zhao

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Yining Wang

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Guoping Huang

Chinese Academy of Sciences

View shared research outputs
Researchain Logo
Decentralizing Knowledge