Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Shuming Ma is active.

Publication


Featured researches published by Shuming Ma.


meeting of the association for computational linguistics | 2017

Improving semantic relevance for sequence-to-Sequence learning of Chinese social media text summarization

Shuming Ma; Xu Sun; Jingjing Xu; Houfeng Wang; Wenjie Li; Qi Su

Current Chinese social media text summarization models are based on an encoder-decoder framework. Although its generated summaries are similar to source texts literally, they have low semantic relevance. In this work, our goal is to improve semantic relevance between source texts and summaries for Chinese social media summarization. We introduce a Semantic Relevance Based neural model to encourage high semantic similarity between texts and summaries. In our model, the source text is represented by a gated attention encoder, while the summary representation is produced by a decoder. Besides, the similarity score between the representations is maximized during training. Our experiments show that the proposed model outperforms baseline systems on a social media corpus.


arXiv: Computation and Language | 2017

Transfer Deep Learning for Low-Resource Chinese Word Segmentation with a Novel Neural Network

Jingjing Xu; Shuming Ma; Yi Zhang; Bingzhen Wei; Xiaoyan Cai; Xu Sun

Recent studies have shown effectiveness in using neural networks for Chinese word segmentation. However, these models rely on large-scale data and are less effective for low-resource datasets because of insufficient training data. We propose a transfer learning method to improve low-resource word segmentation by leveraging high-resource corpora. First, we train a teacher model on high-resource corpora and then use the learned knowledge to initialize a student model. Second, a weighted data similarity method is proposed to train the student model on low-resource data. Experiment results show that our work significantly improves the performance on low-resource datasets: 2.3% and 1.5% F-score on PKU and CTB datasets. Furthermore, this paper achieves state-of-the-art results: 96.1%, and 96.2% F-score on PKU and CTB datasets.


north american chapter of the association for computational linguistics | 2018

Query and Output: Generating Words by Querying Distributed Word Representations for Paraphrase Generation

Shuming Ma; Xu Sun; Wei Li; Sujian Li; Wenjie Li; Xuancheng Ren

Most recent approaches use the sequence-to-sequence model for paraphrase generation. The existing sequence-to-sequence model tends to memorize the words and the patterns in the training dataset instead of learning the meaning of the words. Therefore, the generated sentences are often grammatically correct but semantically improper. In this work, we introduce a novel model based on the encoder-decoder framework, called Word Embedding Attention Network (WEAN). Our proposed model generates the words by querying distributed word representations (i.e. neural word embeddings), hoping to capturing the meaning of the according words. Following previous work, we evaluate our model on two paraphrase-oriented tasks, namely text simplification and short text abstractive summarization. Experimental results show that our model outperforms the sequence-to-sequence baseline by the BLEU score of 6.3 and 5.5 on two English text simplification datasets, and the ROUGE-2 F1 score of 5.7 on a Chinese summarization dataset. Moreover, our model achieves state-of-the-art performances on these three benchmark datasets.


international joint conference on artificial intelligence | 2018

A Hierarchical End-to-End Model for Jointly Improving Text Summarization and Sentiment Classification

Shuming Ma; Xu Sun; Junyang Lin; Xuancheng Ren

Text summarization and sentiment classification both aim to capture the main ideas of the text but at different levels. Text summarization is to describe the text within a few sentences, while sentiment classification can be regarded as a special type of summarization which summarizes the text into a even more abstract fashion, i.e., a sentiment class. Based on this idea, we propose a hierarchical end-to-end model for joint learning of text summarization and sentiment classification, where the sentiment classification label is treated as the further summarization of the text summarization output. Hence, the sentiment classification layer is put upon the text summarization layer, and a hierarchical structure is derived. Experimental results on Amazon online reviews datasets show that our model achieves better performance than the strong baseline systems on both abstractive summarization and sentiment classification.


international conference natural language processing | 2018

Accelerating Graph-Based Dependency Parsing with Lock-Free Parallel Perceptron

Shuming Ma; Xu Sun; Yi Zhang; Bingzhen Wei

Dependency parsing is an important NLP task. A popular approach for dependency parsing is structured perceptron. Still, graph-based dependency parsing has the time complexity of (O(n^3)), and it suffers from slow training. To deal with this problem, we propose a parallel algorithm called parallel perceptron. The parallel algorithm can make full use of a multi-core computer which saves a lot of training time. Based on experiments we observe that dependency parsing with parallel perceptron can achieve 8-fold faster training speed than traditional structured perceptron methods when using 10 threads, and with no loss at all in accuracy.


meeting of the association for computational linguistics | 2018

Bag-of-Words as Target for Neural Machine Translation

Shuming Ma; Xu Sun; Yizhong Wang; Junyang Lin


international conference on machine learning | 2017

meProp: Sparsified Back Propagation for Accelerated Deep Learning with Reduced Overfitting.

Xu Sun; Xuancheng Ren; Shuming Ma; Houfeng Wang


meeting of the association for computational linguistics | 2018

Global Encoding for Abstractive Summarization

Junyang Lin; Xu Sun; Shuming Ma; Qi Su


arXiv: Computation and Language | 2018

Decoding-History-Based Adaptive Control of Attention for Neural Machine Translation.

Junyang Lin; Shuming Ma; Qi Su; Xu Sun


arXiv: Learning | 2018

Label Embedding Network: Learning Label Representation for Soft Training of Deep Networks

Xu Sun; Bingzhen Wei; Xuancheng Ren; Shuming Ma

Collaboration


Dive into the Shuming Ma's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Wei Li

Tsinghua University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Wenjie Li

Hong Kong Polytechnic University

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge