Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Baolin Peng is active.

Publication


Featured researches published by Baolin Peng.


spoken language technology workshop | 2014

Spoken language understanding using long short-term memory neural networks

Kaisheng Yao; Baolin Peng; Yu Zhang; Dong Yu; Geoffrey Zweig; Yangyang Shi

Neural network based approaches have recently produced record-setting performances in natural language understanding tasks such as word labeling. In the word labeling task, a tagger is used to assign a label to each word in an input sequence. Specifically, simple recurrent neural networks (RNNs) and convolutional neural networks (CNNs) have shown to significantly outperform the previous state-of-the-art - conditional random fields (CRFs). This paper investigates using long short-term memory (LSTM) neural networks, which contain input, output and forgetting gates and are more advanced than simple RNN, for the word labeling task. To explicitly model output-label dependence, we propose a regression model on top of the LSTM un-normalized scores. We also propose to apply deep LSTM to the task. We investigated the relative importance of each gate in the LSTM by setting other gates to a constant and only learning particular gates. Experiments on the ATIS dataset validated the effectiveness of the proposed models.


international conference on acoustics, speech, and signal processing | 2015

Contextual spoken language understanding using recurrent neural networks

Yangyang Shi; Kaisheng Yao; Hu Chen; Yi-Cheng Pan; Mei-Yuh Hwang; Baolin Peng

We present a contextual spoken language understanding (contextual SLU) method using Recurrent Neural Networks (RNNs). Previous work has shown that context information, specifically the previously estimated domain assignment, is helpful for domain identification. We further show that other context information such as the previously estimated intent and slot labels are useful for both intent classification and slot filling tasks in SLU. We propose a step-n-gram model to extract sentence-level features from RNNs, which extract sequential features. The step-n-gram model is used together with a stack of Convolution Networks for training domain/intent classification. Our method therefore exploits possible correlations among domain/intent classification and slot filling and incorporates context information from the past predictions of domain/intent and slots. The proposed method obtains new state-of-the-art results on ATIS and improved performances over baseline techniques such as conditional random fields (CRFs) on a large context-sensitive SLU dataset.


arXiv: Computation and Language | 2015

Recurrent Neural Networks with External Memory for Spoken Language Understanding

Baolin Peng; Kaisheng Yao; Li Jing; Kam-Fai Wong

Recurrent Neural Networks RNNs have become increasingly popular for the task of language understanding. In this task, a semantic tagger is deployed to associate a semantic label to each word in an input sequence. The success of RNN may be attributed to its ability to memorise long-term dependence that relates the current-time semantic label prediction to the observations many time instances away. However, the memory capacity of simple RNNs is limited because of the gradient vanishing and exploding problem. We propose to use an external memory to improve memorisation capability of RNNs. Experiments on the ATIS dataset demonstrated that the proposed model was able to achieve the state-of-the-art results. Detailed analysis may provide insights for future research.


empirical methods in natural language processing | 2015

Using Content-level Structures for Summarizing Microblog Repost Trees

Jing Li; Wei Gao; Zhongyu Wei; Baolin Peng; Kam-Fai Wong

A microblog repost tree provides strong clues on how an event described therein develops. To help social media users capture the main clues of events on microblogging sites, we propose a novel repost tree summarization framework by effectively differentiating two kinds of messages on repost trees called leaders and followers, which are derived from contentlevel structure information, i.e., contents of messages and the reposting relations. To this end, Conditional Random Fields (CRF) model is used to detect leaders across repost tree paths. We then present a variant of random-walk-based summarization model to rank and select salient messages based on the result of leader detection. To reduce the error propagation cascaded from leader detection, we improve the framework by enhancing the random walk with adjustment steps for sampling from leader probabilities given all the reposting messages. For evaluation, we construct two annotated corpora, one for leader detection, and the other for repost tree summarization. Experimental results confirm the effectiveness of our method.


international conference on acoustics, speech, and signal processing | 2015

A factorization network based method for multi-lingual domain classification

Yangyang Shi; Yi-Cheng Pan; Mei-Yuh Hwang; Kaisheng Yao; Hu Chen; Yuanhang Zou; Baolin Peng

In many spoken language understanding systems (SLUS), domain classification is the most crucial component, as system responses based on wrong domains often yield very unpleasant user experiences. In multi-lingual domain classification, the training data for some poor-resource languages often comes from machine translation. Some of the higher order n-gram features are distorted during machine translation. Feature co-occurrence becomes reliable feature in multi-lingual domain classification. In this paper, in order to effectively model feature co-occurrences, we propose Factorization Networks that are combinations of Factorization Machines (FMs) with Neural Networks (NNs). FNs extend the linear connections from the input feature layer to the hidden layer in NNs to factorization connections that represent the weights of feature co-occurrences using factorized method. In addition to FNs, we also propose a hybrid model that integrates FNs, NNs and Maximum Entropy (ME) models together. The component models in the hybrid model share the same input features. Based on two data sets (ATIS data set and Microsoft Cortana Chinese data ), the proposed models shows promising results. Especially for large Microsoft Cortana Chinese data which is translated from well annotated English data, FNs using unigram, class and query length features achieve more than 20% relative error reduction over linear (SVMs).


conference on intelligent text processing and computational linguistics | 2015

Trending Sentiment-Topic Detection on Twitter

Baolin Peng; Jing Li; Junwen Chen; Xu Han; Ruifeng Xu; Kam-Fai Wong

Twitter plays a significant role in information diffusion and has evolved to an important information resource as well as news feed. People wonder and care about what is happening on Twitter and what news it is bringing to us every moment. However, with huge amount of data, it is impossible to tell what topic is trending on time manually, which makes real-time topic detection attractive and significant. Furthermore, Twitter provides a platform of opinion sharing and sentiment expression for events, news, products etc. Users intend to tell what they are really thinking about on Twitter thus makes Twitter a valuable source of opinions. Nevertheless, most works about trending topic detection fail to take sentiment into consideration. This work is based on a non-parametric supervised real-time trending topic detection model with sentimental feature. Experiment shows our model successfully detects trending sentimental topic in the shortest time. After a combination of multiple features, e.g. tweet volume and user volume, it demonstrates impressive effectiveness with 82.3% recall and surpasses all the competitors.


Archive | 2015

An Introduction to Computational Networks and the Computational Network Toolkit

Dong Yu; Adam Eversole; Michael L. Seltzer; Kaisheng Yao; Zhiheng Huang; Brian K. Guenter; Oleksii Kuchaiev; Yu Zhang; Frank Seide; Huaming Wang; Jasha Droppo; Geoffrey Zweig; Christopher J. Rossbach; Jon Currey; Jie Gao; Baolin Peng; Andreas Stolcke; Malcolm Slaney


arXiv: Neural and Evolutionary Computing | 2015

Attention with Intention for a Neural Network Conversation Model.

Kaisheng Yao; Geoffrey Zweig; Baolin Peng


arXiv: Artificial Intelligence | 2015

Towards Neural Network-based Reasoning

Baolin Peng; Zhengdong Lu; Hang Li; Kam-Fai Wong


empirical methods in natural language processing | 2017

Composite Task-Completion Dialogue Policy Learning via Hierarchical Deep Reinforcement Learning.

Baolin Peng; Xiujun Li; Lihong Li; Jianfeng Gao; Asli Celikyilmaz; Sungjin Lee; Kam-Fai Wong

Collaboration


Dive into the Baolin Peng's collaboration.

Top Co-Authors

Avatar

Kam-Fai Wong

The Chinese University of Hong Kong

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Xiujun Li

University of Wisconsin-Madison

View shared research outputs
Top Co-Authors

Avatar

Jingjing Liu

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Zhongyu Wei

University of Texas at Dallas

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge