Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Jiaming Xu is active.

Publication


Featured researches published by Jiaming Xu.


Neurocomputing | 2016

Semantic expansion using word embedding clustering and convolutional neural network for improving short text classification

Peng Wang; Bo Xu; Jiaming Xu; Guanhua Tian; Cheng-Lin Liu; Hong-Wei Hao

Text classification can help users to effectively handle and exploit useful information hidden in large-scale documents. However, the sparsity of data and the semantic sensitivity to context often hinder the classification performance of short texts. In order to overcome the weakness, we propose a unified framework to expand short texts based on word embedding clustering and convolutional neural network (CNN). Empirically, the semantically related words are usually close to each other in embedding spaces. Thus, we first discover semantic cliques via fast clustering. Then, by using additive composition over word embeddings from context with variable window width, the representations of multi-scale semantic units11Semantic units are defined as n-grams which have dominant meaning of text. With n varying, multi-scale contextual information can be exploited. in short texts are computed. In embedding spaces, the restricted nearest word embeddings (NWEs)22In order to prevent outliers, a Euclidean distance threshold is preset between semantic cliques and semantic units, which is used as restricted condition. of the semantic units are chosen to constitute expanded matrices, where the semantic cliques are used as supervision information. Finally, for a short text, the projected matrix33The projected matrix is obtained by table looking up, which encodes Unigram level features. and expanded matrices are combined and fed into CNN in parallel. Experimental results on two open benchmarks validate the effectiveness of the proposed method.


international joint conference on natural language processing | 2015

Semantic Clustering and Convolutional Neural Network for Short Text Categorization

Peng Wang; Jiaming Xu; Bo Xu; Cheng-Lin Liu; Heng Zhang; Fangyuan Wang; Hongwei Hao

Short texts usually encounter data sparsity and ambiguity problems in representations for their lack of context. In this paper, we propose a novel method to model short texts based on semantic clustering and convolutional neural network. Particularly, we first discover semantic cliques in embedding spaces by a fast clustering algorithm. Then, multi-scale semantic units are detected under the supervision of semantic cliques, which introduce useful external knowledge for short texts. These meaningful semantic units are combined and fed into convolutional layer, followed by max-pooling operation. Experimental results on two open benchmarks validate the effectiveness of the proposed method.


north american chapter of the association for computational linguistics | 2015

Short Text Clustering via Convolutional Neural Networks

Jiaming Xu; peng wang; Guanhua Tian; Bo Xu; Jun Zhao; Fangyuan Wang; Hongwei Hao

Short text clustering has become an increasing important task with the popularity of social media, and it is a challenging problem due to its sparseness of text representation. In this paper, we propose a Short Text Clustering via Convolutional neural networks (abbr. to STCC), which is more beneficial for clustering by considering one constraint on learned features through a self-taught learning framework without using any external tags/labels. First, we embed the original keyword features into compact binary codes with a localitypreserving constraint. Then, word embeddings are explored and fed into convolutional neural networks to learn deep feature representations, with the output units fitting the pre-trained binary code in the training process. After obtaining the learned representations, we use K-means to cluster them. Our extensive experimental study on two public short text datasets shows that the deep feature representation learned by our approach can achieve a significantly better performance than some other existing features, such as term frequency-inverse document frequency, Laplacian eigenvectors and average embedding, for clustering.


Neural Networks | 2017

Self-Taught convolutional neural networks for short text clustering

Jiaming Xu; Bo Xu; Peng Wang; Suncong Zheng; Guanhua Tian; Jun Zhao

Short text clustering is a challenging problem due to its sparseness of text representation. Here we propose a flexible Self-Taught Convolutional neural network framework for Short Text Clustering (dubbed STC2), which can flexibly and successfully incorporate more useful semantic features and learn non-biased deep text representation in an unsupervised manner. In our framework, the original raw text features are firstly embedded into compact binary codes by using one existing unsupervised dimensionality reduction method. Then, word embeddings are explored and fed into convolutional neural networks to learn deep feature representations, meanwhile the output units are used to fit the pre-trained binary codes in the training process. Finally, we get the optimal clusters by employing K-means to cluster the learned representations. Extensive experimental results demonstrate that the proposed framework is effective, flexible and outperform several popular clustering methods when tested on three public short text datasets.


Knowledge Based Systems | 2016

A neural network framework for relation extraction

Suncong Zheng; Jiaming Xu; Peng Zhou; Hongyun Bao; Zhenyu Qi; Bo Xu

Relation extraction is to identify the relationship of two given entities in the text. It is an important step in the task of knowledge extraction. Most conventional methods for the task of relation extraction focus on designing effective handcrafted features or learning a semantic representation of the whole sentence. Sentences with the same relationship always share the similar expressions. Besides, the semantic properties of given entities can also help to distinguish some confusing relations. Based on the above observations, we propose a neural network based framework for relation classification. It can simultaneously learn the relation patterns information and the semantic properties of given entities. In this framework, we explore two specific models: the CNN-based model and LSTM-based model. We conduct experiments on two public datasets: the SemEval-2010 Task8 dataset and the ACE05 dataset. The proposed method achieves the state-of-the-art result without using any external information. Additionally, the experimental results also show that our approach can represent the semantic relationship of the given entities effectively.


NLPCC | 2013

A Fast Matching Method Based on Semantic Similarity for Short Texts

Jiaming Xu; Pengcheng Liu; Gaowei Wu; Zhengya Sun; Bo Xu; Hongwei Hao

As the emergence of various social media, short texts, such as weibos and instant messages, are very prevalent on today’s websites. In order to mine semantically similar information from massive data, a fast and efficient matching method for short texts has become an urgent task. However, the conventional matching methods suffer from the data sparsity in short documents. In this paper, we propose a novel matching method, referred as semantically similar hashing (SSHash). The basic idea of SSHash is to directly train a topic model from corpus rather than documents, then project texts into hash codes by using latent features. The major advantages of SSHash are that 1) SSHash alleviates the sparse problem in short texts, because we obtain the latent features from whole corpus regardless of document level; and 2) SSHash can accomplish similar matching in an interactive real time by introducing hash method. We carry out extensive experiments on real-world short texts. The results demonstrate that our method significantly outperforms baseline methods on several evaluation metrics.


web intelligence | 2016

Compositional Recurrent Neural Networks for Chinese Short Text Classification

Yujun Zhou; Bo Xu; Jiaming Xu; Lei Yang; Changliang Li

Word segmentation is the first step in Chinese natural language processing, and the error caused by word segmentation can be transmitted to the whole system. In order to reduce the impact of word segmentation and improve the overall performance of Chinese short text classification system, we propose a hybrid model of character-level and word-level features based on recurrent neural network (RNN) with long short-term memory (LSTM). By integrating character-level feature into word-level feature, the missing semantic information by the error of word segmentation will be constructed, meanwhile the wrong semantic relevance will be reduced. The final feature representation is that it suppressed the error of word segmentation in the case of maintaining most of the semantic features of the sentence. The whole model is finally trained end-to-end with supervised Chinese short text classification task. Results demonstrate that the proposed model in this paper is able to represent Chinese short text effectively, and the performances of 32-class and 5-class categorization outperform some remarkable methods.


international joint conference on artificial intelligence | 2018

Listen, Think and Listen Again: Capturing Top-down Auditory Attention for Speaker-independent Speech Separation

Jing Shi; Jiaming Xu; Guangcan Liu; Bo Xu

Recent deep learning methods have made significant progress in multi-talker mixed speech separation. However, most existing models adopt a driftless strategy to separate all the speech channels rather than selectively attend the target one. As a result, those frameworks may be failed to offer a satisfactory solution in complex auditory scene where the number of input sounds is usually uncertain and even dynamic. In this paper, we present a novel neural network based structure motivated by the top-down attention behavior of human when facing complicated acoustical scene. Different from previous works, our method constructs an inferenceattention structure to predict interested candidates and extract each speech channel of them. Our work gets rid of the limitation that the number of channels must be given or the high computation complexity for label permutation problem. We evaluated our model on the WSJ0 mixed-speech tasks. In all the experiments, our model gets highly competitive to reach and even outperform the baselines.


Neural Networks | 2018

Distant supervision for relation extraction with hierarchical selective attention

Peng Zhou; Jiaming Xu; Zhenyu Qi; Hongyun Bao; Zhineng Chen; Bo Xu

Distant supervised relation extraction is an important task in the field of natural language processing. There are two main shortcomings for most state-of-the-art methods. One is that they take all sentences of an entity pair as input, which would result in a large computational cost. But in fact, few of most relevant sentences are enough to recognize the relation of an entity pair. To tackle these problems, we propose a novel hierarchical selective attention network for relation extraction under distant supervision. Our model first selects most relevant sentences by taking coarse sentence-level attention on all sentences of an entity pair and then employs word-level attention to construct sentence representations and fine sentence-level attention to aggregate these sentence representations. Experimental results on a widely used dataset demonstrate that our method performs significantly better than most of existing methods.


international conference on neural information processing | 2017

Hierarchical Hybrid Attention Networks for Chinese Conversation Topic Classification

Yujun Zhou; Changliang Li; Bo Xu; Jiaming Xu; Jie Cao

Topic classification is useful for applications such as forensics analysis and cyber-crime investigation. To improve the overall performance on the task of Chinese conversation topic classification, we propose a hierarchical neural network with automatic semantic features selection, which is a hierarchical architecture that depicts the structure of conversations. The model firstly incorporates speaker information into the character- and word-level attentions and generates sentence representation, then uses attention-based BLSTM to construct the conversation representation. Experimental results on three datasets demonstrate that our model achieves better performance than multiple baselines. It indicates that the proposed architecture can capture the informative and salient features related to the meaning of a conversation for topic classification. And we release the dataset of this paper that can be obtained from https://github.com/njoe9/H-HANs.

Collaboration


Dive into the Jiaming Xu's collaboration.

Top Co-Authors

Avatar

Bo Xu

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Hongwei Hao

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Suncong Zheng

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Hongyun Bao

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Jing Shi

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Guanhua Tian

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Zhenyu Qi

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Jun Zhao

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Peng Wang

Chinese Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Changliang Li

Chinese Academy of Sciences

View shared research outputs
Researchain Logo
Decentralizing Knowledge