Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Hanxiao Liu is active.

Publication


Featured researches published by Hanxiao Liu.


meeting of the association for computational linguistics | 2017

Gated-Attention Readers for Text Comprehension

Bhuwan Dhingra; Hanxiao Liu; Zhilin Yang; William W. Cohen; Ruslan Salakhutdinov

In this paper we study the problem of answering cloze-style questions over documents. Our model, the Gated-Attention (GA) Reader, integrates a multi-hop architecture with a novel attention mechanism, which is based on multiplicative interactions between the query embedding and the intermediate states of a recurrent neural network document reader. This enables the reader to build query-specific representations of tokens in the document for accurate answer selection. The GA Reader obtains state-of-the-art results on three benchmarks for this task--the CNN \& Daily Mail news stories and the Who Did What dataset. The effectiveness of multiplicative interaction is demonstrated by an ablation study, and by comparing to alternative compositional operators for implementing the gated-attention. The code is available at this https URL


web search and data mining | 2015

Concept Graph Learning from Educational Data

Yiming Yang; Hanxiao Liu; Jaime G. Carbonell; Wanli Ma

This paper addresses an open challenge in educational data mining, i.e., the problem of using observed prerequisite relations among courses to learn a directed universal concept graph, and using the induced graph to predict unobserved prerequisite relations among a broader range of courses. This is particularly useful to induce prerequisite relations among courses from different providers (universities, MOOCs, etc.). We propose a new framework for inference within and across two graphs---at the course level and at the induced concept level---which we call Concept Graph Learning (CGL). In the training phase, our system projects the course-level links onto the concept space to induce directed concept links; in the testing phase, the concept links are used to predict (unobserved) prerequisite links for test-set courses within the same institution or across institutions. The dual mappings enable our system to perform an interlingua-style transfer learning, e.g. treating the concept graph as the interlingua, and inducing prerequisite links in a transferable manner across different universities. Experiments on our newly collected data sets of courses from MIT, Caltech, Princeton and CMU show promising results, including the viability of CGL for transfer learning.


conference on information and knowledge management | 2016

Cross-lingual Text Classification via Model Translation with Limited Dictionaries

Ruochen Xu; Yiming Yang; Hanxiao Liu; Andrew Hsi

Cross-lingual text classification (CLTC) refers to the task of classifying documents in different languages into the same taxonomy of categories. An open challenge in CLTC is to classify documents for the languages where labeled training data are not available. Existing approaches rely on the availability of either high-quality machine translation of documents (to the languages where massively training data are available), or rich bilingual dictionaries for effective translation of trained classification models (to the languages where labeled training data are lacking). This paper studies the CLTC challenge under the assumption that neither condition is met. That is, we focus on the problem of translating classification models with highly incomplete bilingual dictionaries. Specifically, we propose two new approaches that combines unsupervised word embedding in different languages, supervised mapping of embedded words across languages, and probabilistic translation of classification models. The approaches show significant performance improvement in CLTC on a benchmark corpus of Reuters news stories (RCV1/RCV2) in English, Spanish, German, French and Chinese and an internal dataset in Uzbek, compared to representative baseline methods using conventional bilingual dictionaries or highly incomplete ones.


Journal of Artificial Intelligence Research | 2016

Learning concept graphs from online educational data

Hanxiao Liu; Wanli Ma; Yiming Yang; Jaime G. Carbonell

This paper addresses an open challenge in educational data mining, i.e., the problem of automatically mapping online courses from different providers (universities, MOOCs, etc.) onto a universal space of concepts, and predicting latent prerequisite dependencies (directed links) among both concepts and courses. We propose a novel approach for inference within and across course-level and concept-level directed graphs. In the training phase, our system projects partially observed course-level prerequisite links onto directed concept-level links; in the testing phase, the induced concept-level links are used to infer the unknown course-level prerequisite links. Whereas courses may be specific to one institution, concepts are shared across different providers. The bi-directional mappings enable our system to perform interlingua-style transfer learning, e.g. treating the concept graph as the interlingua and transferring the prerequisite relations across universities via the interlingua. Experiments on our newly collected datasets of courses from MIT, Caltech, Princeton and CMU show promising results.


international acm sigir conference on research and development in information retrieval | 2018

Modeling Long- and Short-Term Temporal Patterns with Deep Neural Networks

Guokun Lai; Wei-Cheng Chang; Yiming Yang; Hanxiao Liu

Multivariate time series forecasting is an important machine learning problem across many domains, including predictions of solar plant energy output, electricity consumption, and traffic jam situation. Temporal data arise in these real-world applications often involves a mixture of long-term and short-term patterns, for which traditional approaches such as Autoregressive models and Gaussian Process may fail. In this paper, we proposed a novel deep learning framework, namely Long- and Short-term Time-series network (LSTNet), to address this open challenge. LSTNet uses the Convolution Neural Network (CNN) and the Recurrent Neural Network (RNN) to extract short-term local dependency patterns among variables and to discover long-term patterns for time series trends. Furthermore, we leverage traditional autoregressive model to tackle the scale insensitive problem of the neural network model. In our evaluation on real-world data with complex mixtures of repetitive patterns, LSTNet achieved significant performance improvements over that of several state-of-the-art baseline methods. All the data and experiment codes are available online.


2017 IEEE International Conference on Software Architecture (ICSA) | 2017

Experiments in Curation: Towards Machine-Assisted Construction of Software Architecture Knowledge Bases

Ian Gorton; Rouchen Xu; Yiming Yang; Hanxiao Liu; Guoqing Zheng

Software architects inhabit a complex, rapidly evolving technological landscape. An ever growing collection of competing architecturally significant technologies, ranging from distributed databases to middleware and cloud platforms, makes rigorously comparing alternatives and selecting appropriate solutions a daunting engineering task. To address this problem, we envisage an ecosystem of curated, automatically updated knowledge bases that enable straightforward and streamlined technical comparisons of related products. These knowledge bases would emulate engineering handbooks that are commonly found in other engineering disciplines. As a first step towards this vision, we have built a curated knowledge base for comparing distributed databases based on a semantically defined feature taxonomy. We report in this paper on the initial results of using supervised machine learning to assist with knowledge base curation. Our results show immense promise in recommending Web pages that are highly relevant to curators. We also describe the major obstacles, both practical and scientific, that our work has uncovered. These must be overcome by future research in order to make our vision of curated knowledge bases a reality.


international conference on learning representations | 2018

Hierarchical Representations for Efficient Architecture Search

Hanxiao Liu; Karen Simonyan; Oriol Vinyals; Chrisantha Fernando; Koray Kavukcuoglu


empirical methods in natural language processing | 2017

RACE: Large-scale ReAding Comprehension Dataset From Examinations

Guokun Lai; Qizhe Xie; Hanxiao Liu; Yiming Yang; Eduard H. Hovy


international conference on machine learning | 2017

Analogical Inference for Multi-relational Embeddings

Hanxiao Liu; Yuexin Wu; Yiming Yang


international conference on machine learning | 2015

Bipartite Edge Prediction via Transductive Learning over Product Graphs

Hanxiao Liu; Yiming Yang

Collaboration


Dive into the Hanxiao Liu's collaboration.

Top Co-Authors

Avatar

Yiming Yang

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Guokun Lai

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Yuexin Wu

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Bhuwan Dhingra

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Wanli Ma

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

William W. Cohen

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Wei-Cheng Chang

National Taiwan University

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge