Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Luu Anh Tuan is active.

Publication


Featured researches published by Luu Anh Tuan.


international acm sigir conference on research and development in information retrieval | 2017

Learning to Rank Question Answer Pairs with Holographic Dual LSTM Architecture

Yi Tay; Minh C. Phan; Luu Anh Tuan; Siu Cheung Hui

We describe a new deep learning architecture for learning to rank question answer pairs. Our approach extends the long short-term memory (LSTM) network with holographic composition to model the relationship between question and answer representations. As opposed to the neural tensor layer that has been adopted recently, the holographic composition provides the benefits of scalable and rich representational learning approach without incurring huge parameter costs. Overall, we present Holographic Dual LSTM (HD-LSTM), a unified architecture for both deep sentence modeling and semantic matching. Essentially, our model is trained end-to-end whereby the parameters of the LSTM are optimized in a way that best explains the correlation between question and answer representations. In addition, our proposed deep learning architecture requires no extensive feature engineering. Via extensive experiments, we show that HD-LSTM outperforms many other neural architectures on two popular benchmark QA datasets. Empirical studies confirm the effectiveness of holographic composition over the neural tensor layer.


conference on information and knowledge management | 2017

Dyadic Memory Networks for Aspect-based Sentiment Analysis

Yi Tay; Luu Anh Tuan; Siu Cheung Hui

This paper proposes Dyadic Memory Networks (DyMemNN), a novel extension of end-to-end memory networks (memNN) for aspect-based sentiment analysis (ABSA). Originally designed for question answering tasks, memNN operates via a memory selection operation in which relevant memory pieces are adaptively selected based on the input query. In the problem of ABSA, this is analogous to aspects and documents in which the relationship between each word in the document is compared with the aspect vector. In the standard memory networks, simple dot products or feed forward neural networks are used to model the relationship between aspect and words which lacks representation learning capability. As such, our dyadic memory networks ameliorates this weakness by enabling rich dyadic interactions between aspect and word embeddings by integrating either parameterized neural tensor compositions or holographic compositions into the memory selection operation. To this end, we propose two variations of our dyadic memory networks, namely the Tensor DyMemNN and Holo DyMemNN. Overall, our two models are end-to-end neural architectures that enable rich dyadic interaction between aspect and document which intuitively leads to better performance. Via extensive experiments, we show that our proposed models achieve the state-of-the-art performance and outperform many neural architectures across six benchmark datasets.


web search and data mining | 2018

Hyperbolic Representation Learning for Fast and Efficient Neural Question Answering

Yi Tay; Luu Anh Tuan; Siu Cheung Hui

The dominant neural architectures in question answer retrieval are based on recurrent or convolutional encoders configured with complex word matching layers. Given that recent architectural innovations are mostly new word interaction layers or attention-based matching mechanisms, it seems to be a well-established fact that these components are mandatory for good performance. Unfortunately, the memory and computation cost incurred by these complex mechanisms are undesirable for practical applications. As such, this paper tackles the question of whether it is possible to achieve competitive performance with simple neural architectures. We propose a simple but novel deep learning architecture for fast and efficient question-answer ranking and retrieval. More specifically, our proposed model, HyperQA, is a parameter efficient neural network that outperforms other parameter intensive models such as Attentive Pooling BiLSTMs and Multi-Perspective CNNs on multiple QA benchmarks. The novelty behind HyperQA is a pairwise ranking objective that models the relationship between question and answer embeddings in Hyperbolic space instead of Euclidean space. This empowers our model with a self-organizing ability and enables automatic discovery of latent hierarchies while learning embeddings of questions and answers. Our model requires no feature engineering, no similarity matrix matching, no complicated attention mechanisms nor over-parameterized layers and yet outperforms and remains competitive to many models that have these functionalities on multiple benchmarks.Many state-of-the-art deep learning models for question answer retrieval are highly complex, ox89en having a huge number of parameters or complicated word interaction mechanisms. x8cis paper studies if it is possible to achieve equally competitive performance with smaller and faster neural architectures. Overall, our proposed approach is a simple neural network that performs question-answer matching and ranking in Hyperbolic space. We show that QA embeddings learned in Hyperbolic space results in highly competitive performance on multiple benchmarks, outperforming models with signix80cantly much larger parameters. Our proposed approach (90K parameters) remains competitive to models with millions of parameters such as Ax8aentive Pooling BiLSTMs or Multi-Perspective Convolutional Neural Networks (MP-CNN).


conference on information and knowledge management | 2017

Multi-Task Neural Network for Non-discrete Attribute Prediction in Knowledge Graphs

Yi Tay; Luu Anh Tuan; Minh C. Phan; Siu Cheung Hui

Many popular knowledge graphs such as Freebase, YAGO or DBPedia maintain a list of non-discrete attributes for each entity. Intuitively, these attributes such as height, price or population count are able to richly characterize entities in knowledge graphs. This additional source of information may help to alleviate the inherent sparsity and incompleteness problem that are prevalent in knowledge graphs. Unfortunately, many state-of-the-art relational learning models ignore this information due to the challenging nature of dealing with non-discrete data types in the inherently binary-natured knowledge graphs. In this paper, we propose a novel multi-task neural network approach for both encoding and prediction of non-discrete attribute information in a relational setting. Specifically, we train a neural network for triplet prediction along with a separate network for attribute value regression. Via multi-task learning, we are able to learn representations of entities, relations and attributes that encode information about both tasks. Moreover, such attributes are not only central to many predictive tasks as an information source but also as a prediction target. Therefore, models that are able to encode, incorporate and predict such information in a relational learning context are highly attractive as well. We show that our approach outperforms many state-of-the-art methods for the tasks of relational triplet classification and attribute value prediction.


knowledge discovery and data mining | 2018

Multi-Cast Attention Networks

Yi Tay; Luu Anh Tuan; Siu Cheung Hui

Attention is typically used to select informative sub-phrases that are used for prediction. This paper investigates the novel use of attention as a form of feature augmentation, i.e, casted attention. We propose Multi-Cast Attention Networks (MCAN), a new attention mechanism and general model architecture for a potpourri of ranking tasks in the conversational modeling and question answering domains. Our approach performs a series of soft attention operations, each time casting a scalar feature upon the inner word embeddings. The key idea is to provide a real-valued hint (feature) to a subsequent encoder layer and is targeted at improving the representation learning process. There are several advantages to this design, e.g., it allows an arbitrary number of attention mechanisms to be casted, allowing for multiple attention types (e.g., co-attention, intra-attention) and attention variants (e.g., alignment-pooling, max-pooling, mean-pooling) to be executed simultaneously. This not only eliminates the costly need to tune the nature of the co-attention layer, but also provides greater extents of explainability to practitioners. Via extensive experiments on four well-known benchmark datasets, we show that MCAN achieves state-of-the-art performance. On the Ubuntu Dialogue Corpus, MCAN outperforms existing state-of-the-art models by 9%. MCAN also achieves the best performing score to date on the well-studied TrecQA dataset.


arXiv: Computation and Language | 2018

A Compare-Propagate Architecture with Alignment Factorization for Natural Language Inference.

Yi Tay; Luu Anh Tuan; Siu Cheung Hui


arXiv: Computation and Language | 2018

Multi-range Reasoning for Machine Comprehension.

Yi Tay; Luu Anh Tuan; Siu Cheung Hui


international conference on weblogs and social media | 2018

CoupleNet: Paying Attention to Couples with Coupled Attention for Relationship Recommendation.

Yi Tay; Luu Anh Tuan; Siu Cheung Hui


arXiv: Information Retrieval | 2018

Self-Attentive Neural Collaborative Filtering

Yi Tay; Shuai Zhang; Luu Anh Tuan; Siu Cheung Hui


arXiv: Computation and Language | 2018

Multi-Cast Attention Networks for Retrieval-based Question Answering and Response Prediction.

Yi Tay; Luu Anh Tuan; Siu Cheung Hui

Collaboration


Dive into the Luu Anh Tuan's collaboration.

Top Co-Authors

Avatar

Siu Cheung Hui

Nanyang Technological University

View shared research outputs
Top Co-Authors

Avatar

Yi Tay

Nanyang Technological University

View shared research outputs
Top Co-Authors

Avatar

Minh C. Phan

Nanyang Technological University

View shared research outputs
Researchain Logo
Decentralizing Knowledge