Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where John D. Kelleher is active.

Publication


Featured researches published by John D. Kelleher.


meeting of the association for computational linguistics | 2006

Incremental Generation of Spatial Referring Expressions in Situated Dialog

John D. Kelleher; Geert-Jan M. Kruijff

This paper presents an approach to incrementally generating locative expressions. It addresses the issue of combinatorial explosion inherent in the construction of relational context models by: (a) contextually defining the set of objects in the context that may function as a landmark, and (b) sequencing the order in which spatial relations are considered using a cognitively motivated hierarchy of relations, and visual and discourse salience.


Computational Linguistics | 2009

Applying computational models of spatial prepositions to visually situated dialog

John D. Kelleher; Fintan Costello

This article describes the application of computational models of spatial prepositions to visually situated dialog systems. In these dialogs, spatial prepositions are important because people often use them to refer to entities in the visual context of a dialog. We first describe a generic architecture for a visually situated dialog system and highlight the interactions between the spatial cognition module, which provides the interface to the models of prepositional semantics, and the other components in the architecture. Following this, we present two new computational models of topological and projective spatial prepositions. The main novelty within these models is the fact that they account for the contextual effect which other distractor objects in a visual scene can have on the region described by a given preposition. We next present psycholinguistic tests evaluating our approach to distractor interference on prepositional semantics, and illustrate how these models are used for both interpretation and generation of prepositional expressions.


perception and interactive technologies | 2006

Information fusion for visual reference resolution in dynamic situated dialogue

Geert-Jan M. Kruijff; John D. Kelleher; Nick Hawes

Human-Robot Interaction (HRI) invariably involves dialogue about objects in the environment in which the agents are situated. The paper focuses on the issue of resolving discourse references to such visual objects. The paper addresses the problem using strategies for intra-modal fusion (identifying that different occurrences concern the same object), and inter-modal fusion, (relating object references across different modalities). Core to these strategies are sensorimotoric coordination, and ontology-based mediation between content in different modalities. The approach has been fully implemented, and is illustrated with several working examples.


human-robot interaction | 2006

Structural descriptions in human-assisted robot visual learning

Geert-Jan M. Kruijff; John D. Kelleher; Gregor Berginc; Aleš Leonardis

The paper presents an approach to using structural descriptions, obtained through a human-robot tutoring dialogue, as labels for the visual object models a robot learns. The paper shows how structural descriptions enable relating models for different aspects of one and the same object, and how being able to relate descriptions for visual models and discourse referents enables incremental updating of model descriptions through dialogue (either robot- or human initiated). The approach has been implemented in an integrated architecture for human-assisted robot visual learning.


Artificial Intelligence Review | 2006

Attention driven reference resolution in multimodal contexts

John D. Kelleher

In recent years a number of psycholinguistic experiments have pointed to the interaction between language and vision. In particular, the interaction between visual attention and linguistic reference. In parallel with this, several theories of discourse have attempted to provide an account of the relationship between types of referential expressions on the one hand and the degree of mental activation on the other. Building on both of these traditions, this paper describes an attention based approach to visually situated reference resolution. The framework uses the relationship between referential form and preferred mode of interpretation as a basis for a weighted integration of linguistic and visual attention scores for each entity in the multimodal context. The resulting integrated attention scores are then used to rank the candidate referents during the resolution process, with the candidate scoring the highest selected as the referent. One advantage of this approach is that the resolution process occurs within the full multimodal context, in so far as the referent is selected from a full list of the objects in the multimodal context. As a result situations where the intended target of the reference is erroneously excluded, due to an individual assumption within the resolution process, are avoided. Moreover, the system can recognise situations where attention cues from different modalities make a reference potentially ambiguous.


Proceedings of the 3rd Workshop on Hybrid Approaches to Machine Translation (HyTra) | 2014

An Empirical Study of the Impact of Idioms on Phrase Based Statistical Machine Translation of English to Brazilian-Portuguese

Giancarlo D. Salton; Robert J. Ross; John D. Kelleher

This paper describes an experiment to evaluate the impact of idioms on Statistical Machine Translation (SMT) process using the language pair English/BrazilianPortuguese. Our results show that on sentences containing idioms a standard SMT system achieves about half the BLEU score of the same system when applied to sentences that do not contain idioms. We also provide a short error analysis and outline our planned work to overcome this limitation.


international conference on natural language generation | 2008

Referring expression generation challenge 2008 DIT system descriptions

John D. Kelleher; B. Mac Namee

This section describes the two systems developed at DIT for the attribute selection track of the REG 2008 challenge. Both of theses systems use an incremental greedy search to generate descriptions, similar to the incremental algorithm described in (Dale and Reiter, 1995). The output of these incremental algorithms are, to a large extent, determined by the order in which the algorithm tests the target objects attributes for inclusion in the description. Indeed, the major difference between the two systems described in this section is the mechanism used to order the attributes for inclusion.


international symposium on ambient intelligence | 2016

Using Topic Modelling Algorithms for Hierarchical Activity Discovery

Eoin Rogers; John D. Kelleher; Robert J. Ross

Activity discovery is the unsupervised process of discovering patterns in data produced from sensor networks that are monitoring the behaviour of human subjects. Improvements in activity discovery may simplify the training of activity recognition models by enabling the automated annotation of datasets and also the construction of systems that can detect and highlight deviations from normal behaviour. With this in mind, we propose an approach to activity discovery based on topic modelling techniques, and evaluate it on a dataset that mimics complex, interleaved sensor data in the real world. We also propose a means for discovering hierarchies of aggregated activities and discuss a mechanism for visualising the behaviour of such algorithms graphically.


Cognitive Processing | 2011

The effect of occlusion on the semantics of projective spatial terms: a case study in grounding language in perception

John D. Kelleher; Robert J. Ross; Colm Sloan; Brian Mac Namee

Although data-driven spatial template models provide a practical and cognitively motivated mechanism for characterizing spatial term meaning, the influence of perceptual rather than solely geometric and functional properties has yet to be systematically investigated. In the light of this, in this paper, we investigate the effects of the perceptual phenomenon of object occlusion on the semantics of projective terms. We did this by conducting a study to test whether object occlusion had a noticeable effect on the acceptance values assigned to projective terms with respect to a 2.5-dimensional visual stimulus. Based on the data collected, a regression model was constructed and presented. Subsequent analysis showed that the regression model that included the occlusion factor outperformed an adaptation of Regier & Carlson’s well-regarded AVS model for that same spatial configuration.


national conference on artificial intelligence | 2010

Visual Salience and Reference Resolution in Situated Dialogues: A Corpus-based Evaluation

Niels Schuette; John D. Kelleher; Brian Mac Namee

Dialogues between humans and robots are necessarily situated. Exophoric references to objects in the shared visual context are very frequent in situated dialogues, for example when a human is verbally guiding a tele-operated mobile robot. We present an approach to automatically resolving exophoric referring expressions in a situated dialogue based on the visual salience of possible referents. We evaluate the effectiveness of this approach and a range of different salience metrics using data from the SCARE corpus which we have augmented with visual information. The results of our evaluation show that our computationally lightweight approach is successful, and so promising for use in human-robot dialogue systems.

Collaboration


Dive into the John D. Kelleher's collaboration.

Top Co-Authors

Avatar

Brian Mac Namee

Dublin Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Robert J. Ross

Dublin Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Simon Dobnik

University of Gothenburg

View shared research outputs
Top Co-Authors

Avatar

Niels Schütte

Dublin Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Nick Hawes

University of Birmingham

View shared research outputs
Top Co-Authors

Avatar

Robert J. Ross

Dublin Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Colm Sloan

Dublin Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Fintan Costello

University College Dublin

View shared research outputs
Top Co-Authors

Avatar

Mark Dunne

Dublin Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Abhijit Mahalunkar

Dublin Institute of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge