Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Dominic Widdows is active.

Publication


Featured researches published by Dominic Widdows.


international conference on computational linguistics | 2002

A graph model for unsupervised lexical acquisition

Dominic Widdows; Beate Dorow

This paper presents an unsupervised method for assembling semantic knowledge from a part-of-speech tagged corpus using graph algorithms. The graph model is built by linking pairs of words which participate in particular syntactic relationships. We focus on the symmetric relationship between pairs of nouns which occur together in lists. An incremental cluster-building algorithm using this part of the graph achieves 82% accuracy at a lexical acquisition task, evaluated against WordNet classes. The model naturally realises domain and corpus specific ambiguities as distinct components in the graph surrounding an ambiguous word.


meeting of the association for computational linguistics | 2003

An Empirical Model of Multiword Expression Decomposability

Timothy Baldwin; Colin Bannard; Takaaki Tanaka; Dominic Widdows

This paper presents a construction-inspecific model of multiword expression decomposability based on latent semantic analysis. We use latent semantic analysis to determine the similarity between a multiword expression and its constituent words, and claim that higher similarities indicate greater decomposability. We test the model over English noun-noun compounds and verb-particles, and evaluate its correlation with similarities and hyponymy values in WordNet. Based on mean hyponymy over partitions of data ranked on similarity, we furnish evidence for the calculated similarities being correlated with the semantic relational content of WordNet.


Journal of Biomedical Informatics | 2009

Empirical Distributional Semantics: Methods and Biomedical Applications

Trevor Cohen; Dominic Widdows

Over the past 15 years, a range of methods have been developed that are able to learn human-like estimates of the semantic relatedness between terms from the way in which these terms are distributed in a corpus of unannotated natural language text. These methods have also been evaluated in a number of applications in the cognitive science, computational linguistics and the information retrieval literatures. In this paper, we review the available methodologies for derivation of semantic relatedness from free text, as well as their evaluation in a variety of biomedical and other applications. Recent methodological developments, and their applicability to several existing applications are also discussed.


north american chapter of the association for computational linguistics | 2003

Using LSA and noun coordination information to improve the precision and recall of automatic hyponymy extraction

Scott Cederberg; Dominic Widdows

In this paper we demonstrate methods of improving both the recall and the precision of automatic methods for extraction of hyponymy (IS_A) relations from free text. By applying latent semantic analysis (LSA) to filter extracted hyponymy relations we reduce the rate of error of our initial pattern-based hyponymy extraction by 30%, achieving precision of 58%. Applying a graph-based model of noun-noun similarity learned automatically from coordination patterns to previously extracted correct hyponymy relations, we achieve roughly a five-fold increase in the number of correct hyponymy relations extracted.


north american chapter of the association for computational linguistics | 2003

Unsupervised methods for developing taxonomies by combining syntactic and statistical information

Dominic Widdows

This paper describes an unsupervised algorithm for placing unknown words into a taxonomy and evaluates its accuracy on a large and varied sample of words. The algorithm works by first using a large corpus to find semantic neighbors of the unknown word, which we accomplish by combining latent semantic analysis with part-of-speech information. We then place the unknown word in the part of the taxonomy where these neighbors are most concentrated, using a class-labelling algorithm developed especially for this task. This method is used to reconstruct parts of the existing Word-Net database, obtaining results for common nouns, proper nouns and verbs. We evaluate the contribution made by part-of-speech tagging and show that automatic filtering using the class-labelling algorithm gives a fourfold improvement in accuracy.


Journal of Biomedical Informatics | 2010

Reflective Random Indexing and indirect inference: A scalable method for discovery of implicit connections

Trevor Cohen; Roger W. Schvaneveldt; Dominic Widdows

The discovery of implicit connections between terms that do not occur together in any scientific document underlies the model of literature-based knowledge discovery first proposed by Swanson. Corpus-derived statistical models of semantic distance such as Latent Semantic Analysis (LSA) have been evaluated previously as methods for the discovery of such implicit connections. However, LSA in particular is dependent on a computationally demanding method of dimension reduction as a means to obtain meaningful indirect inference, limiting its ability to scale to large text corpora. In this paper, we evaluate the ability of Random Indexing (RI), a scalable distributional model of word associations, to draw meaningful implicit relationships between terms in general and biomedical language. Proponents of this method have achieved comparable performance to LSA on several cognitive tasks while using a simpler and less computationally demanding method of dimension reduction than LSA employs. In this paper, we demonstrate that the original implementation of RI is ineffective at inferring meaningful indirect connections, and evaluate Reflective Random Indexing (RRI), an iterative variant of the method that is better able to perform indirect inference. RRI is shown to lead to more clearly related indirect connections and to outperform existing RI implementations in the prediction of future direct co-occurrence in the MEDLINE corpus.


conference of the european chapter of the association for computational linguistics | 2003

Discovering corpus-specific word senses

Beate Dorow; Dominic Widdows

This paper presents an unsupervised algorithm which automatically discovers word senses from text. The algorithm is based on a graph model representing words and relationships between them. Sense clusters are iteratively computed by clustering the local graph of similar words around an ambiguous word. Discrimination against previously extracted sense clusters enables us to discover new senses. We use the same data for both recognising and resolving ambiguity.


ieee international conference semantic computing | 2010

The Semantic Vectors Package: New Algorithms and Public Tools for Distributional Semantics

Dominic Widdows; Trevor Cohen

Distributional semantics is the branch of natural language processing that attempts to model the meanings of words, phrases and documents from the distribution and usage of words in a corpus of text. In the past three years, research in this area has been accelerated by the availability of the Semantic Vectors package, a stable, fast, scalable, and free software package for creating and exploring concepts in distributional models. This paper introduces the broad field of distributional semantics, the role of vector models within this field, and describes some of the results that have been made possible by the Semantic Vectors package. These applications of Semantic Vectors have so far included contributions to medical informatics and knowledge discovery, analysis of scientific articles, and even Biblical scholarship. Of particular interest is the recent emergence of models that take word order and other ordered structures into account, using permutation of coordinates to model directional relationships and semantic predicates.


arXiv: Quantum Physics | 2009

A quantum logic of down below

Peter D. Bruza; Dominic Widdows; John Woods

This chapter is offered as a contribution to the logic of down below. We attempt to demonstrate that the nature of human agency necessitates that there actually be such a logic. The ensuing sections develop the suggestion that cognition down below has a structure strikingly similar to the physical structure of quantum states. In its general form, this is not an idea that originates with the present authors. It is known that there exist mathematical models from the cognitive science of cognition down below that have certain formal similarities to quantum mechanics. We want to take this idea seriously. We will propose that the subspaces of von Neumann-Birkhoff lattices are too crisp for modelling requisite cognitive aspects in relation to subsymbolic logic. Instead, we adopt an approach which relies on projections into nonorthogonal density states. The projection operator is motivated from cues which probe human memory.


Journal of Biomedical Informatics | 2009

Methodological Review: Empirical distributional semantics: Methods and biomedical applications

Trevor Cohen; Dominic Widdows

Over the past 15 years, a range of methods have been developed that are able to learn human-like estimates of the semantic relatedness between terms from the way in which these terms are distributed in a corpus of unannotated natural language text. These methods have also been evaluated in a number of applications in the cognitive science, computational linguistics and the information retrieval literatures. In this paper, we review the available methodologies for derivation of semantic relatedness from free text, as well as their evaluation in a variety of biomedical and other applications. Recent methodological developments, and their applicability to several existing applications are also discussed.

Collaboration


Dive into the Dominic Widdows's collaboration.

Top Co-Authors

Avatar

Trevor Cohen

University of Texas Health Science Center at Houston

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Peter D. Bruza

Queensland University of Technology

View shared research outputs
Top Co-Authors

Avatar

Beate Dorow

University of Stuttgart

View shared research outputs
Top Co-Authors

Avatar

Beate Dorow

University of Stuttgart

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Thomas C. Rindflesch

National Institutes of Health

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Donald A. Sofge

United States Naval Research Laboratory

View shared research outputs
Researchain Logo
Decentralizing Knowledge