Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Afra Alishahi is active.

Publication


Featured researches published by Afra Alishahi.


Cognitive Science | 2010

A Probabilistic Computational Model of Cross-Situational Word Learning

Afsaneh Fazly; Afra Alishahi; Suzanne Stevenson

Words are the essence of communication: They are the building blocks of any language. Learning the meaning of words is thus one of the most important aspects of language acquisition: Children must first learn words before they can combine them into complex utterances. Many theories have been developed to explain the impressive efficiency of young children in acquiring the vocabulary of their language, as well as the developmental patterns observed in the course of lexical acquisition. A major source of disagreement among the different theories is whether children are equipped with special mechanisms and biases for word learning, or their general cognitive abilities are adequate for the task. We present a novel computational model of early word learning to shed light on the mechanisms that might be at work in this process. The model learns word meanings as probabilistic associations between words and semantic elements, using an incremental and probabilistic learning mechanism, and drawing only on general cognitive abilities. The results presented here demonstrate that much about word meanings can be learned from naturally occurring child-directed utterances (paired with meaning representations), without using any special biases or constraints, and without any explicit developmental changes in the underlying learning mechanism. Furthermore, our model provides explanations for the occasionally contradictory child experimental data, and offers predictions for the behavior of young word learners in novel situations.


Cognitive Science | 2008

A Computational Model of Early Argument Structure Acquisition

Afra Alishahi; Suzanne Stevenson

How children go about learning the general regularities that govern language, as well as keeping track of the exceptions to them, remains one of the challenging open questions in the cognitive science of language. Computational modeling is an important methodology in research aimed at addressing this issue. We must determine appropriate learning mechanisms that can grasp generalizations from examples of specific usages, and that exhibit patterns of behavior over the course of learning similar to those in children. Early learning of verb argument structure is an area of language acquisition that provides an interesting testbed for such approaches due to the complexity of verb usages. A range of linguistic factors interact in determining the felicitous use of a verb in various constructions-associations between syntactic forms and properties of meaning that form the basis for a number of linguistic and psycholinguistic theories of language. This article presents a computational model for the representation, acquisition, and use of verbs and constructions. The Bayesian framework is founded on a novel view of constructions as a probabilistic association between syntactic and semantic features. The computational experiments reported here demonstrate the feasibility of learning general constructions, and their exceptions, from individual usages of verbs. The behavior of the model over the timecourse of acquisition mimics, in relevant aspects, the stages of learning exhibited by children. Therefore, this proposal sheds light on the possible mechanisms at work in forming linguistic generalizations and maintaining knowledge of exceptions.


Computational Linguistics | 2017

Representation of linguistic form and function in recurrent neural networks

Ákos Kádár; Grzegorz Chrupała; Afra Alishahi

We present novel methods for analyzing the activation patterns of recurrent neural networks from a linguistic point of view and explore the types of linguistic structure they learn. As a case study, we use a standard standalone language model, and a multi-task gated recurrent network architecture consisting of two parallel pathways with shared word embeddings: The Visual pathway is trained on predicting the representations of the visual scene corresponding to an input sentence, and the Textual pathway is trained to predict the next word in the same sentence. We propose a method for estimating the amount of contribution of individual tokens in the input to the final prediction of the networks. Using this method, we show that the Visual pathway pays selective attention to lexical categories and grammatical functions that carry semantic information, and learns to treat word types differently depending on their grammatical function and their position in the sequential structure of the sentence. In contrast, the language models are comparatively more sensitive to words with a syntactic function. Further analysis of the most informative n-gram contexts for each model shows that in comparison with the Visual pathway, the language models react more strongly to abstract contexts that represent syntactic constructions.


Language and Cognitive Processes | 2010

A computational model of learning semantic roles from child-directed language

Afra Alishahi; Suzanne Stevenson

Semantic roles are a critical aspect of linguistic knowledge because they indicate the relations of the participants in an event to the main predicate. Experimental studies on children and adults show that both groups use associations between general semantic roles such as Agent and Theme, and grammatical positions such as Subject and Object, even in the absence of familiar verbs. Other studies suggest that semantic roles evolve over time, and might best be viewed as a collection of verb-based or general semantic properties. A usage-based account of language acquisition suggests that general roles and their association with grammatical positions can be learned from the data children are exposed to, through a process of generalisation and categorisation. In this paper, we propose a probabilistic usage-based model of semantic role learning. Our model can acquire associations between the semantic properties of the arguments of an event, and the syntactic positions that the arguments appear in. These probabilistic associations enable the model to learn general conceptions of roles, based only on exposure to individual verb usages, and without requiring explicit labelling of the roles in the input. The acquired role properties are a good intuitive match to the expected properties of various roles, and are useful in guiding comprehension in the model to the most likely interpretation in the face of ambiguity. The learned roles can also be used to select the correct meaning of a novel verb in an ambiguous situation.


meeting of the association for computational linguistics | 2017

Representations of language in a model of visually grounded speech signal

Grzegorz Chrupała; Lieke Gelderloos; Afra Alishahi

We present a visually grounded model of speech perception which projects spoken utterances and images to a joint semantic space. We use a multi-layer recurrent highway network to model the temporal nature of spoken speech, and show that it learns to extract both form and meaning-based linguistic knowledge from the input signal. We carry out an in-depth analysis of the representations used by different components of the trained model and show that encoding of semantic aspects tends to become richer as we go up the hierarchy of layers, whereas encoding of form-related aspects of the language input tends to initially increase and then plateau or decrease.


international joint conference on natural language processing | 2015

Learning language through pictures

Grzegorz Chrupała; Ákos Kádár; Afra Alishahi

We propose Imaginet, a model of learning visually grounded representations of language from coupled textual and visual input. The model consists of two Gated Recurrent Unit networks with shared word embeddings, and uses a multi-task objective by receiving a textual description of a scene and trying to concurrently predict its visual representation and the next word in the sentence. Like humans, it acquires meaning representations for individual words from descriptions of visual scenes. Moreover, it learns to effectively use sequential structure in semantic interpretation of multi-word phrases.


conference on computational natural language learning | 2008

Fast Mapping in Word Learning: What Probabilities Tell Us

Afra Alishahi; Afsaneh Fazly; Suzanne Stevenson

Children can determine the meaning of a new word from hearing it used in a familiar context---an ability often referred to as fast mapping. In this paper, we study fast mapping in the context of a general probabilistic model of word learning. We use our model to simulate fast mapping experiments on children, such as referent selection and retention. The word learning model can perform these tasks through an inductive interpretation of the acquired probabilities. Our results suggest that fast mapping occurs as a natural consequence of learning more words, and provides explanations for the (occasionally contradictory) child experimental data.


meeting of the association for computational linguistics | 2009

Computational Modeling of Human Language Acquisition

Afra Alishahi

The nature and amount of information needed for learning a natural language, and the underlying mechanisms involved in this process, are the subject of much debate: is it possible to learn a language from usage data only, or some sort of innate knowledge and/or bias is needed to boost the process? This is a topic of interest to (psycho)linguists who study human language acquisition, as well as computational linguists who develop the knowledge sources necessary for largescale natural language processing systems. Children are a source of inspiration for any such study of language learnability. They learn language with ease, and their acquired knowledge of language is flexible and robust.


Proceedings of the Workshop on Cognitive Aspects of Computational Language Acquisition | 2007

A Cognitive Model for the Representation and Acquisition of Verb Selectional Preferences

Afra Alishahi; Suzanne Stevenson

We present a cognitive model of inducing verb selectional preferences from individual verb usages. The selectional preferences for each verb argument are represented as a probability distribution over the set of semantic properties that the argument can possess---a semantic profile. The semantic profiles yield verb-specific conceptualizations of the arguments associated with a syntactic position. The proposed model can learn appropriate verb profiles from a small set of noisy training data, and can use them in simulating human plausibility judgments and analyzing implicit object alternation.


conference on computational natural language learning | 2017

Encoding of phonology in a recurrent neural model of grounded speech

Afra Alishahi; Marie Barking; Grzegorz Chrupała

We study the representation and encoding of phonemes in a recurrent neural network model of grounded speech. We use a model which processes images and their spoken descriptions, and projects the visual and auditory representations into the same semantic space. We perform a number of analyses on how information about individual phonemes is encoded in the MFCC features extracted from the speech signal, and the activations of the layers of the model. Via experiments with phoneme decoding and phoneme discrimination we show that phoneme representations are most salient in the lower layers of the model, where low-level signals are processed at a fine-grained level, although a large amount of phonological information is retain at the top recurrent layer. We further find out that the attention mechanism following the top recurrent layer significantly attenuates encoding of phonology and makes the utterance embeddings much more invariant to synonymy. Moreover, a hierarchical clustering of phoneme representations learned by the network shows an organizational structure of phonemes similar to those proposed in linguistics.

Collaboration


Dive into the Afra Alishahi's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Thierry Poibeau

École Normale Supérieure

View shared research outputs
Researchain Logo
Decentralizing Knowledge