Proceedings of the National Academy of Sciences | 2019

Linguistic inferences without words

 
 
 

Abstract


Significance Linguistic meaning encompasses a rich typology of inferences, characterized by distinct patterns of interaction with logical expressions. For example, “Robin has continued to smoke” triggers the presuppositional inference that Robin smoked before, characterized by the preservation of the inference under negation in “Robin hasn’t continued to smoke.” We show experimentally that four main inference types can be robustly replicated with iconic gestures and visual animations. These nonlinguistic objects thus display the same type of logical behavior as spoken words. Because the gestures and animations were novel to the participants, the results suggest that people may productively divide new informational content among the components of the inferential typology using general algorithms that apply to linguistic and nonlinguistic objects alike. Contemporary semantics has uncovered a sophisticated typology of linguistic inferences, characterized by their conversational status and their behavior in complex sentences. This typology is usually thought to be specific to language and in part lexically encoded in the meanings of words. We argue that it is neither. Using a method involving “composite” utterances that include normal words alongside novel nonlinguistic iconic representations (gestures and animations), we observe successful “one-shot learning” of linguistic meanings, with four of the main inference types (implicatures, presuppositions, supplements, homogeneity) replicated with gestures and animations. The results suggest a deeper cognitive source for the inferential typology than usually thought: Domain-general cognitive algorithms productively divide both linguistic and nonlinguistic information along familiar parts of the linguistic typology.

Volume 116
Pages 9796 - 9801
DOI 10.1073/pnas.1821018116
Language English
Journal Proceedings of the National Academy of Sciences

Full Text