Matthew Mulholland
Princeton University
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Matthew Mulholland.
meeting of the association for computational linguistics | 2014
Michael Heilman; Aoife Cahill; Nitin Madnani; Melissa Lopez; Matthew Mulholland; Joel R. Tetreault
Automated methods for identifying whether sentences are grammatical have various potential applications (e.g., machine translation, automated essay scoring, computer-assisted language learning). In this work, we construct a statistical model of grammaticality using various linguistic features (e.g., misspelling counts, parser outputs, n-gram language model scores). We also present a new publicly available dataset of learner sentences judged for grammaticality on an ordinal scale. In evaluations, we compare our system to the one from Post (2011) and find that our approach yields state-of-the-art performance.
international conference on multimodal interfaces | 2015
Chee Wee Leong; Lei Chen; Gary Feng; Chong Min Lee; Matthew Mulholland
Body language plays an important role in learning processes and communication. For example, communication research produced evidence that mathematical knowledge can be embodied in gestures made by teachers and students. Likewise, body postures and gestures are also utilized by speakers in oral presentations to convey ideas and important messages. Consequently, capturing and analyzing non-verbal behaviors is an important aspect in multimodal learning analytics (MLA) research. With regard to sensing capabilities, the introduction of depth sensors such as the Microsoft Kinect has greatly facilitated research and development in this area. However, the rapid advancement in hardware and software capabilities is not always in sync with the expanding set of features reported in the literature. For example, though Anvil is a widely used state-of-the-art annotation and visualization toolkit for motion traces, its motion recording component based on OpenNI is outdated. As part of our research in developing multimodal educational assessments, we began an effort to develop and standardize algorithms for purposes of multimodal feature extraction and creating automated scoring models. This paper provides an overview of relevant work in multimodal research on educational tasks, and proceeds to summarize our work using multimodal sensors in developing assessments of communication skills, with attention on the use of depth sensors. Specifically, we focus on the task of public speaking assessment using Microsoft Kinect. Additionally, we introduce an open-source Python package for computing expressive body language features from Kinect motion data, which we hope will benefit the MLA research community.
international conference on acoustics, speech, and signal processing | 2016
Matthew Mulholland; Melissa Lopez; Keelan Evanini; Anastassia Loukina; Yao Qian
In this paper, we compare ASR and human transcriptions of non-native speech to investigate to what extent the accuracy and the patterns of errors of a modern ASR system match those of human listeners in the context of automated assessment of L2 English language proficiency. We obtained multiple naïve transcriptions of short fragments of non-native spontaneous speech with different proficiency levels using crowdsourcing and matched these against the output of an ASR system. We compare WER and recall at the fragment level and consider human-ASR agreement at the word level. We find that we are able to attain a commensurate level of transcription quality using ASR, but the patterns of errors between the two groups differ at the word level.
spoken language technology workshop | 2016
Xinhao Wang; Keelan Evanini; James Bruno; Matthew Mulholland
This paper addresses the task of automatically detecting plagiarized responses in the context of a test of spoken English proficiency for non-native speakers. Text-to-text content similarity features are used jointly with speaking proficiency features extracted using an automated speech scoring system to train classifiers to distinguish between plagiarized and non-plagiarized spoken responses. A large data set drawn from an operational English proficiency assessment is used to simulate the performance of the detection system in a practical application. The best classifier on this heavily imbalanced data set resulted in an F1-score of 0.706 on the plagiarized class. These results indicate that the proposed system can potentially be used to improve the validity of both human and automated assessment of non-native spoken English.
Bergen Language and Linguistics Studies | 2015
Michael Flor; Yoko Futagi; Melissa Lopez; Matthew Mulholland
conference of the international speech communication association | 2018
Keelan Evanini; Matthew Mulholland; Rutuja Ubale; Yao Qian; Robert A. Pugh; Vikram Ramanarayanan; Aoife Cahill
ETS Research Report Series | 2018
Nitin Madnani; Aoife Cahill; Daniel Blanchard; Slava Andreyev; Diane Napolitano; Binod Gyawali; Michael Heilman; Chong Min Lee; Chee Wee Leong; Matthew Mulholland; Brian Riordan
symposium on languages, applications and technologies | 2017
Keelan Evanini; Matthew Mulholland; Eugene Tsuprun; Yao Qian
symposium on languages, applications and technologies | 2017
Xinhao Wang; Keelan Evanini; Klaus Zechner; Matthew Mulholland
conference of the international speech communication association | 2017
Su-Youn Yoon; Chong Min Lee; Ikkyu Choi; Xinhao Wang; Matthew Mulholland; Keelan Evanini