Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Borhan Samei is active.

Publication


Featured researches published by Borhan Samei.


intelligent tutoring systems | 2014

Context-Based Speech Act Classification in Intelligent Tutoring Systems

Borhan Samei; Haiying Li; Fazel Keshtkar; Vasile Rus; Arthur C. Graesser

In intelligent tutoring systems with natural language dialogue, speech act classification, the task of detecting learners’ intentions, informs the system’s response mechanism. In this paper, we propose supervised machine learning models for speech act classification in the context of an online collaborative learning game environment. We explore the role of context (i.e. speech acts of previous utterances) for speech act classification. We compare speech act classification models trained and tested with contextual and non-contextual features (contents of the current utterance). The accuracy of the proposed models is high. A surprising finding is the modest role of context in automatically predicting the speech acts.


international conference on multimodal interfaces | 2015

Multimodal Capture of Teacher-Student Interactions for Automated Dialogic Analysis in Live Classrooms

Sidney K. D'Mello; Andrew Olney; Nathaniel Blanchard; Borhan Samei; Xiaoyi Sun; Brooke Ward; Sean Kelly

We focus on data collection designs for the automated analysis of teacher-student interactions in live classrooms with the goal of identifying instructional activities (e.g., lecturing, discussion) and assessing the quality of dialogic instruction (e.g., analysis of questions). Our designs were motivated by multiple technical requirements and constraints. Most importantly, teachers could be individually micfied but their audio needed to be of excellent quality for automatic speech recognition (ASR) and spoken utterance segmentation. Individual students could not be micfied but classroom audio quality only needed to be sufficient to detect student spoken utterances. Visual information could only be recorded if students could not be identified. Design 1 used an omnidirectional laptop microphone to record both teacher and classroom audio and was quickly deemed unsuitable. In Designs 2 and 3, teachers wore a wireless Samson AirLine 77 vocal headset system, which is a unidirectional microphone with a cardioid pickup pattern. In Design 2, classroom audio was recorded with dual first- generation Microsoft Kinects placed at the front corners of the class. Design 3 used a Crown PZM-30D pressure zone microphone mounted on the blackboard to record classroom audio. Designs 2 and 3 were tested by recording audio in 38 live middle school classrooms from six U.S. schools while trained human coders simultaneously performed live coding of classroom discourse. Qualitative and quantitative analyses revealed that Design 3 was suitable for three of our core tasks: (1) ASR on teacher speech (word recognition rate of 66% and word overlap rate of 69% using Google Speech ASR engine); (2) teacher utterance segmentation (F-measure of 97%); and (3) student utterance segmentation (F-measure of 66%). Ideas to incorporate video and skeletal tracking with dual second-generation Kinects to produce Design 4 are discussed.


artificial intelligence in education | 2015

A Study of Automatic Speech Recognition in Noisy Classroom Environments for Automated Dialog Analysis

Nathaniel Blanchard; Michael Connolly Brady; Andrew Olney; Marci Glaus; Xiaoyi Sun; Martin Nystrand; Borhan Samei; Sean Kelly; Sidney D’Mello

The development of large-scale automatic classroom dialog analysis systems requires accurate speech-to-text translation. A variety of automatic speech recognition (ASR) engines were evaluated for this purpose. Recordings of teachers in noisy classrooms were used for testing. In comparing ASR results, Google Speech and Bing Speech were more accurate with word accuracy scores of 0.56 for Google and 0.52 for Bing compared to 0.41 for AT&T Watson, 0.08 for Microsoft, 0.14 for Sphinx with the HUB4 model, and 0.00 for Sphinx with the WSJ model. Further analysis revealed both Google and Bing engines were largely unaffected by speakers, speech class sessions, and speech characteristics. Bing results were validated across speakers in a laboratory study, and a method of improving Bing results is presented. Results provide a useful understanding of the capabilities of contemporary ASR engines in noisy classroom environments. Results also highlight a list of issues to be aware of when selecting an ASR engine for difficult speech recognition tasks.


annual meeting of the special interest group on discourse and dialogue | 2016

Identifying Teacher Questions Using Automatic Speech Recognition in Classrooms.

Nathaniel Blanchard; Patrick Donnelly; Andrew Olney; Borhan Samei; Brooke Ward; Xiaoyi Sun; Sean Kelly; Martin Nystrand; Sidney K. D'Mello

We investigate automatic question detection from recordings of teacher speech collected in live classrooms. Our corpus contains audio recordings of 37 class sessions taught by 11 teachers. We automatically segment teacher speech into utterances using an amplitude envelope thresholding approach followed by filtering non-speech via automatic speech recognition (ASR). We manually code the segmented utterances as containing a teacher question or not based on an empirically-validated scheme for coding classroom discourse. We compute domain-independent natural language processing (NLP) features from transcripts generated by three ASR engines (AT&T, Bing Speech, and Azure Speech). Our teacher-independent supervised machine learning model detects questions with an overall weighted F1 score of 0.59, a 51% improvement over chance. Furthermore, the proportion of automatically-detected questions per class session strongly correlates (Pearson’s r = 0.85) with human-coded question rates. We consider our results to reflect a substantial (37%) improvement over the state-of-the-art in automatic question detection from naturalistic audio. We conclude by discussing applications of our work for teachers, researchers, and other stakeholders.


north american chapter of the association for computational linguistics | 2016

Evaluation Dataset (DT-Grade) and Word Weighting Approach towards Constructed Short Answers Assessment in Tutorial Dialogue Context.

Rajendra Banjade; Nabin Maharjan; Nobal B. Niraula; Dipesh Gautam; Borhan Samei; Vasile Rus

Evaluating student answers often requires contextual information, such as previous utterances in conversational tutoring systems. For example, students use coreferences and write elliptical responses, i.e. incomplete but can be interpreted in context. The DT-Grade corpus which we present in this paper consists of short constructed answers extracted from tutorial dialogues between students and an Intelligent Tutoring System and annotated for their correctness in the given context and whether the contextual information was useful. The dataset contains 900 answers (of which about 25% required contextual information to properly interpret them). We also present a baseline system developed to predict the correctness label (such as correct, correct but incomplete) in which weights for the words are assigned based on context.


international conference on multimodal interfaces | 2016

Multi-sensor modeling of teacher instructional segments in live classrooms

Patrick Donnelly; Nathaniel Blanchard; Borhan Samei; Andrew Olney; Xiaoyi Sun; Brooke Ward; Sean Kelly; Martin Nystrand; Sidney K. D'Mello

We investigate multi-sensor modeling of teachers’ instructional segments (e.g., lecture, group work) from audio recordings collected in 56 classes from eight teachers across five middle schools. Our approach fuses two sensors: a unidirectional microphone for teacher audio and a pressure zone microphone for general classroom audio. We segment and analyze the audio streams with respect to discourse timing, linguistic, and paralinguistic features. We train supervised classifiers to identify the five instructional segments that collectively comprised a majority of the data, achieving teacher-independent F1 scores ranging from 0.49 to 0.60. With respect to individual segments, the individual sensor models and the fused model were on par for Question & Answer and Procedures & Directions segments. For Supervised Seatwork, Small Group Work, and Lecture segments, the classroom model outperformed both the teacher and fusion models. Across all segments, a multi-sensor approach led to an average 8% improvement over the state of the art approach that only analyzed teacher audio. We discuss implications of our findings for the emerging field of multimodal learning analytics.


Geophysical Journal International | 2016

Seismic features and automatic discrimination of deep and shallow induced-microearthquakes using neural network and logistic regression

S. Mostafa Mousavi; Stephen P. Horton; Charles A. Langston; Borhan Samei


educational data mining | 2014

Domain Independent Assessment of Dialogic Properties of Classroom Discourse

Borhan Samei; Andrew Olney; Sean Kelly; Martin Nystrand; Sidney K. D'Mello; Nathaniel Blanchard; Xiaoyi Sun; Marci Glaus; Arthur C. Graesser


international conference on user modeling adaptation and personalization | 2016

Automatic Teacher Modeling from Live Classroom Audio

Patrick Donnelly; Nathan Blanchard; Borhan Samei; Andrew Olney; Xiaoyi Sun; Brooke Ward; Sean Kelly; Martin Nystran; Sidney K. D'Mello


educational data mining | 2014

Building an Intelligent PAL from the Tutor.com Session Database Phase 1: Data Mining.

Donald M. Morrison; Benjamin D. Nye; Borhan Samei; Vivek Varma Datla; Craig Kelly; Vasile Rus

Collaboration


Dive into the Borhan Samei's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Xiaoyi Sun

University of Wisconsin-Madison

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Martin Nystrand

University of Wisconsin-Madison

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Brooke Ward

University of Wisconsin-Madison

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Sean Kelly

University of Notre Dame

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge