Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Rajendra Banjade is active.

Publication


Featured researches published by Rajendra Banjade.


international conference on computational linguistics | 2013

Similarity measures based on latent dirichlet allocation

Vasile Rus; Nobal B. Niraula; Rajendra Banjade

We present in this paper the results of our investigation on semantic similarity measures at word- and sentence-level based on two fully-automated approaches to deriving meaning from large corpora: Latent Dirichlet Allocation, a probabilistic approach, and Latent Semantic Analysis, an algebraic approach. The focus is on similarity measures based on Latent Dirichlet Allocation, due to its novelty aspects, while the Latent Semantic Analysis measures are used for comparison purposes. We explore two types of measures based on Latent Dirichlet Allocation: measures based on distances between probability distribution that can be applied directly to larger texts such as sentences and a word-to-word similarity measure that is then expanded to work at sentence-level. We present results using paraphrase identification data in the Microsoft Research Paraphrase corpus.


north american chapter of the association for computational linguistics | 2015

NeRoSim: A System for Measuring and Interpreting Semantic Textual Similarity

Rajendra Banjade; Nobal B. Niraula; Nabin Maharjan; Vasile Rus; Dan Stefanescu; Mihai C. Lintean; Dipesh Gautam

We present in this paper our system developed for SemEval 2015 Shared Task 2 (2a - English Semantic Textual Similarity, STS, and 2c - Interpretable Similarity) and the results of the submitted runs. For the English STS subtask, we used regression models combining a wide array of features including semantic similarity scores obtained from various methods. One of our runs achieved weighted mean correlation score of 0.784 for sentence similarity subtask (i.e., English STS) and was ranked tenth among 74 runs submitted by 29 teams. For the interpretable similarity pilot task, we employed a rule-based approach blended with chunk alignment labeling and scoring based on semantic similarity features. Our system for interpretable text similarity was among the top three best performing systems.


conference on intelligent text processing and computational linguistics | 2015

Lemon and Tea Are Not Similar: Measuring Word-to-Word Similarity by Combining Different Methods

Rajendra Banjade; Nabin Maharjan; Nobal B. Niraula; Vasile Rus; Dipesh Gautam

Substantial amount of work has been done on measuring word-to-word relatedness which is also commonly referred as similarity. Though relatedness and similarity are closely related, they are not the same as illustrated by the words lemon and tea which are related but not similar. The relatedness takes into account a broader ranLemge of relations while similarity only considers subsumption relations to assess how two objects are similar. We present in this paper a method for measuring the semantic similarity of words as a combination of various techniques including knowledge-based and corpus-based methods that capture different aspects of similarity. Our corpus based method exploits state-of-the-art word representations. We performed experiments with a recently published significantly large dataset called Simlex-999 and achieved a significantly better correlation (ρ = 0.642, P < 0.001) with human judgment compared to the individual performance.


international conference on computational linguistics | 2014

A Sentence Similarity Method Based on Chunking and Information Content

Dan Ştefănescu; Rajendra Banjade; Vasile Rus

This paper introduces a method for assessing the semantic similarity between sentences, which relies on the assumption that the meaning of a sentence is captured by its syntactic constituents and the dependencies between them. We obtain both the constituents and their dependencies from a syntactic parser. Our algorithm considers that two sentences have the same meaning if it can find a good mapping between their chunks and also if the chunk dependencies in one text are preserved in the other. Moreover, the algorithm takes into account that every chunk has a different importance with respect to the overall meaning of a sentence, which is computed based on the information content of the words in the chunk. The experiments conducted on a well-known paraphrase data set show that the performance of our method is comparable to state of the art.


SLSP'13 Proceedings of the First international conference on Statistical Language and Speech Processing | 2013

Experiments with semantic similarity measures based on LDA and LSA

Nobal B. Niraula; Rajendra Banjade; Dan Ştefănescu; Vasile Rus

We present in this paper experiments with several semantic similarity measures based on the unsupervised method Latent Dirichlet Allocation. For comparison purposes, we also report experimental results using an algebraic method, Latent Semantic Analysis. The proposed semantic similarity methods were evaluated using one dataset that includes student answers from conversational intelligent tutoring systems and a standard paraphrase dataset, the Microsoft Research Paraphrase corpus. Results indicate that the method based on word representations as topic vectors outperforms methods based on distributions over topics and words. The proposed evaluation methods can also be regarded as an extrinsic method for evaluating topic coherence or selecting the number of topics in LDA models, i.e. a task-based evaluation of topic coherence and selection of number of topics in LDA.


Archive | 2017

Dialogue Act Classification In Human-to-Human Tutorial Dialogues

Vasile Rus; Nabin Maharjan; Rajendra Banjade

We present in this paper preliminary results with dialogue act classification in human-to-human tutorial dialogues. Dialogue acts are ways to characterize the actions of tutors and students based on the language-as-action theory. This work serves our larger goal of identifying patterns of tutors’ actions, in the form of dialogue acts, that relate to learning. The preliminary results we obtained for dialogue act classification using a machine learning approach are promising.


Archive | 2017

A Study On Two Hint-level Policies in Conversational Intelligent Tutoring Systems

Vasile Rus; Rajendra Banjade; Nobal B. Niraula; Elizabeth Gire; Donald R. Franceschetti

In this work, we compared two hint-level instructional strategies, minimum scaffolding vs. maximum scaffolding, in the context of conversational intelligent tutoring systems (ITSs). The two strategies are called policies because they have a clear bias, as detailed in the paper. To this end, we conducted a randomized controlled trial experiment with two conditions corresponding to two versions of the same underlying state-of-the-art conversational ITS, i.e. DeepTutor. Each version implemented one of the two hint-level strategies. Experimental data analysis revealed that pre-post learning gains were significant in both conditions. We also learned that, in general, students need more than just a minimally informative hint in order to infer the next steps in the solution to a challenging problem; this is the case in the context of a problem selection strategy that picks challenging problems for students to work on.


north american chapter of the association for computational linguistics | 2016

DTSim at SemEval-2016 Task 2: Interpreting Similarity of Texts Based on Automated Chunking, Chunk Alignment and Semantic Relation Prediction.

Rajendra Banjade; Nabin Maharjan; Nobal B. Niraula; Vasile Rus

In this paper we describe our system (DTSim) submitted at SemEval-2016 Task 2: Interpretable Semantic Textual Similarity (iSTS). We participated in both gold chunks category (texts chunked by human experts and provided by the task organizers) and system chunks category (participants had to automatically chunk the input texts). We developed a Conditional Random Fields based chunker and applied rules blended with semantic similarity methods in order to predict chunk alignments, alignment types and similarity scores. Our system obtained F1 score up to 0.648 in predicting the chunk alignment types and scores together and was one of the top performing systems overall.


north american chapter of the association for computational linguistics | 2016

DTSim at SemEval-2016 Task 1: Semantic Similarity Model Including Multi-Level Alignment and Vector-Based Compositional Semantics

Rajendra Banjade; Nabin Maharjan; Dipesh Gautam; Vasile Rus

In this paper we describe our system (DTSim) submitted at SemEval-2016 Task 1: Semantic Textual Similarity (STS Core). We developed Support Vector Regression model with various features including the similarity scores calculated using alignment based methods and semantic composition based methods. The correlations between our system output and the human ratings were above 0.8 in three datasets.


north american chapter of the association for computational linguistics | 2016

Evaluation Dataset (DT-Grade) and Word Weighting Approach towards Constructed Short Answers Assessment in Tutorial Dialogue Context.

Rajendra Banjade; Nabin Maharjan; Nobal B. Niraula; Dipesh Gautam; Borhan Samei; Vasile Rus

Evaluating student answers often requires contextual information, such as previous utterances in conversational tutoring systems. For example, students use coreferences and write elliptical responses, i.e. incomplete but can be interpreted in context. The DT-Grade corpus which we present in this paper consists of short constructed answers extracted from tutorial dialogues between students and an Intelligent Tutoring System and annotated for their correctness in the given context and whether the contextual information was useful. The dataset contains 900 answers (of which about 25% required contextual information to properly interpret them). We also present a baseline system developed to predict the correctness label (such as correct, correct but incomplete) in which weights for the words are assigned based on context.

Collaboration


Dive into the Rajendra Banjade's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge