Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Prem Natarajan is active.

Publication


Featured researches published by Prem Natarajan.


computer vision and pattern recognition | 2016

Pose-Aware Face Recognition in the Wild

Iacopo Masi; Stephen Rawls; Gérard G. Medioni; Prem Natarajan

We propose a method to push the frontiers of unconstrained face recognition in the wild, focusing on the problem of extreme pose variations. As opposed to current techniques which either expect a single model to learn pose invariance through massive amounts of training data, or which normalize images to a single frontal pose, our method explicitly tackles pose variation by using multiple pose specific models and rendered face images. We leverage deep Convolutional Neural Networks (CNNs) to learn discriminative representations we call Pose-Aware Models (PAMs) using 500K images from the CASIA WebFace dataset. We present a comparative evaluation on the new IARPA Janus Benchmark A (IJB-A) and PIPA datasets. On these datasets PAMs achieve remarkably better performance than commercial products and surprisingly also outperform methods that are specifically fine-tuned on the target dataset.


SACH'06 Proceedings of the 2006 conference on Arabic and Chinese handwriting recognition | 2006

Multi-lingual offline handwriting recognition using hidden Markov models: a script-independent approach

Prem Natarajan; Shirin Saleem; Rohit Prasad; Ehry MacRostie; Krishna Subramanian

This paper introduces a script-independent methodology for multilingual offline handwriting recognition (OHR) based on the use of Hidden Markov Models (HMM). The OHR methodology extends our script-independent approach for OCR of machine-printed text images. The feature extraction, training, and recognition components of the system are all designed to be script independent. The HMM training and recognition components are based on our Byblos continuous speech recognition system. The HMM parameters are estimated automatically from the training data, without the need for laborious hand-written rules. The system does not require pre-segmentation of the data, neither at the word level nor at the character level. Thus, the system can handle languages with cursive handwritten scripts in a straightforward manner. The script independence of the system is demonstrated with experimental results in three scripts that exhibit significant differences in glyph characteristics: English, Chinese, and Arabic. Results from an initial set of experiments are presented to demonstrate the viability of the proposed methodology.


workshop on applications of computer vision | 2016

Face recognition using deep multi-pose representations

Yue Wu; Stephen Rawls; Shai Harel; Tal Hassner; Iacopo Masi; Jongmoo Choi; Jatuporn Lekust; Jungyeon Kim; Prem Natarajan; Ram Nevatia; Gérard G. Medioni

We introduce our method and system for face recognition using multiple pose-aware deep learning models. In our representation, a face image is processed by several pose-specific deep convolutional neural network (CNN) models to generate multiple pose-specific features. 3D rendering is used to generate multiple face poses from the input image. Sensitivity of the recognition system to pose variations is reduced since we use an ensemble of pose-specific CNN features. The paper presents extensive experimental results on the effect of landmark detection, CNN layer selection and pose model selection on the performance of the recognition pipeline. Our novel representation achieves better results than the state-of-the-art on IARPAs CS2 and NISTs IJB-A in both verification and identification (i.e. search) tasks.


international conference on image processing | 2011

Automated image quality assessment for camera-captured OCR

Xujun Peng; Huaigu Cao; Krishna Subramanian; Rohit Prasad; Prem Natarajan

Camera-captured optical character recognition (OCR) is a challenging area because of artifacts introduced during image acquisition with consumer-domain hand-held and Smart phone cameras. Critical information is lost if the user does not get immediate feedback on whether the acquired image meets the quality requirements for OCR. To avoid such information loss, we propose a novel automated image quality assessment method that predicts the degree of degradation on OCR. Unlike other image quality assessment algorithms which only deal with blurring, the proposed method quantifies image quality degradation across several artifacts and accurately predicts the impact on OCR error rate. We present evaluation results on a set of machine-printed document images which have been captured using digital cameras with different degradations.


international semantic web conference | 2015

Building and Using a Knowledge Graph to Combat Human Trafficking

Pedro A. Szekely; Craig A. Knoblock; Jason Slepicka; Andrew Philpot; Amandeep Singh; Chengye Yin; Dipsy Kapoor; Prem Natarajan; Daniel Marcu; Kevin Knight; David Stallard; Subessware S. Karunamoorthy; Rajagopal Bojanapalli; Steven Minton; Brian Amanatullah; Todd Hughes; Mike Tamayo; David Flynt; Rachel Artiss; Shih-Fu Chang; Tao Chen; Gerald Hiebel; Lidia Ferreira

There is a huge amount of data spread across the web and stored in databases that we can use to build knowledge graphs. However, exploiting this data to build knowledge graphs is difficult due to the heterogeneity of the sources, scale of the amount of data, and noise in the data. In this paper we present an approach to building knowledge graphs by exploiting semantic technologies to reconcile the data continuously crawled from diverse sources, to scale to billions of triples extracted from the crawled content, and to support interactive queries on the data. We applied our approach, implemented in the DIG system, to the problem of combating human trafficking and deployed it to six law enforcement agencies and several non-governmental organizations to assist them with finding traffickers and helping victims.


Computer Speech & Language | 2013

Batch-mode semi-supervised active learning for statistical machine translation

Sankaranarayanan Ananthakrishnan; Rohit Prasad; David Stallard; Prem Natarajan

The development of high-performance statistical machine translation (SMT) systems is contingent on the availability of substantial, in-domain parallel training corpora. The latter, however, are expensive to produce due to the labor-intensive nature of manual translation. We propose to alleviate this problem with a novel, semi-supervised, batch-mode active learning strategy that attempts to maximize in-domain coverage by selecting sentences, which represent a balance between domain match, translation difficulty, and batch diversity. Simulation experiments on an English-to-Pashto translation task show that the proposed strategy not only outperforms the random selection baseline, but also traditional active selection techniques based on dissimilarity to existing training data.


Proceedings of the International Workshop on Multilingual OCR | 2009

A stroke regeneration method for cleaning rule-lines in handwritten document images

Huaigu Cao; Rohit Prasad; Prem Natarajan

We describe a rule-line removal algorithm for handwritten document images in this paper. Compared to the existing approaches, our algorithm obtains more scalability to higher-resolution images and thicker rule-lines. Derived from the simple gap-filling methods using line-drawing algorithms, we present a novel approach to regenerating the missing portions of text strokes. Using this approach, the deformed text can be restored to its original shape. We also explore the noise filtering method for binarized document images, in particular by choosing the morphological operator in accordance with the noise power of the input image. Our approach has proven to be effective by experiments on both real and synthetic handwritten document images.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2017

Facial Landmark Detection with Tweaked Convolutional Neural Networks

Yue Wu; Tal Hassner; KangGeon Kim; Gérard G. Medioni; Prem Natarajan

This paper concerns the problem of facial landmark detection. We provide a unique new analysis of the features produced at intermediate layers of a convolutional neural network (CNN) trained to regress facial landmark coordinates. This analysis shows that while being processed by the CNN, face images can be partitioned in an unsupervised manner into subsets containing faces in similar poses (i.e., 3D views) and facial properties (e.g., presence or absence of eye-wear). Based on this finding, we describe a novel CNN architecture, specialized to regress the facial landmark coordinates of faces in specific poses and appearances. To address the shortage of training data, particularly in extreme profile poses, we additionally present data augmentation techniques designed to provide sufficient training examples for each of these specialized sub-networks. The proposed Tweaked CNN (TCNN) architecture is shown to outperform existing landmark detection methods in an extensive battery of tests on the AFW, ALFW, and 300W benchmarks. Finally, to promote reproducibility of our results, we make code and trained models publicly available through our project webpage.


international conference on document analysis and recognition | 2011

Graph Clustering-Based Ensemble Method for Handwritten Text Line Segmentation

Vasant Manohar; Shiv Naga Prasad Vitaladevuni; Huaigu Cao; Rohit Prasad; Prem Natarajan

Handwritten text line segmentation on real-world data presents significant challenges that cannot be overcome by any single technique. Given the diversity of approaches and the recent advances in ensemble-based combination for pattern recognition problems, it is possible to improve the segmentation performance by combining the outputs from different line finding methods. In this paper, we propose a novel graph clustering-based approach to combine the output of an ensemble of text line segmentation algorithms. A weighted undirected graph is constructed with nodes corresponding to connected components and edge connecting pairs of connected components. Text line segmentation is then posed as the problem of minimum cost partitioning of the nodes in the graph such that each cluster corresponds to a unique line in the document image. Experimental results on a challenging Arabic field dataset using the ensemble method shows a relative gain of 18% in the F1 score over the best individual method within the ensemble.


international conference on document analysis and recognition | 2011

Handwritten and Typewritten Text Identification and Recognition Using Hidden Markov Models

Huaigu Cao; Rohit Prasad; Prem Natarajan

We present a system for identification and recognition of handwritten and typewritten text from document images using hidden Markov models (HMMs) in this paper. Our text type identification uses OCR decoding to generate word boundaries followed by word-level handwritten/typewritten identification using HMMs. We show that the contextual constraints from the HMM significantly improves the identification performance over the conventional Gaussian mixture model (GMM)-based method. Type identification is then used to estimate the frame sample rates and frame width of feature sequences for HMM OCR system for each type independently. This type-dependent approach to computing the frame sample rate and frame width shows significant improvement in OCR accuracy over type-independent approaches.

Collaboration


Dive into the Prem Natarajan's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Stephen Rawls

University of Southern California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Yue Wu

University of Southern California

View shared research outputs
Researchain Logo
Decentralizing Knowledge