Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Ramakrishnan Kannan is active.

Publication


Featured researches published by Ramakrishnan Kannan.


Knowledge and Information Systems | 2014

Bounded matrix factorization for recommender system

Ramakrishnan Kannan; Mariya Ishteva; Haesun Park

Matrix factorization has been widely utilized as a latent factor model for solving the recommender system problem using collaborative filtering. For a recommender system, all the ratings in the rating matrix are bounded within a pre-determined range. In this paper, we propose a new improved matrix factorization approach for such a rating matrix, called Bounded Matrix Factorization (BMF), which imposes a lower and an upper bound on every estimated missing element of the rating matrix. We present an efficient algorithm to solve BMF based on the block coordinate descent method. We show that our algorithm is scalable for large matrices with missing elements on multicore systems with low memory. We present substantial experimental results illustrating that the proposed method outperforms the state of the art algorithms for recommender system such as stochastic gradient descent, alternating least squares with regularization, SVD++ and Bias-SVD on real-world datasets such as Jester, Movielens, Book crossing, Online dating and Netflix.


acm sigplan symposium on principles and practice of parallel programming | 2016

A high-performance parallel algorithm for nonnegative matrix factorization

Ramakrishnan Kannan; Grey Ballard; Haesun Park

Non-negative matrix factorization (NMF) is the problem of determining two non-negative low rank factors W and H, for the given input matrix A, such that A ≈ WH. NMF is a useful tool for many applications in different domains such as topic modeling in text mining, background separation in video analysis, and community detection in social networks. Despite its popularity in the data mining community, there is a lack of efficient distributed algorithms to solve the problem for big data sets. We propose a high-performance distributed-memory parallel algorithm that computes the factorization by iteratively solving alternating non-negative least squares (NLS) subproblems for W and H. It maintains the data and factor matrices in memory (distributed across processors), uses MPI for interprocessor communication, and, in the dense case, provably minimizes communication costs (under mild assumptions). As opposed to previous implementations, our algorithm is also flexible: (1) it performs well for both dense and sparse matrices, and (2) it allows the user to choose any one of the multiple algorithms for solving the updates to low rank factors W and H within the alternating iterations. We demonstrate the scalability of our algorithm and compare it with baseline implementations, showing significant performance improvements.


international conference on data mining | 2012

Bounded Matrix Low Rank Approximation

Ramakrishnan Kannan; Mariya Ishteva; Haesun Park

Matrix lower rank approximations such as non-negative matrix factorization (NMF) have been successfully used to solve many data mining tasks. In this paper, we propose a new matrix lower rank approximation called Bounded Matrix Low Rank Approximation (BMA) which imposes a lower and an upper bound on every element of a lower rank matrix that best approximates a given matrix with missing elements. This new approximation models many real world problems, such as recommender systems, and performs better than other methods, such as singular value decompositions (SVD) or NMF. We present an efficient algorithm to solve BMA based on coordinate descent method. BMA is different from NMF as it imposes bounds on the approximation itself rather than on each of the low rank factors. We show that our algorithm is scalable for large matrices with missing elements on multi core systems with low memory. We present substantial experimental results illustrating that the proposed method outperforms the state of the art algorithms for recommender systems such as Stochastic Gradient Descent, Alternating Least Squares with regularization, SVD++, Bias-SVD on real world data sets such as Jester, Movie lens, Book crossing, Online dating and Netflix.


international conference on conceptual structures | 2014

A Dynamic Data Driven Application System for Vehicle Tracking

Richard M. Fujimoto; Angshuman Guin; Michael Hunter; Haesun Park; Gaurav Kanitkar; Ramakrishnan Kannan; Michael Milholen; Sabra Neal; Philip Pecher

Abstract Tracking the movement of vehicles in urban environments using fixed position sensors, mobile sensors, and crowd-sourced data is a challenging but important problem in applications such as law enforcement and defense. A dynamic data driven application system (DDDAS) is described to track a vehicles movements by repeatedly identifying the vehicle under investigation from live image and video data, predicting probable future locations, and repositioning sensors or retargeting requests for information in order to reacquire the vehicle. An overview of the envisioned system is described that includes image processing algorithms to detect and recapture the vehicle from live image data, a computational framework to predict probable vehicle locations at future points in time, and a power aware data distribution management system to disseminate data and requests for information over ad hoc wireless communication networks. A testbed under development in the midtown area of Atlanta, Georgia in the United States is briefly described.


visual analytics science and technology | 2014

VisIRR: Visual analytics for information retrieval and recommendation with large-scale document data

Jaegul Choo; Changhyun Lee; Hannah Kim; Hanseung Lee; Zhicheng Liu; Ramakrishnan Kannan; Charles D. Stolper; John T. Stasko; Barry L. Drake; Haesun Park

We present VisIRR, an interactive visual information retrieval and recommendation system for large-scale document data. Starting with a query, VisIRR visualizes the retrieved documents in a scatter plot along with their topic summary. Next, based on interactive personalized preference feedback on the documents, VisIRR collects and visualizes potentially relevant documents out of the entire corpus so that an integrated analysis of both retrieved and recommended documents can be performed seamlessly.


international conference on big data | 2016

Mini-apps for high performance data analysis

Sreenivas R. Sukumar; Michael A. Matheson; Ramakrishnan Kannan; Seung-Hwan Lim

Scaling-up scientific data analysis and machine learning algorithms for data-driven discovery is a grand challenge that we face today. Despite the growing need for analysis from science domains that are generating ‘Big Data’ from instruments and simulations, building high-performance analytical workflows of data-intensive algorithms have been daunting because: (i) the ‘Big Data’ hardware and software architecture landscape is constantly evolving, (ii) newer architectures impose new programming models, and (iii) data-parallel kernels of analysis algorithms and their performance facets on different architectures are poorly understood. To address these problems, we have: (i) identified scalable data-parallel kernels of popular data analysis algorithms, (ii) implemented ‘Mini-Apps’ of those kernels using different programming models (e.g. Map Reduce, MPI, etc.), (iii) benchmarked and validated the performance of the kernels in diverse architectures. In this paper, we discuss two of those Mini-Apps and show the execution of principal component analysis built as a workflow of the Mini-Apps. We show that Mini-Apps enable scientists to (i) write domain-specific data analysis code that scales on most HPC hardware and (ii) and offers the ability (most times with over a 10x speed-up) to analyze data sizes 100 times the size of what off-the-shelf desktop/workstations of today can handle.


parallel computing | 2015

Behavioral clusters in dynamic graphs

James P. Fairbanks; Ramakrishnan Kannan; Haesun Park; David A. Bader

Discover interesting patterns on longitudinal studies of vertex behavior over time.Apply machine learning to parallel graph analysis in a modular fashion.Finding behavior clusters of vertices in a dynamic graph.Apply NMF to group vertices that rise and fall in influence together. This paper contributes a method for combining sparse parallel graph algorithms with dense parallel linear algebra algorithms in order to understand dynamic graphs including the temporal behavior of vertices. Our method is the first to cluster vertices in a dynamic graph based on arbitrary temporal behaviors. In order to successfully implement this method, we develop a feature based pipeline for dynamic graphs and apply Nonnegative Matrix Factorization (NMF) to these features. We demonstrate these steps with a sample of the Twitter mentions graph as well as a CAIDA network traffic graph. We contribute and analyze a parallel NMF algorithm presenting both theoretical and empirical studies of performance. This work can be leveraged by graph/network analysts to understand the temporal behavior cluster structure and segmentation structure of dynamic graphs.


international conference on big data | 2016

Kernels for scalable data analysis in science: Towards an architecture-portable future

Sreenivas R. Sukumar; Ramakrishnan Kannan; Seung-Hwan Lim; Michael A. Matheson

In this paper, we pose and address some of the unique challenges in the analysis of scientific Big Data on supercomputing platforms. Our approach identifies, implements and scales numerical kernels that are critical to the instantiation of theory-inspired analytic workflows on modern computing architectures. We present the benefits of scalable kernels towards constructing algorithms such as principal component analysis and non-negative matrix factorization on an image-analysis use case at the Oak Ridge Leadership Computing Facility (OLCF). Based on experience with the use-case, we conclude that piecing scalable analytic kernels into user-defined analytic workflows are a flexible, modular and agile way to enable architecture-portable productivity for the data-intensive sciences.


2016 1st Joint International Workshop on Parallel Data Storage and data Intensive Scalable Computing Systems (PDSW-DISCS) | 2016

Fatman vs. littleboy: scaling up linear algebraic operations in scale-out data platforms

Luna Xu; Seung-Hwan Lim; Ali Raza Butt; Sreenivas R. Sukumar; Ramakrishnan Kannan

Linear algebraic operations such as matrix manipulations form the kernel of many machine learning and other crucial algorithms. Scaling up as well as scaling out such algorithms are highly desirable to enable efficient processing over millions of data points. To this end, we present a matrix manipulation approach to effectively scale-up each node in a scale-out data parallel platform such as Apache Spark. Specifically, we enable hardware acceleration for matrix multiplications in a distributed Spark setup without user intervention. Our approach supports both dense and sparse distributed matrices, and provides flexible control of acceleration by matrix density. We demonstrate the benefit of our approach for generalized matrix multiplication operations over large matrices with up to four billion elements. To connect the effectiveness of our approach with machine learning applications, we performed Gramian matrix computation via generalized matrix multiplications. Our experiments show that our approach achieves more than 2× performance speed-up, and up to 96.1% computation improvement, compared to a state of the art Spark MLlib for dense matrices.


international conference on parallel processing | 2018

Partitioning and Communication Strategies for Sparse Non-negative Matrix Factorization

Oguz Kaya; Ramakrishnan Kannan; Grey Ballard

Non-negative matrix factorization (NMF), the problem of finding two non-negative low-rank factors whose product approximates an input matrix, is a useful tool for many data mining and scientific applications such as topic modeling in text mining and unmixing in microscopy. In this paper, we focus on scaling algorithms for NMF to very large sparse datasets and massively parallel machines by employing effective algorithms, communication patterns, and partitioning schemes that leverage the sparsity of the input matrix. We consider two previous works developed for related problems, one that uses a fine-grained partitioning strategy using a point-to-point communication pattern and one that uses a Cartesian, or checkerboard, partitioning strategy using a collective-based communication pattern. We show that a combination of the previous approaches balances the demands of the various computations within NMF algorithms and achieves high efficiency and scalability. From the experiments, we see that our proposed strategy runs up to 10x faster than the state of the art on real-world datasets.

Collaboration


Dive into the Ramakrishnan Kannan's collaboration.

Top Co-Authors

Avatar

Haesun Park

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Grey Ballard

Sandia National Laboratories

View shared research outputs
Top Co-Authors

Avatar

Seung-Hwan Lim

Oak Ridge National Laboratory

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Barry L. Drake

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Changhyun Lee

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Charles D. Stolper

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Hanseung Lee

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

John T. Stasko

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Sreenivas R. Sukumar

Oak Ridge National Laboratory

View shared research outputs
Researchain Logo
Decentralizing Knowledge