Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where E. Jason Riedy is active.

Publication


Featured researches published by E. Jason Riedy.


2012 IEEE Conference on High Performance Extreme Computing | 2012

STINGER: High performance data structure for streaming graphs

David Ediger; Robert McColl; E. Jason Riedy; David A. Bader

The current research focus on “big data” problems highlights the scale and complexity of analytics required and the high rate at which data may be changing. In this paper, we present our high performance, scalable and portable software, Spatio-Temporal Interaction Networks and Graphs Extensible Representation (STINGER), that includes a graph data structure that enables these applications. Key attributes of STINGER are fast insertions, deletions, and updates on semantic graphs with skewed degree distributions. We demonstrate a process of algorithmic and architectural optimizations that enable high performance on the Cray XMT family and Intel multicore servers. Our implementation of STINGER on the Cray XMT processes over 3 million updates per second on a scale-free graph with 537 million edges.


parallel processing and applied mathematics | 2011

Parallel community detection for massive graphs

E. Jason Riedy; Henning Meyerhenke; David Ediger; David A. Bader

Tackling the current volume of graph-structured data requires parallel tools. We extend our work on analyzing such massive graph data with the first massively parallel algorithm for community detection that scales to current data sizes, scaling to graphs of over 122 million vertices and nearly 2 billion edges in under 7300 seconds on a massively multithreaded Cray XMT. Our algorithm achieves moderate parallel scalability without sacrificing sequential operational complexity. Community detection partitions a graph into subgraphs more densely connected within the subgraph than to the rest of the graph. We take an agglomerative approach similar to Clauset, Newman, and Moores sequential algorithm, merging pairs of connected intermediate subgraphs to optimize different graph properties. Working in parallel opens new approaches to high performance. On smaller data sets, we find the outputs modularity compares well with the standard sequential algorithms.


ieee international symposium on parallel distributed processing workshops and phd forum | 2010

Massive streaming data analytics: A case study with clustering coefficients

David Ediger; Karl Jiang; E. Jason Riedy; David A. Bader

We present a new approach for parallel massive graph analysis of streaming, temporal data with a dynamic and extensible representation. Handling the constant stream of new data from health care, security, business, and social network applications requires new algorithms and data structures. We examine data structure and algorithm trade-offs that extract the parallelism necessary for high-performance updating analysis of massive graphs. Static analysis kernels often rely on storing input data in a specific structure. Maintaining these structures for each possible kernel with high data rates incurs a significant performance cost. A case study computing clustering coefficients on a general-purpose data structure demonstrates incremental updates can be more efficient than global recomputation. Within this kernel, we compare three methods for dynamically updating local clustering coefficients: a brute-force local recalculation, a sorting algorithm, and our new approximation method using a Bloom filter. On 32 processors of a Cray XMT with a synthetic scale-free graph of 224 ≈ 16 million vertices and 229 ≈ 537 million edges, the brute-force method processes a mean of over 50 000 updates per second and our Bloom filter approaches 200 000 updates per second.


international parallel and distributed processing symposium | 2012

Scalable Multi-threaded Community Detection in Social Networks

E. Jason Riedy; David A. Bader; Henning Meyerhenke

The volume of existing graph-structured data requires improved parallel tools and algorithms. Finding communities, smaller sub graphs densely connected within the sub graph than to the rest of the graph, plays a role both in developing new parallel algorithms as well as opening smaller portions of the data to current analysis tools. We improve performance of our parallel community detection algorithm by 20% on the massively multithreaded Cray XMT, evaluate its performance on the next-generation Cray XMT2, and extend its reach to Intel-based platforms with OpenMP. To our knowledge, not only is this the first massively parallel community detection algorithm but also the only such algorithm that achieves excellent performance and good parallel scalability across all these platforms. Our implementation analyzes a moderate sized graph with 105 million vertices and 3.3 billion edges in around 500 seconds on a four processor, 80-logical-core Intel-based system and 1100 seconds on a 64-processor Cray XMT2.


ieee international symposium on parallel & distributed processing, workshops and phd forum | 2011

Tracking Structure of Streaming Social Networks

David Ediger; E. Jason Riedy; David A. Bader; Henning Meyerhenke

Current online social networks are massive and still growing. For example, Face book has over 500 million active users sharing over 30 billion items per month. The scale within these data streams has outstripped traditional graph analysis methods. Real-time monitoring for anomalies may require dynamic analysis rather than repeated static analysis. The massive state behind multiple persistent queries requires shared data structures and flexible representations. We present a framework based on the STINGER data structure that can monitor a global property, connected components, on a graph of 16 million vertices at rates of up to 240,000 updates per second on 32 processors of a Cray XMT. For very large scale-free graphs, our implementation uses novel batching techniques that exploit the scale-free nature of the data and run over three times faster than prior methods. Our framework handles, for the first time, real-world data rates, opening the door to higher-level analytics such as community and anomaly detection.


IEEE Transactions on Parallel and Distributed Systems | 2013

GraphCT: Multithreaded Algorithms for Massive Graph Analysis

David Ediger; Karl Jiang; E. Jason Riedy; David A. Bader

The digital world has given rise to massive quantities of data that include rich semantic and complex networks. A social graph, for example, containing hundreds of millions of actors and tens of billions of relationships is not uncommon. Analyzing these large data sets, even to answer simple analytic queries, often pushes the limits of algorithms and machine architectures. We present GraphCT, a scalable framework for graph analysis using parallel and multithreaded algorithms on shared memory platforms. Utilizing the unique characteristics of the Cray XMT, GraphCT enables fast network analysis at unprecedented scales on a variety of input data sets. On a synthetic power law graph with 2 billion vertices and 17 billion edges, we can find the connected components in 2 minutes. We can estimate the betweenness centrality of a similar graph with 537 million vertices and over 8 billion edges in under 1 hour. GraphCT is built for portability and performance.


international conference on acoustics, speech, and signal processing | 2012

Analysis of streaming social networks and graphs on multicore architectures

E. Jason Riedy; Henning Meyerhenke; David A. Bader; David Ediger; Timothy G. Mattson

Analyzing static snapshots of massive, graph-structured data cannot keep pace with the growth of social networks, financial transactions, and other valuable data sources. We introduce a framework, STING (Spatio-Temporal Interaction Networks and Graphs), and evaluate its performance on multicore, multisocket Intel®-based platforms. STING achieves rates of around 100 000 edge updates per second on large, dynamic graphs with a single, general data structure. We achieve speedups of up to 1000× over parallel static computation, improve monitoring a dynamic graphs connected components, and show an exact algorithm for maintaining local clustering coefficients performs better on Intel-based platforms than our earlier approximate algorithm.


ieee high performance extreme computing conference | 2014

Optimizing energy consumption and parallel performance for static and dynamic betweenness centrality using GPUs

Adam McLaughlin; E. Jason Riedy; David A. Bader

Applications of high-performance graph analysis range from computational biology to network security and even transportation. These applications often consider graphs under rapid change and are moving beyond HPC platforms into energy-constrained embedded systems. This paper optimizes one successful and demanding analysis kernel, betweenness centrality, for NVIDIA GPU accelerators in both environments. Our algorithm for static analysis is capable of exceeding 2 million traversed edges per second per watt (MTEPS/W). Optimizing the parallel algorithm and treating the dynamic problem directly achieves a 6.9× average speed-up and 83% average reduction in energy consumption.


international parallel and distributed processing symposium | 2016

Updating PageRank for Streaming Graphs

E. Jason Riedy

Incremental graph algorithms can respond quickly to small changes in massive graphs by updating rather than recomputing analysis metrics. Here we use the linear system formulation of PageRank and ideas from iterative refinement to compute the update to a PageRank vector accurately and quickly. The core idea is to express the residual of the original solution with respect to the updated matrix representing the graph. The update to the residual is sparse. Solving for the solution update with a straight-forward iterative method spreads the change outward from the change locations but converges before traversing the entire graph. We achieve speed-ups of 2x to over 40x relative to a restarted, highly parallel PageRank iteration for small, low-latency batches of edge insertions. These cases traverse 2x to nearly 10,000x fewer edges than the restarted PageRank iteration. This provides an interesting test case for the ongoing GraphBLAS effort: Can the APIs support our incremental algorithms cleanly and efficiently?


ACM Crossroads Student Magazine | 2013

Massive streaming data analytics: a graph-based approach

E. Jason Riedy; David A. Bader

Analyzing massive streaming graphs efficiently requires new algorithms, data structures, and computing platforms.

Collaboration


Dive into the E. Jason Riedy's collaboration.

Top Co-Authors

Avatar

David A. Bader

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

David Ediger

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Henning Meyerhenke

Karlsruhe Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Adam McLaughlin

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Anita Zakrzewska

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Eisha Nathan

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Karl Jiang

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Richard W. Vuduc

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Christine Klymko

Lawrence Livermore National Laboratory

View shared research outputs
Researchain Logo
Decentralizing Knowledge