Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Eric Goodman is active.

Publication


Featured researches published by Eric Goodman.


international joint conference on neural network | 2006

Spatiotemporal Pattern Recognition via Liquid State Machines

Eric Goodman; Dan Ventura

The applicability of complex networks of spiking neurons as a general purpose machine learning technique remains open. Building on previous work using macroscopic exploration of the parameter space of an (artificial) neural microcircuit, we investigate the possibility of using a liquid state machine to solve two real-world problems: stockpile surveillance signal alignment and spoken phoneme recognition.


extended semantic web conference | 2011

High-performance computing applied to semantic databases

Eric Goodman; Edward Steven Jimenez; David Mizell; Sinan Al-Saffar; Bob Adolf; David J. Haglin

To-date, the application of high-performance computing resources to Semantic Web data has largely focused on commodity hardware and distributed memory platforms. In this paper we make the case that more specialized hardware can offer superior scaling and close to an order of magnitude improvement in performance. In particular we examine the Cray XMT. Its key characteristics, a large, global sharedmemory, and processors with a memory-latency tolerant design, offer an environment conducive to programming for the Semantic Web and have engendered results that far surpass current state of the art. We examine three fundamental pieces requisite for a fully functioning semantic database: dictionary encoding, RDFS inference, and query processing. We show scaling up to 512 processors (the largest configuration we had available), and the ability to process 20 billion triples completely in-memory.


ieee international symposium on parallel distributed processing workshops and phd forum | 2010

Hashing strategies for the Cray XMT

Eric Goodman; David J. Haglin; Chad Scherrer; Daniel G. Chavarría-Miranda; Jace A. Mogill; John Feo

Two of the most commonly used hashing strategies-linear probing and hashing with chaining-are adapted for efficient execution on a Cray XMT. These strategies are designed to minimize memory contention. Datasets that follow a power law distribution cause significant performance challenges to shared memory parallel hashing implementations. Experimental results show good scalability up to 128 processors on two power law datasets with different data types: integer and string. These implementations can be used in a wide range of applications.


international symposium on neural networks | 2005

Effectively using recurrently-connected spiking neural networks

Eric Goodman; Dan Ventura

Recurrently connected spiking neural networks are difficult to use and understand because of the complex nonlinear dynamics of the system. Through empirical studies of spiking networks, we deduce several principles which are critical to success. Network parameters such as synaptic time delays and time constants and the connection probabilities can be adjusted to have a significant impact on accuracy. We show how to adjust these parameters to fit the type of problem.


ieee international conference on high performance computing data and analytics | 2011

Scalable hashing for shared memory supercomputers

Eric Goodman; M. Nicole Lemaster; Edward Steven Jimenez

Hashing is a fundamental technique in computer science to allow O(l) insert and lookups of items in an associative array. Here we present several thread coordination and hashing strategies and compare and contrast their performance on large, shared memory symmetric multiprocessor machines, each possessing between a half to a full terabyte of memory. We show how our approach can be used as a key kernel for fundamental paradigms such as dynamic programming and MapReduce. We further show that a set of approaches yields close to linear speedup for both uniform random and more difficult power law distributions. This scalable performance is in spite of the fact that our set of approaches is not completely lock-free. Our experimental results utilize and compare an SGI Altix UV with 4 Xeon processors (32 cores) and a Cray XMT with 128 processors. On the scale of data we addressed, on the order of 5 billion integers, we show that the Altix UV far exceeds the performance of the Cray XMT for power law distributions. However, the Cray XMT exhibits greater scalability.


international conference on machine learning and applications | 2015

Using Bipartite Anomaly Features for Cyber Security Applications

Eric Goodman; Joe Ingram; Shawn Martin; Dirk Grunwald

In this paper we use anomaly scores derived from a technique for bipartite graphs as features for a supervised machine learning algorithm for two cyber security problems: classifying Short Message Service (SMS) text messages as either spam or non-spam and detecting malicious lateral movement within a network. While disparate problems, both spam and lateral movement detection can be viewed as bipartite graphs and we can compute bipartite anomaly scores for each situation. The bipartite anomaly scores by themselves are not very predictive, but used as auxiliary features can boost the receiver operating characteristic (ROC) curve of a supervised classifier. We examine the UCI SMS Spam Collection Data Set for the SPAM problem and use an authentication graph from Los Alamos National Laboratory. We create features by dimensionality reduction through principal component analysis (PCA) on the message-term or user-computer matrix, and then augment those features with anomaly scores. By using the anomaly scores we are able to improve the area under the curve (AUC) for the receiver operating characteristic (ROC) up to 27.5% for the spam data and 21.4% for the authentication data.


Proceedings of SPIE | 2014

Irregular large-scale computed tomography on multiple graphics processors improves energy-efficiency metrics for industrial applications

Edward Steven Jimenez; Eric Goodman; Ryeojin Park; Laurel Jeffers Orr; Kyle R. Thompson

This paper will investigate energy-efficiency for various real-world industrial computed-tomography reconstruction algorithms, both CPU- and GPU-based implementations. This work shows that the energy required for a given reconstruction is based on performance and problem size. There are many ways to describe performance and energy efficiency, thus this work will investigate multiple metrics including performance-per-watt, energy-delay product, and energy consumption. This work found that irregular GPU-based approaches1 realized tremendous savings in energy consumption when compared to CPU implementations while also significantly improving the performance-per- watt and energy-delay product metrics. Additional energy savings and other metric improvement was realized on the GPU-based reconstructions by improving storage I/O by implementing a parallel MIMD-like modularization of the compute and I/O tasks.


international conference on big data | 2014

Optimizing graph queries with graph joins and Sprinkle SPARQL

Eric Goodman; Edward Steven Jimenez; Cliff Joslyn; David J. Haglin; Sinan Al-Saffar; Dirk Grunwald

Big data problems are often more akin to sparse graphs rather than relational tables. As such we argue that graph-based physical representations provide advantages in terms of both size and speed for executing queries. Drawing from research in sparse matrices, we use a compressed sparse row (CSR) format to model graph-oriented data. We also present two novel mechanisms for exploiting the CSR format that both find optimal join strategies and also prune variable bindings before expensive join operations occur. The first tactic we call Sprinkle SPARQL, which takes triple patterns of SPARQL queries and performs low-cost, linear-time set intersections to produce a constrained list of variable bindings for each variable in a query. Besides constrained lists of variable bindings, Sprinkle SPARQL also produces metrics that are consumed by the join algorithm to select an optimal execution path. The second tactic, graph joins, utilizes the CSR data structure as an index to efficiently join two variables expressed in a triple pattern together. We evaluate our approach on two data sets with over a billion edges: LUBM(8000) and an R-MAT graph generated with Graph5001 parameters and extended to have edge labels.


Archive | 2010

Scalable in-memory RDFS closure on billions of triples.

Eric Goodman; David Mizell


Archive | 2010

High performance semantic factoring of giga-scale semantic graph databases.

Cliff A. Joslyn; Robert D. Adolf; Sinan Al-Saffar; John Feo; Eric Goodman; David J. Haglin; Greg Mackey; David Mizell

Collaboration


Dive into the Eric Goodman's collaboration.

Top Co-Authors

Avatar

David J. Haglin

Minnesota State University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

John Feo

Pacific Northwest National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Sinan Al-Saffar

Pacific Northwest National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Dan Ventura

Brigham Young University

View shared research outputs
Top Co-Authors

Avatar

Bob Adolf

Pacific Northwest National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Chad Scherrer

Pacific Northwest National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Cliff A. Joslyn

Pacific Northwest National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Daniel G. Chavarría-Miranda

Pacific Northwest National Laboratory

View shared research outputs
Researchain Logo
Decentralizing Knowledge