Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Daniel Kunkle is active.

Publication


Featured researches published by Daniel Kunkle.


ieee international conference on high performance computing, data, and analytics | 2008

A load balancing framework for clustered storage systems

Daniel Kunkle; Jiri Schindler

The load balancing framework for high-performance clustered storagesystems presented in this paper provides a general method for reconfiguringa system facing dynamic workload changes. It simultaneously balances load andminimizes the cost of reconfiguration. It can be used for automatic reconfigurationor to present an administrator with a range of (near) optimal reconfigurationoptions, allowing a tradeoff between load distribution and reconfiguration cost.The framework supports a wide range of measures for load imbalance and reconfigurationcost, as well as several optimization techniques. The effectivenessof this framework is demonstrated by balancing the workload on a NetApp DataONTAP GX system, a commercial scale-out clustered NFS server implementation.The evaluation scenario considers consolidating two real world systems,with hundreds of users each: a six-node clustered storage system supporting engineeringworkloads and a legacy system supporting three email severs.


Journal of Computer Science and Technology | 2008

Mining Frequent Generalized Itemsets and Generalized Association Rules Without Redundancy

Daniel Kunkle; Donghui Zhang; Gene Cooperman

This paper presents some new algorithms to efficiently mine max frequent generalized itemsets (g-itemsets) and essential generalized association rules (g-rules). These are compact and general representations for all frequent patterns and all strong association rules in the generalized environment. Our results fill an important gap among algorithms for frequent patterns and association rules by combining two concepts. First, generalized itemsets employ a taxonomy of items, rather than a flat list of items. This produces more natural frequent itemsets and associations such as (meat, milk) instead of (beef, milk), (chicken, milk), etc. Second, compact representations of frequent itemsets and strong rules, whose result size is exponentially smaller, can solve a standard dilemma in mining patterns: with small threshold values for support and confidence, the user is overwhelmed by the extraordinary number of identified patterns and associations; but with large threshold values, some interesting patterns and associations fail to be identified.Our algorithms can also expand those max frequent g-itemsets and essential g-rules into the much larger set of ordinary frequent g-itemsets and strong g-rules. While that expansion is not recommended in most practical cases, we do so in order to present a comparison with existing algorithms that only handle ordinary frequent g-itemsets. In this case, the new algorithm is shown to be thousands, and in some cases millions, of the time faster than previous algorithms. Further, the new algorithm succeeds in analyzing deeper taxonomies, with the depths of seven or more. Experimental results for previous algorithms limited themselves to taxonomies with depth at most three or four.In each of the two problems, a straightforward lattice-based approach is briefly discussed and then a classification-based algorithm is developed. In particular, the two classification-based algorithms are MFGI_class for mining max frequent g-itemsets and EGR_class for mining essential g-rules. The classification-based algorithms are featured with conceptual classification trees and dynamic generation and pruning algorithms.


parallel symbolic computation | 2010

Parallel disk-based computation for large, monolithic binary decision diagrams

Daniel Kunkle; Vlad Slavici; Gene Cooperman

Binary Decision Diagrams (BDDs) are widely used in formal verification. They are also widely known for consuming large amounts of memory. For larger problems, a BDD computation will often start thrashing due to lack of memory within minutes. This work uses the parallel disks of a cluster or a SAN (storage area network) as an extension of RAM, in order to efficiently compute with BDDs that are orders of magnitude larger than what is available on a typical computer. The use of parallel disks overcomes the bandwidth problem of single disk methods, since the bandwidth of 50 disks is similar to the bandwidth of a single RAM sub-system. In order to overcome the latency issues of disk, the Roomy library is used for the sake of its latency-tolerant data structures. A breadth-first algorithm is implemented. A further advantage of the algorithm is that RAM usage can be very modest, since its largest use is as buffers for open files. The success of the method is demonstrated by solving the 16-queens problem, and by solving a more unusual problem --- counting the number of tie games in a three-dimensional 4x4x4 tic-tac-toe board.


Journal of Symbolic Computation | 2009

Harnessing parallel disks to solve Rubik's cube

Daniel Kunkle; Gene Cooperman

The number of moves required to solve any configuration of Rubiks cube has held a fascination for over 25 years. A new upper bound of 26 is produced. More important, a new methodology is described for finding upper bounds. The novelty is two-fold. First, parallel disks are employed. This allows 1.4x10^1^2 states representing symmetrized cosets to be enumerated in seven terabytes. Second, a faster table-based multiplication is described for symmetrized cosets that attempts to keep most tables in the CPU cache. This enables the product of a symmetrized coset by a generator at a rate of 10 million moves per second.


Communications of The ACM | 2008

Solving Rubik's Cube: disk is the new RAM

Daniel Kunkle; Gene Cooperman

Substituting disk for RAM, disk-based computation is a way to increase working memory and achieve results that are not otherwise economical.


parallel symbolic computation | 2010

Roomy: a system for space limited computations

Daniel Kunkle

There are numerous examples of problems in symbolic algebra in which the required storage grows far beyond the limitations even of the distributed RAM of a cluster. Often this limitation determines how large a problem one can solve in practice. Roomy provides a minimally invasive system to modify the code for such a computation, in order to use the local disks of a cluster or a SAN as a transparent extension of RAM. Roomy is implemented as a C/C++ library. It provides some simple data structures (arrays, unordered lists, and hash tables). Some typical programming constructs that one might employ in Roomy are: map, reduce, duplicate elimination, chain reduction, pair reduction, and breadth-first search. All aspects of parallelism and remote I/O are hidden within the Roomy library.


international symposium on symbolic and algebraic computation | 2009

Biased tadpoles: a fast algorithm for centralizers in large matrix groups

Daniel Kunkle; Gene Cooperman

Centralizers are an important tool in in computational group theory. Yet for large matrix groups, they tend to be slow. We demonstrate a O(√|G|(1/logε)) black box randomized algorithm that produces a centralizer using space logarithmic in the order of the centralizer, even for typical matrix groups of order 1020 . An optimized version of this algorithm (larger space and no longer black box) typically runs in seconds for groups of order 1015 and minutes for groups of order 1020. Further, the algorithm trivially parallelizes, and so linear speedup is achieved in an experiment on a computer with four CPU cores. The novelty lies in the use of a biased tadpole, which delivers an order of magnitude speedup as compared to the classical tadpole algorithm. The biased tadpole also allows a test for membership in a conjugacy class in a fraction of a second. Finally, the same methodology quickly finds the order of a matrix group via a vector stabilizer. This allows one to eliminate the already small possibility of error in the randomized centralizer algorithm.


conference on information and knowledge management | 2006

Efficient mining of max frequent patterns in a generalized environment

Daniel Kunkle; Donghui Zhang; Gene Cooperman

This poster paper summarizes our solution for mining max frequent generalized itemsets (g-itemsets), a compact representation for frequent patterns in the generalized environment.


international symposium on symbolic and algebraic computation | 2010

Fast multiplication of large permutations for disk, flash memory and RAM

Vlad Slavici; Xin Dong; Daniel Kunkle; Gene Cooperman

Permutation multiplication (or permutation composition) is perhaps the simplest of all algorithms in computer science. Yet for large permutations, the standard algorithm is not the fastest for disk or for flash, and surprisingly, it is not even the fastest algorithm for RAM on recent multi-core CPUs. On a recent commodity eight-core machine we demonstrate a novel algorithm that is 50% faster than the traditional algorithm. For larger permutations on flash or disk, the novel algorithm is orders of magnitude faster. A disk-parallel algorithm is demonstrated that can multiply two permutations with 12.8 billion points using 16 parallel local disks of a cluster in under one hour. Such large permutations are important in computational group theory, where they arise as the result of the well-known Todd-Coxeter coset enumeration algorithm. The novel algorithm emphasizes several passes of streaming access to the data instead of the traditional single pass using random access to the data. Similar novel algorithms are presented for permutation inverse and permutation multiplication by an inverse, thus providing a complete library of the underlying permutation operations needed for computations with permutation groups.


Nonlinear Dynamics, Psychology, and Life Sciences | 2004

Emergence of constraint in self-organizing systems.

Stephen Guerin; Daniel Kunkle

Collaboration


Dive into the Daniel Kunkle's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Vlad Slavici

Northeastern University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Sean Boyle

London School of Economics and Political Science

View shared research outputs
Top Co-Authors

Avatar

Jonathan Arney

Rochester Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Peter G. Anderson

Rochester Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Xin Dong

Northeastern University

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge