Roger G. Grimes
Washington University in St. Louis
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Roger G. Grimes.
ieee international conference on high performance computing data and analytics | 1987
C. Cleveland Ashcraft; Roger G. Grimes; John G. Lewis; Barry W. Peyton; Horst D. Simon; Petter E. Bjørstad
This paper summarizes progress in the use of direct methods for solving very large sparse symmetric positive definite systems of linear equations on vector supercomputers. Sparse di rect solvers based on the multifrontal method or the general sparse method now outperform band or envelope solvers on vector supercomputers such as the CRAY X-MP. This departure from conventional wisdom is due to several advances. The hardware gather/scatter feature or indirect address feature of some recent vector super computers permits vectorization of the general sparse factorization. Other advances are algo rithmic. The new multiple minimum degree algo rithm calculates a powerful ordering much faster than its predecessors. Exploiting the supernode structure of the factored matrix provides vectori zation over nested loops, giving greater speed in the factorization module for the multifrontal and general sparse methods. Out-of-core versions of both methods are now available. Numerical re sults on the CRAY X-MP for several structural engineering problems demonstrate the impact of these improvements.
Siam Journal on Scientific and Statistical Computing | 1988
C. Cleveland Ashcraft; Roger G. Grimes
We consider the problem of vectorizing the recursive calculations found in modified incomplete factorizations and SSOR preconditioners for the conjugate gradient method. We examine matrix problems derived from partial differential equations which are discretized on regular 2-D and 3-D grids, where the grid nodes are ordered in the natural ordering. By performing data dependency analyses of the computations, we show that there is concurrency in both the factorization and the forward and backsolves. The computations may be performed with an average vector length of
ACM Signum Newsletter | 1982
Iain S. Duff; Roger G. Grimes; John G. Lewis; Bill Poole
O(n)
ieee international conference on high performance computing data and analytics | 2010
Robert F. Lucas; Gene Wagenbreth; Dan M. Davis; Roger G. Grimes
on an
Siam Journal on Scientific and Statistical Computing | 1987
A. M. Erisman; Roger G. Grimes; John G. Lewis; W. G. Poole; Horst D. Simon
n^2
Siam Journal on Scientific and Statistical Computing | 1981
Roger G. Grimes; John G. Lewis
or
Journal of Computational Physics | 1987
Roger G. Grimes; Henry Krakauer; John G. Lewis; Horst D. Simon; Su-Hai Wei
n^3
The Journal of Supercomputing | 1988
Roger G. Grimes
grid in two and three dimensions. Numerical studies on four model problems show that the conjugate gradient method using these vectorized implementations of the modified incomplete factorizations and SSOR preconditioners achieves overall speeds approaching 100 megaflops on a Cray X-MP/24 vector computer. Furthermore, these methods require considerably less overall execution time than diagonal scaling and no-fill red-black incomplete factorization preconditioners, both of which allow full vectorization but are not as convergent.
North-holland Mathematics Studies | 1986
Roger G. Grimes; John G. Lewis; Horst D. Simon
The development, analysis and production of algorithms in sparse linear algebra often requires the use of test problems to demonstrate the effectiveness and applicability of the algorithms. Many algorithms have been developed in the context of specific application areas and have been tested in the context of sets of test problems collected by the developers. Comparisons of algorithms across application areas and comparisons between algorithms has often been incomplete, due to the lack of a comprehensive set of test problems. Additionally we believe that a comprehensive set of test problems will lead to a better understanding of the range of structures in sparse matrix problems and thence to better classification and development of algorithms. We have agreed to sponsor and maintain a general library of sparse matrix test problems, available on request to anyone for a nominal fee to cover postal charges. Contributors to the library will, of course, receive a free copy.
The Journal of Supercomputing | 1988
Richard E. Anderson; Roger G. Grimes; Horst D. Simon
The use of GPUs to accelerate the factoring of large sparse symmetric matrices shows the potential of yielding important benefits to a large group of widely used applications. This paper examines how a multifrontal sparse solver performs when exploiting both the GPU and its multi-core host. It demonstrates that the GPU can dramatically accelerate the solver relative to one host CPU. Furthermore, the solver can profitably exploit both the GPU to factor its larger frontal matrices and multiple threads on the host to handle the smaller frontal matrices.