Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Joseph W. H. Liu is active.

Publication


Featured researches published by Joseph W. H. Liu.


SIAM Journal on Matrix Analysis and Applications | 1999

A Supernodal Approach to Sparse Partial Pivoting

James Demmel; Stanley C. Eisenstat; John R. Gilbert; Xiaoye S. Li; Joseph W. H. Liu

We investigate several ways to improve the performance of sparse LU factorization with partial pivoting, as used to solve unsymmetric linear systems. We introduce the notion of unsymmetric supernodes to perform most of the numerical computation in dense matrix kernels. We introduce unsymmetric supernode-panel updates and two-dimensional data partitioning to better exploit the memory hierarchy. We use Gilbert and Peierlss depth-first search with Eisenstat and Lius symmetric structural reductions to speed up symbolic factorization. We have developed a sparse LU code using all these ideas. We present experiments demonstrating that it is significantly faster than earlier partial pivoting codes. We also compare its performance with UMFPACK, which uses a multifrontal approach; our code is very competitive in time and storage requirements, especially for large problems.


SIAM Journal on Matrix Analysis and Applications | 1990

The role of elimination trees in sparse factorization

Joseph W. H. Liu

In this paper, the role of elimination trees in the direct solution of large sparse linear systems is examined. The notion of elimination trees is described and its relation to sparse Cholesky factorization is discussed. The use of elimination trees in the various phases of direct factorization are surveyed: in reordering, sparse storage schemes, symbolic factorization, numeric factorization, and different computing environments.


Siam Review | 1992

The multifrontal method for sparse matrix solution: theory and practice

Joseph W. H. Liu

This paper presents an overview of the multifrontal method for the solution of large sparse symmetric positive definite linear systems. The method is formulated in terms of frontal matrices, update matrices, and an assembly tree. Formal definitions of these notions are given based on the sparse matrix structure. Various advances to the basic method are surveyed. They include the role of matrix reorderings, the use of supernodes, and other implementatjon techniques. The use of the method in different computational environments is also described.


ACM Transactions on Mathematical Software | 1985

Modification of the minimum-degree algorithm by multiple elimination

Joseph W. H. Liu

The most widely used ordering scheme to reduce fills and operations in sparse matrix computation is the minimum-degree algorithm. The notion of multiple elimination is introduced here as a modification to the conventional scheme. The motivation is discussed using the k-by-k grid model problem. Experimental results indicate that the modified version retains the fill-reducing property of (and is often better than) the original ordering algorithm and yet requires less computer time. The reduction in ordering time is problem dependent, and for some problems the modified algorithm can run a few times faster than existing implementations of the minimum-degree algorithm. The use of external degree in the algorithm is also introduced.


Archive | 1993

Graph theory and sparse matrix computation

Alan George; John R. Gilbert; Joseph W. H. Liu

When reality is modelled by computation, matrices are often the connection between the continuous physical world and the finite algorithmic one. Usually, the more detailed the model, the bigger the matrix; however, efficiency demands that every possible advantage be exploited. The articles in this volume are based on recent research on sparse matrix computations. They examine graph theory as it connects to linear algebra, parallel computing, data structures, geometry and both numerical and discrete algorithms. The articles are grouped into three general categories: graph models of symmetric matrices and factorizations; graph models of algorithms on nonsymmetric matrices; and parallel sparse matrix algorithms.


Siam Journal on Scientific and Statistical Computing | 1988

Sparse Cholesky factorization on a local-memory multiprocessor

Alan George; Michael T. Heath; Joseph W. H. Liu; Esmond Ng

This article deals with the problem of factoring a large sparse positive definite matrix on a multiprocessor system. The processors are assumed to have substantial local memory but no globally shared memory. They communicate among themselves and with a host processor through message passing. Our primary interest is in designing an algorithm which exploits parallelism, rather than in exploiting features of the underlying topology of the hardware. However, part of our study is aimed at determining, for certain sparse matrix problems, whether hardware based on the binary hypercube topology adequately supports the communication requirements for such problems. Numerical results from experiments conducted on a hypercube multiprocessor are included.


parallel computing | 1989

Communication results for parallel sparse Cholesky factorization on a hypercube

Alan George; Joseph W. H. Liu; Esmond G. Ng

Abstract We consider the problem of reducing data traffic among processor nodes during the parallel factorization of a sparse matrix on a hypercube multiprocessor. A task assignment strategy based on the structure of an elimination tree is presented. This assignment is aimed at achieving load balancing among the processors and also reducing the amount of processor-to-processor data communication. An analysis of regular grid problems is presented, providing a bound on communication volume generated by the new strategy, and showing that the allocation scheme is optimal in the asymptotic sense. Some experimental results on the performance of this scheme are presented.


ACM Transactions on Mathematical Software | 1986

A compact row storage scheme for Cholesky factors using elimination trees

Joseph W. H. Liu

For a given sparse symmetric positive definite matrix, a compact row-oriented storage scheme for its Cholesky factor is introduced. The scheme is based on the structure of an elimination tree defined for the given matrix. This new storage scheme has the distinct advantage of having the amount of overhead storage required for indexing always bounded by the number of nonzeros in the original matrix. The structural representation may be viewed as storing the minimal structure of the given matrix that will preserve the symbolic Cholesky factor. Experimental results on practical problems indicate that the amount of savings in overhead storage can be substantial when compared with Shermans compressed column storage scheme.


parallel computing | 1986

Computational models and task scheduling for parallel sparse Cholesky factorization

Joseph W. H. Liu

Abstract In this paper, a systematic and unified treatment of computational task models for parallel sparse Cholesky factorization is presented. They are classified as fine-, medium-, and large-grained graph models. In particular, a new medium-grained model based on column-oriented tasks is introduced, and it is shown to correspond structurally to the filled graph of the given sparse matrix. The task scheduling problem for the various task graphs is also discussed. A practical algorithm to schedule the column tasks of the medium-grained model for multiple processors is described. It is based on a heuristic critical path scheduling method. This will give an overall scheme for parallel sparse Cholesky factorization, appropriate for parallel machines with shared-memory architecture like the Denelcor HEP.


ACM Transactions on Mathematical Software | 1986

On the storage requirement in the out-of-core multifrontal method for sparse factorization

Joseph W. H. Liu

Two techniques are introduced to reduce the working storage requirement for the recent multifrontal method of Duff and Reid used in the sparse out-of-core factorization of symmetric matrices. For a given core size, the reduction in working storage allows some large problems to be solved without having to use auxiliary storage for the working arrays. Even if the working arrays exceed the core size, it will reduce the amount of input-output traffic necessary to manipulate the working vectors. Experimental results are provided to demonstrate significant storage reduction on practical problems using the proposed techniques.

Collaboration


Dive into the Joseph W. H. Liu's collaboration.

Top Co-Authors

Avatar

Alan George

University of Waterloo

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Alan George

University of Waterloo

View shared research outputs
Top Co-Authors

Avatar

Esmond G. Ng

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Esmond Ng

Oak Ridge National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Barry W. Peyton

Oak Ridge National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Esmond Ng

Oak Ridge National Laboratory

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Eleanor Chu

University of Waterloo

View shared research outputs
Researchain Logo
Decentralizing Knowledge