Piotr Luszczek
MathWorks
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Piotr Luszczek.
Journal of Physics: Conference Series | 2009
Emmanuel Agullo; James W. Demmel; Jack J. Dongarra; Bilel Hadri; Jakub Kurzak; Julien Langou; Hatem Ltaief; Piotr Luszczek; Stanimire Tomov
The emergence and continuing use of multi-core architectures and graphics processing units require changes in the existing software and sometimes even a redesign of the established algorithms in order to take advantage of now prevailing parallelism. Parallel Linear Algebra for Scalable Multi-core Architectures (PLASMA) and Matrix Algebra on GPU and Multics Architectures (MAGMA) are two projects that aims to achieve high performance and portability across a wide range of multi-core architectures and hybrid systems respectively. We present in this document a comparative study of PLASMAs performance against established linear algebra packages and some preliminary results of MAGMA on hybrid multi-core and GPU systems.
Computer Physics Communications | 2009
Marc Baboulin; Alfredo Buttari; Jack J. Dongarra; Jakub Kurzak; Julie Langou; Julien Langou; Piotr Luszczek; Stanimire Tomov
On modern architectures, the performance of 32-bit operations is often at least twice as fast as the performance of 64-bit operations. By using a combination of 32-bit and 64-bit floating point arithmetic, the performance of many dense and sparse linear algebra algorithms can be significantly enhanced while maintaining the 64-bit accuracy of the resulting solution. The approach presented here can apply not only to conventional processors but also to other technologies such as Field Programmable Gate Arrays (FPGA), Graphical Processing Units (GPU), and the STI Cell BE processor. Results on modern processor architectures and the STI Cell BE are presented.
ACM Transactions on Mathematical Software | 2008
Alfredo Buttari; Jack J. Dongarra; Jakub Kurzak; Piotr Luszczek; Stanimire Tomov
By using a combination of 32-bit and 64-bit floating point arithmetic, the performance of many sparse linear algebra algorithms can be significantly enhanced while maintaining the 64-bit accuracy of the resulting solution. These ideas can be applied to sparse multifrontal and supernodal direct techniques and sparse iterative techniques such as Krylov subspace methods. The approach presented here can apply not only to conventional processors but also to exotic technologies such as Field Programmable Gate Arrays (FPGA), Graphical Processing Units (GPU), and the Cell BE processor.
Computing in Science and Engineering | 2008
Jakub Kurzak; Alfredo Buttari; Piotr Luszczek; Jack J. Dongarra
Is real-world gaming technology the next big thing in the more academically based high-performance computing arena? The authors put PlayStation 3 to the test.
Advances in Computers | 2008
Jack J. Dongarra; Robert Graybill; William Harrod; Robert F. Lucas; Ewing L. Lusk; Piotr Luszczek; Janice McMahon; Allan Snavely; Jeffrey S. Vetter; Katherine A. Yelick; Sadaf R. Alam; Roy L. Campbell; Laura Carrington; Tzu-Yi Chen; Omid Khalili; Jeremy S. Meredith; Mustafa M. Tikir
Abstract The historical context with regard to the origin of the DARPA High Productivity Computing Systems (HPCS) program is important for understanding why federal government agencies launched this new, long-term high-performance computing program and renewed their commitment to leadership computing in support of national security, large science and space requirements at the start of the 21st century. In this chapter, we provide an overview of the context for this work as well as various procedures being undertaken for evaluating the effectiveness of this activity including such topics as modelling the proposed performance of the new machines, evaluating the proposed architectures, understanding the languages used to program these machines as well as understanding programmer productivity issues in order to better prepare for the introduction of these machines in the 2011–2015 timeframe.
Archive | 2017
Jakub Kurzak; Piotr Luszczek
The goal of the BONSAI project is to develop a software infrastructure for using parallel hybrid systems at any scale to carry out large, concurrent autotuning sweeps in order to dramatically accelerate the optimization process of computational kernels for GPU accelerators and many-core coprocessors.
Archive | 2011
Piotr Luszczek; Jakub Kurzak; Jack Dongarra
Over the years, computational physics and chemistry served as an ongoing 5 source of problems that demanded the ever increasing performance from 6 hardware as well as the software that ran on top of it. Most of these 7 problems could be translated into solutions for systems of linear equa8 tions: the very topic of numerical linear algebra. Seemingly then, a set of 9 efficient linear solvers could be solving important scientific problems for 10 years to come. We argue that dramatic changes in hardware designs pre11 cipitated by the shifting nature of the marketplace of computer hardware 12 had a continuous effect on the software for numerical linear algebra. The 13 extraction of high percentages of peak performance continues to require 14 adaptation of software. If the past history of this adaptive nature of linear 15 algebra software is any guide then the future theme will feature changes 16 as well — changes aimed at harnessing the incredible advances of the 17 evolving hardware infrastructure. 18
Archive | 2008
Piotr Luszczek; John N. Little; Jocelyn Luke Martin; Halldor N. Stefansson; Edric Ellis; Penelope L. Anderson; Brett Baker; Loren Dean; Roy E. Lurie
Archive | 2008
Halldor N. Stefansson; Penelope L. Anderson; Brett Baker; Edric Ellis; Joseph F. Hicklin; John N. Little; Jocelyn Luke Martin; Piotr Luszczek; Nausheen B. Moulana
Archive | 2008
Piotr Luszczek; John N. Little; Joseph F. Hicklin; Jocelyn Luke Martin; Halldor N. Stefansson; Edric Ellis; Penelope L. Anderson; Nausheen B. Moulana; Brett Baker; Loren Dean; Roy E. Lurie