Uwe Küster
Rus
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Uwe Küster.
international conference on computational science | 2006
Sunil R. Tiyyagura; Uwe Küster; Stefan Borowski
Many applications based on finite element and finite difference methods include the solution of large sparse linear systems using preconditioned iterative methods. Matrix vector multiplication is one of the key operations that has a significant impact on the performance of any iterative solver. In this paper, recent developments in sparse storage formats on vector machines are reviewed. Then, several improvements to memory access in the sparse matrix vector product are suggested. Particularly, algorithms based on dense blocks are discussed and reasons for their superior performance are explained. Finally, the performance gain by the presented modifications is demonstrated.
ieee international conference on high performance computing data and analytics | 2008
Sunil R. Tiyyagura; Panagiotis Adamidis; Rolf Rabenseifner; Peter Lammers; Stefan Borowski; F. Lippold; F. Svensson; Olaf Marxen; Stefan Haberhauer; Ari P. Seitsonen; J. Furthmüller; Katharina Benkert; Martin Galle; Thomas Bönisch; Uwe Küster; Michael M. Resch
This paper provides a comprehensive performance evaluation of the NEC SX-8 system at the High Performance Computing Center Stuttgart which has been in operation since July 2005. It provides a description of the installed hardware together with its performance for some synthetic benchmarks and five real world applications. All the applications achieved sustained Tflop/s performance. Additionally, the measurements presented show the ability of the system to solve not only large problems with a very high performance, but also medium sized problems with high efficiency using a large number of processors.
ieee international conference on high performance computing data and analytics | 2013
Daniel F. Harlacher; Sabine Roller; Florian Hindenlang; Claus-Dieter Munz; Tim Kraus; Martin Fischer; Koen Geurts; Matthias Meinke; Tobias Klühspies; Yevgeniya Kovalenko; Uwe Küster
The most important aspect for simulations in industrial design processes is the time to solution. To obtain highly detailed results nevertheless, massive computational resources have to be deployed. Feasibility and applicability of HPC systems to this purpose is the main focus of this paper. Two different numerical approaches, implemented with parallelism in mind, are investigated with respect to quality as well as turn around times on large super computing systems. The one approach compares the efficiency of high order schemes on coarser meshes to lower order schemes on finer meshes. The second approach employs a zonal coupling of LES and RANS to limit the computational effort by using solution adapted models. Three industrial use-cases evaluate the performance and quality of these approaches. General optimizations are presented as well as solutions for load-balancing.
ieee international conference on high performance computing data and analytics | 2011
Michael M. Resch; Uwe Küster
Moore’s law has come to an end with respect to the clock speed of the single processor. Clock rates are no longer increasing. Parallelism carries the day and accelerators are making the most of this. What is the future of processors for HPC going to look like? This talk will give a short overview and discuss some potential solutions.
Praxis Der Informationsverarbeitung Und Kommunikation | 2006
Michael M. Resch; Uwe Küster
ABSTRACT Supercomputing – or High Performance Computing (HPC) – has become to be dominated by standard components. A quick look at the Top500 list shows that clusters built from such standard components have become the architecture of choice. The fraction of clusters in the list has increased from about 2% in 2000 to about 73% in 2006. The key driving factor is the availability of competitive processor technology in the mass market on the one hand and a growing awareness of this potential in the user community on the other hand. These trends have motivated small companies to seek business opportunities in building large systems from standard components. Such commodity based systems rely mainly on massive parallelism in order to compete with traditional approaches. At the same time other approaches have been brought into the supercomputing arena which promise to allow for another leap in peak performance. This paper sets out to describe the current situation in supercomputing from a point of view of architectures and to describe futures scenarios and challenges that arise from the current trends in hardware and software development.
Archive | 2006
Natalia Currle-Linde; Benedetto Risio; Uwe Küster; Michael M. Resch
Currently, numerical simulation using automated parameter studies is already a key tool in discovering functional optima in complex systems such as biochemical drug design and car crash analysis. In the future, such studies of complex systems will be extremely important for the purpose of steering simulations. One such example is the optimum design and steering of computation equipment for power plants. The performance of today’s high performance computers enables simulation studies with results that are comparable to those obtained from physical experimentation. Recently, Grid technology has supported this development by providing uniform and secure access to computing resources over wide area networks (WANs), making it possible for industries to investigate large numbers of parameter sets using sophisticated optimization simulations. However, the large scale of such studies requires organized support for the submission, monitoring, and termination of jobs, as well as mechanisms for the collection of results, and the dynamic generation of new parameter sets in order to intelligently approach an optimum. In this paper, we describe a solution to these problems which we call Science Experimental Grid Laboratory (SEGL). The system defines complex workflows which can be executed in the Grid environment, and supports the dynamic generation of parameter sets.
ieee international conference on high performance computing data and analytics | 1994
Uwe Küster; Manuela Zürn
To formulate data structures suited to numerical problems like selfadapting multilevel algorithms several data constructs, conditions and attributes are necessary. These constructs are offered by languages like C++ or Fortran 90.
parallel computing | 2005
Natalia Currle-Linde; Uwe Küster; Michael M. Resch; Benedetto Risio
Archive | 2016
Andreas Kaminski; Björn Schembera; Michael M. Resch; Uwe Küster
Archive | 2003
Michael M. Resch; Uwe Küster; Matthias S. Müller; Ulrich Lang