Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Anne C. Elster is active.

Publication


Featured researches published by Anne C. Elster.


Medical Image Analysis | 2014

Comparing algorithms for automated vessel segmentation in computed tomography scans of the lung: the VESSEL12 study

Rina Dewi Rudyanto; Sjoerd Kerkstra; Eva M. van Rikxoort; Catalin I. Fetita; Pierre-Yves Brillet; Christophe Lefevre; Wenzhe Xue; Xiangjun Zhu; Jianming Liang; Ilkay Oksuz; Devrim Unay; Kamuran Kadipaşaogˇlu; Raúl San José Estépar; James C. Ross; George R. Washko; Juan-Carlos Prieto; Marcela Hernández Hoyos; Maciej Orkisz; Hans Meine; Markus Hüllebrand; Christina Stöcker; Fernando Lopez Mir; Valery Naranjo; Eliseo Villanueva; Marius Staring; Changyan Xiao; Berend C. Stoel; Anna Fabijańska; Erik Smistad; Anne C. Elster

The VESSEL12 (VESsel SEgmentation in the Lung) challenge objectively compares the performance of different algorithms to identify vessels in thoracic computed tomography (CT) scans. Vessel segmentation is fundamental in computer aided processing of data generated by 3D imaging modalities. As manual vessel segmentation is prohibitively time consuming, any real world application requires some form of automation. Several approaches exist for automated vessel segmentation, but judging their relative merits is difficult due to a lack of standardized evaluation. We present an annotated reference dataset containing 20 CT scans and propose nine categories to perform a comprehensive evaluation of vessel segmentation algorithms from both academia and industry. Twenty algorithms participated in the VESSEL12 challenge, held at International Symposium on Biomedical Imaging (ISBI) 2012. All results have been published at the VESSEL12 website http://vessel12.grand-challenge.org. The challenge remains ongoing and open to new participants. Our three contributions are: (1) an annotated reference dataset available online for evaluation of new algorithms; (2) a quantitative scoring system for objective comparison of algorithms; and (3) performance analysis of the strengths and weaknesses of the various vessel segmentation methods in the presence of various lung diseases.


Journal of open research software | 2014

Summary of the First Workshop on Sustainable Software for Science: Practice and Experiences (WSSSPE1)

Daniel S. Katz; Sou-Cheng T. Choi; Hilmar Lapp; Ketan Maheshwari; Frank Löffler; Matthew J. Turk; Marcus D. Hanwell; Nancy Wilkins-Diehr; James Hetherington; James Howison; Shel Swenson; Gabrielle Allen; Anne C. Elster; G. Bruce Berriman; Colin C. Venters

Challenges related to development, deployment, and maintenance of reusable software for science are becoming a growing concern. Many scientists’ research increasingly depends on the quality and availability of software upon which their works are built. To highlight some of these issues and share experiences, the First Workshop on Sustainable Software for Science: Practice and Experiences (WSSSPE1) was held in November 2013 in conjunction with the SC13 Conference. The workshop featured keynote presentations and a large number (54) of solicited extended abstracts that were grouped into three themes and presented via panels. A set of collaborative notes of the presentations and discussion was taken during the workshop. Unique perspectives were captured about issues such as comprehensive documentation, development and deployment practices, software licenses and career paths for developers. Attribution systems that account for evidence of software contribution and impact were also discussed. These include mechanisms such as Digital Object Identifiers, publication of “software papers”, and the use of online systems, for example source code repositories like GitHub. This paper summarizes the issues and shared experiences that were discussed, including cross-cutting issues and use cases. It joins a nascent literature seeking to understand what drives software work in science, and how it is impacted by the reward systems of science. These incentives can determine the extent to which developers are motivated to build software for the long-term, for the use of others, and whether to work collaboratively or separately. It also explores community building, leadership, and dynamics in relation to successful scientific software.


international conference on cloud computing and services science | 2016

CLOUDLIGHTNING: A Framework for a Self-organising and Self-managing Heterogeneous Cloud

Theo Lynn; Huanhuan Xiong; Dapeng Dong; Bilal Momani; George A. Gravvanis; Christos K. Filelis-Papadopoulos; Anne C. Elster; Malik Muhammad Zaki Murtaza Khan; Dimitrios Tzovaras; Konstantinos M. Giannoutakis; Dana Petcu; Marian Neagul; Ioan Dragon; Perumal Kuppudayar; Suryanarayanan Natarajan; Michael J. McGrath; Georgi Gaydadjiev; Tobias Becker; Anna Gourinovitch; David Kenny; John P. Morrison

As clouds increase in size and as machines of different types are added to the infrastructure in order to maximize performance and power efficiency, heterogeneous clouds are being created. However, exploiting different architectures poses significant challenges. To efficiently access heterogeneous resources and, at the same time, to exploit these resources to reduce application development effort, to make optimisations easier and to simplify service deployment, requires a re-evaluation of our approach to service delivery. We propose a novel cloud management and delivery architecture based on the principles of self-organisation and self-management that shifts the deployment and optimisation effort from the consumer to the software stack running on the cloud infrastructure. Our goal is to address inefficient use of resources and consequently to deliver savings to the cloud provider and consumer in terms of reduced power consumption and improved service delivery, with hyperscale systems particularly in mind. The framework is general but also endeavours to enable cloud services for high performance computing. Infrastructure-as-a-Service provision is the primary use case, however, we posit that genomics, oil and gas exploration, and ray tracing are three downstream use cases that will benefit from the proposed architecture.


international conference on acoustics, speech, and signal processing | 1989

Fast bit-reversal algorithms

Anne C. Elster

A novel fast algorithm for computing a sequence of bit-reversed integers is presented. In finding a mapping function from a sequence of integers to a sequence of their bit-reverse, a recursive approach is taken to overcome the logarithmic factor burdening the standard scheme. The associated constant for the timing factor is shown to be very low even at the register level. The method generalizes to radix-r and mixed-radix cases and provides an efficient vectorizable scheme with the same low constant.<<ETX>>


Journal of Real-time Image Processing | 2015

Real-time gradient vector flow on GPUs using OpenCL

Erik Smistad; Anne C. Elster; Frank Lindseth

The Gradient Vector Flow (GVF) is a feature-preserving spatial diffusion of gradients. It is used extensively in several image segmentation and skeletonization algorithms. Calculating the GVF is slow as many iterations are needed to reach convergence. However, each pixel or voxel can be processed in parallel for each iteration. This makes GVF ideal for execution on Graphic Processing Units (GPUs). In this paper, we present a highly optimized parallel GPU implementation of GVF written in OpenCL. We have investigated memory access optimization for GPUs, such as using texture memory, shared memory and a compressed storage format. Our results show that this algorithm really benefits from using the texture memory and the compressed storage format on the GPU. Shared memory, on the other hand, makes the calculations slower with or without the other optimizations because of an increased kernel complexity and synchronization. With these optimizations our implementation can process 2D images of large sizes (5122) in real-time and 3D images (2563) using only a few seconds on modern GPUs.


ieee international symposium on parallel & distributed processing, workshops and phd forum | 2011

Bandwidth Reduction through Multithreaded Compression of Seismic Images

Ahmed A. Aqrawi; Anne C. Elster

One of the main challenges of modern computer systems is to overcome the ever more prominent limitations of disk I/O and memory bandwidth, which today are thousands-fold slower than computational speeds. In this paper, we investigate reducing memory bandwidth and overall I/O and memory access times by using multithreaded compression and decompression of large datasets. Since the goal is to achieve a significant overall speedup of I/O, both level of compression achieved and efficiency of the compression and decompression algorithms, are of importance. Several compression methods for efficient disk access for large seismic datasets are implemented and empirically tested on on several modern CPUs and GPUs, including the Intel i7 and NVIDIA c2050 GPU. To reduce I/O time, both lossless and lossy symmetrical compression algorithms as well as hardware alternatives, are tested. Results show that I/O speedup may double by using an SSD vs. HDD disk on larger seismic datasets. Lossy methods investigated include variations of DCT-based methods in several dimensions, and combining these with lossless compression methods such as RLE (Run-Length Encoding) and Huffman encoding. Our best compression rate (0.16%) and speedups (6 for HDD and 3.2 for SSD) are achieved by using DCT in 3D and combining this with a modified RLE for lossy methods. It has an average error of 0.46% which is very acceptable for seismic applications. A simple predictive model for the execution time is also developed and shows an error of maximum 5% vs. our obtained results. It should thus be a good tool for predicting when to take advantage of multithreaded compression. This model and other techniques developed in this paper should also be applicable to several other data intensive applications.


international parallel and distributed processing symposium | 2015

Machine Learning Based Auto-Tuning for Enhanced OpenCL Performance Portability

Thomas L. Falch; Anne C. Elster

Heterogeneous computing, which combines devices with different architectures, is rising in popularity, and promises increased performance combined with reduced energy consumption. OpenCL has been proposed as a standard for programing such systems, and offers functional portability. It does, however, suffer from poor performance portability, code tuned for one device must be re-tuned to achieve good performance on another device. In this paper, we use machine learning-based auto-tuning to address this problem. Benchmarks are run on a random subset of the entire tuning parameter configuration space, and the results are used to build an artificial neural network based model. The model can then be used to find interesting parts of the parameter space for further search. We evaluate our method with different benchmarks, on several devices, including an Intel i7 3770 CPU, an Nvidia K40 GPU and an AMD Radeon HD 7970 GPU. Our model achieves a mean relative error as low as 6.1%, and is able to find configurations as little as 1.3% worse than the global minimum.


computational methods in systems biology | 2012

Population dynamics p systems on CUDA

Miguel A. Martínez-del-Amor; Ignacio Pérez-Hurtado; Adolfo Gastalver-Rubio; Anne C. Elster; Mario J. Pérez-Jiménez

Population Dynamics P systems (PDP systems, in short) provide a new formal bio-inspired modeling framework, which has been successfully used by ecologists. These models are validated using software tools against actual measurements. The goal is to use P systems simulations to adopt a priori management strategies for real ecosystems. Software for PDP systems is still in an early stage. The simulation of PDP systems is both computationally and data intensive for large models. Therefore, the development of efficient simulators is needed for this field. In this paper, we introduce a novel simulator for PDP systems accelerated by the use of the computational power of GPUs. We discuss the implementation of each part of the simulator, and show how to achieve up to a 7x speedup on a NVIDA Tesla C1060 compared to an optimized multicore version on a Intel 4-core i5 Xeon for large systems. Other results and testing methodologies are also included.


international parallel and distributed processing symposium | 2009

A super-efficient adaptable bit-reversal algorithm for multithreaded architectures

Anne C. Elster; Jan Christian Meyer

Fast bit-reversal algorithms have been of strong interest for many decades, especially after Cooley and Tukey introduced their FFT implementation in 1965. Many recent algorithms, including FFTW try to avoid the bit-reversal all together by doing in-place algorithms within their FFTs. We therefore motivate our work by showing that for FFTs of up to 65.536 points, a minimally tuned Cooley-Tukey FFT in C using our bit-reversal algorithm performs comparable or better than the default FFTW algorithm. In this paper, we present an extremely fast linear bit-reversal adapted for modern multithreaded architectures. Our bit-reversal algorithm takes advantage of recursive calls combined with the fact that it only generates pairs of indices for which the corresponding elements need to be exchanges, thereby avoiding any explicit tests. In addition we have implemented an adaptive approach which explores the trade-off between compile time and run-time work load. By generating look-up tables at compile time, our algorithm becomes even faster at run-time. Our results also show that by using more than one thread on tightly coupled architectures, further speed-up can be achieved.


parallel computing | 2006

Parallel methods for real-time visualization of snow

Ingar Saltvik; Anne C. Elster; Henrik Rojas Nagel

Snow is a familiar scene in the Nordic countries during the winter months. This paper discusses some of the complex numerical algorithms behind snow simulations. Previous methods for snow simulation have either covered only a very limited aspect of snow, or have been unsuitable for real-time performance. Here, some of these methods are combined into a model for real-time snow simulation that handles snowflake motion through the air, wind simulation, as well as accumulation of snow on objects including the ground. With our goal towards achieving realtime performance with more than 25 frames per second, some new parallel methods for the snow model are introduced. The algorithms are first parallelized by dividing the data structures among threads. This scheme is then improved by overlapping inherently sequential algorithms with computations for the following frame, to eliminate processor idle time. SMP and multi-core systems are considered.

Collaboration


Dive into the Anne C. Elster's collaboration.

Top Co-Authors

Avatar

Thomas L. Falch

Norwegian University of Science and Technology

View shared research outputs
Top Co-Authors

Avatar

Jan Christian Meyer

Norwegian University of Science and Technology

View shared research outputs
Top Co-Authors

Avatar

Erik Smistad

Norwegian University of Science and Technology

View shared research outputs
Top Co-Authors

Avatar

Thorvald Natvig

Norwegian University of Science and Technology

View shared research outputs
Top Co-Authors

Avatar

Frank Lindseth

Norwegian University of Science and Technology

View shared research outputs
Top Co-Authors

Avatar

Malik Muhammad Zaki Murtaza Khan

Norwegian University of Science and Technology

View shared research outputs
Top Co-Authors

Avatar

Rune Erlend Jensen

Norwegian University of Science and Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Ian Karlin

Norwegian University of Science and Technology

View shared research outputs
Top Co-Authors

Avatar

Øystein E. Krog

Norwegian University of Science and Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge