Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Celestine Dünner is active.

Publication


Featured researches published by Celestine Dünner.


international conference on communications | 2015

Endurance limits of MLC NAND flash

Thomas Parnell; Celestine Dünner; Thomas Mittelholzer; Nikolaos Papandreou; Haralampos Pozidis

An extensive effort is being undertaken by the flash community to develop signal processing and error-correction coding schemes that make use of soft information. Using experimental data from a state-of-the-art MLC flash device we demonstrate that the theoretical endurance improvement that such schemes can bring is limited. To investigate further, we develop a parametric channel model that takes into account the effects of cell-to-cell interference and demonstrate that it is the presence of programming errors in the channel that restricts the potential endurance enhancement that soft information can offer.


international conference on parallel processing | 2017

High-Performance Recommender System Training Using Co-Clustering on CPU/GPU Clusters

Kubilay Atasu; Thomas P. Parnell; Celestine Dünner; Michail Vlachos; Haralampos Pozidis

Recommender systems are becoming the crystal ball of the Internet because they can anticipate what the users may want, even before the users know they want it. However, the machine-learning algorithms typically involved in the training of such systems can be computationally expensive, and often may require several days for retraining. Here, we present a distributed approach for load-balancing the training of a recommender system based on state-of-art non-negative matrix factorization principles. The approach can exploit the presence of a cluster of mixed CPUs and GPUs, and results in a 466-fold performance improvement compared with the serial CPU implementation, and a 15-fold performance improvement compared with the best previously reported results for the popular Netflix data set.


IEEE Journal on Selected Areas in Communications | 2016

Capacity of the MLC NAND Flash Channel

Thomas Parnell; Celestine Dünner; Thomas Mittelholzer; Nikolaos Papandreou

In this paper, we develop a framework for evaluating the symmetric capacity of multilevel-cell (MLC) NAND flash devices while making very few assumptions regarding the underlying device physics. A set of recursive equations are derived that allow one to measure the symmetric capacity for any given page in a flash device using simple conditional statistics that can be extracted experimentally. Using data captured from two different 1y nm MLC devices, we demonstrate that the symmetric capacity of a flash page not only depends on the amount of program/erase cycling and data retention stress that has accumulated, but also on the position of the page within the flash block. We then study the effect on symmetric capacity of using optimized read-back schemes (both hard and soft) and show that while there is significant benefit, not all pages in the block are improved by the same amount. Finally, we show that it is possible to design error correction architectures that harness the inherent variation of symmetric capacity within a flash block to dramatically extend the program/erase cycling endurance of flash-based storage systems.


international conference on big data | 2017

Understanding and optimizing the performance of distributed machine learning applications on apache spark

Celestine Dünner; Thomas Parnell; Kubilay Atasu; Manolis Sifalakis; Haralampos Pozidis


international conference on machine learning | 2016

Primal-dual rates and certificates

Celestine Dünner; Simone Forte; Martin Takáč; Martin Jaggi


international conference on machine learning | 2018

A Distributed Second-Order Algorithm You Can Trust

Celestine Dünner; Matilde Gargiani; Aurelien Lucchi; An Bian; Thomas Hofmann; Martin Jaggi


international conference on big data | 2017

Linear-complexity relaxed word Mover's distance with GPU acceleration

Kubilay Atasu; Thomas Parnell; Celestine Dünner; Manolis Sifalakis; Haralampos Pozidis; Vasileios Vasileiadis; Michail Vlachos; Cesar Berrospi; Abdel Labbi


Archive | 2016

High-Performance Distributed Machine Learning using Apache SPARK.

Celestine Dünner; Thomas P. Parnell; Kubilay Atasu; Manolis Sifalakis; Haralampos Pozidis


neural information processing systems | 2018

Snap ML: A Hierarchical Framework for Machine Learning

Celestine Dünner; Thomas Parnell; Dimitrios Sarigiannis; Nikolas Ioannou; Andreea Anghel; Gummadi Ravi; Madhusudanan Kandasamy; Haralampos Pozidis


arXiv: Learning | 2018

Snap Machine Learning

Celestine Dünner; Thomas P. Parnell; Dimitrios Sarigiannis; Nikolas Ioannou; Haralampos Pozidis

Researchain Logo
Decentralizing Knowledge