Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Alexandros Daglis is active.

Publication


Featured researches published by Alexandros Daglis.


architectural support for programming languages and operating systems | 2014

Scale-out NUMA

Stanko Novakovic; Alexandros Daglis; Edouard Bugnion; Babak Falsafi; Boris Grot

Emerging datacenter applications operate on vast datasets that are kept in DRAM to minimize latency. The large number of servers needed to accommodate this massive memory footprint requires frequent server-to-server communication in applications such as key-value stores and graph-based applications that rely on large irregular data structures. The fine-grained nature of the accesses is a poor match to commodity networking technologies, including RDMA, which incur delays of 10-1000x over local DRAM operations. We introduce Scale-Out NUMA (soNUMA) -- an architecture, programming model, and communication protocol for low-latency, distributed in-memory processing. soNUMA layers an RDMA-inspired programming model directly on top of a NUMA memory fabric via a stateless messaging protocol. To facilitate interactions between the application, OS, and the fabric, soNUMA relies on the remote memory controller -- a new architecturally-exposed hardware block integrated into the nodes local coherence hierarchy. Our results based on cycle-accurate full-system simulation show that soNUMA performs remote reads at latencies that are within 4x of local DRAM, can fully utilize the available memory bandwidth, and can issue up to 10M remote memory operations per second per core.


international symposium on computer architecture | 2015

Manycore network interfaces for in-memory rack-scale computing

Alexandros Daglis; Stanko Novakovic; Edouard Bugnion; Babak Falsafi; Boris Grot

Datacenter operators rely on low-cost, high-density technologies to maximize throughput for data-intensive services with tight tail latencies. In-memory rack-scale computing is emerging as a promising paradigm in scale-out datacenters capitalizing on commodity SoCs, low-latency and high-bandwidth communication fabrics and a remote memory access model to enable aggregation of a racks memory for critical data-intensive applications such as graph processing or key-value stores. Low latency and high bandwidth not only dictate eliminating communication bottlenecks in the software protocols and off-chip fabrics but also a careful on-chip integration of network interfaces. The latter is a key challenge especially in architectures with RDMA-inspired one-sided operations that aim to achieve low latency and high bandwidth through on-chip Network Interface (NI) support. This paper proposes and evaluates network interface architectures for tiled manycore SoCs for in-memory rack-scale computing. Our results indicate that a careful splitting of NI functionality per chip tile and at the chips edge along a NOC dimension enables a rack-scale architecture to optimize for both latency and bandwidth. Our best manycore NI architecture achieves latencies within 3% of an idealized hardware NUMA and efficiently uses the full bisection bandwidth of the NOC, without changing the on-chip coherence protocol or the cores microarchitecture.


symposium on cloud computing | 2016

The Case for RackOut: Scalable Data Serving Using Rack-Scale Systems

Stanko Novakovic; Alexandros Daglis; Edouard Bugnion; Babak Falsafi; Boris Grot

To provide low latency and high throughput guarantees, most large key-value stores keep the data in the memory of many servers. Despite the natural parallelism across lookups, the load imbalance, introduced by heavy skew in the popularity distribution of keys, limits performance. To avoid violating tail latency service-level objectives, systems tend to keep server utilization low and organize the data in micro-shards, which provides units of migration and replication for the purpose of load balancing. These techniques reduce the skew, but incur additional monitoring, data replication and consistency maintenance overheads. In this work, we introduce RackOut, a memory pooling technique that leverages the one-sided remote read primitive of emerging rack-scale systems to mitigate load imbalance while respecting service-level objectives. In RackOut, the data is aggregated at rack-scale granularity, with all of the participating servers in the rack jointly servicing all of the racks micro-shards. We develop a queuing model to evaluate the impact of RackOut at the datacenter scale. In addition, we implement a RackOut proof-of-concept key-value store, evaluate it on two experimental platforms based on RDMA and Scale-Out NUMA, and use these results to validate the model. Our results show that RackOut can increase throughput up to 6x for RDMA and 8.6x for Scale-Out NUMA compared to a scale-out deployment, while respecting tight tail latency service-level objectives.


international symposium on computer architecture | 2017

The Mondrian Data Engine

Mario Drumond; Alexandros Daglis; Nooshin S. Mirzadeh; Dmitrii Ustiugov; Javier Picorel; Babak Falsafi; Boris Grot; Dionisios N. Pnevmatikatos

The increasing demand for extracting value out of ever-growing data poses an ongoing challenge to system designers, a task only made trickier by the end of Dennard scaling. As the performance density of traditional CPU-centric architectures stagnates, advancing compute capabilities necessitates novel architectural approaches. Near-memory processing (NMP) architectures are reemerging as promising candidates to improve computing efficiency through tight coupling of logic and memory. NMP architectures are especially fitting for data analytics, as they provide immense bandwidth to memory-resident data and dramatically reduce data movement, the main source of energy consumption. Modern data analytics operators are optimized for CPU execution and hence rely on large caches and employ random memory accesses. In the context of NMP, such random accesses result in wasteful DRAM row buffer activations that account for a significant fraction of the total memory access energy. In addition, utilizing NMPs ample bandwidth with fine-grained random accesses requires complex hardware that cannot be accommodated under NMPs tight area and power constraints. Our thesis is that efficient NMP calls for an algorithm-hardware co-design that favors algorithms with sequential accesses to enable simple hardware that accesses memory in streams. We introduce an instance of such a co-designed NMP architecture for data analytics, the Mondrian Data Engine. Compared to a CPU-centric and a baseline NMP system, the Mondrian Data Engine improves the performance of basic data analytics operators by up to 49× and 5×, and efficiency by up to 28× and 5×, respectively.


international symposium on microarchitecture | 2016

SABRes: atomic object reads for in-memory rack-scale computing

Alexandros Daglis; Dmitrii Ustiugov; Stanko Novakovic; Edouard Bugnion; Babak Falsafi; Boris Grot

Modern in-memory services rely on large distributed object stores to achieve the high scalability essential to ser-vice thousands of requests concurrently. The independent and unpredictable nature of incoming requests results in random accesses to the object store, triggering frequent remote memory accesses. State-of-the-art distributed memory frameworks leverage the one-sided operations offered by RDMA technology to mitigate the traditionally high cost of remote memory access. Unfortunately, the limited semantics of RDMA one-sided operations bound remote memory access atomicity to a single cache block; therefore, atomic remote object access relies on software mechanisms. Emerging highly integrated rack-scale systems that reduce the latency of one-sided operations to a small multiple of DRAM latency expose the overhead of these software mechanisms as a major latency contributor. This technology-triggered paradigm shift calls for new one-sided operations with stronger semantics. We take a step in that direction by proposing SABRes, a new one-sided operation that provides atomic remote object reads in hardware. We then present LightSABRes, a lightweight hardware accelerator for SABRes that removes all atomicity-associated software overheads. Compared to a state-of-the-art software atomicity mechanism, LightSABRes improve the throughput of a microbenchmark atomically accessing 128B-8KB objects from remote memory by 15-97%, and the throughput of a modern in-memory distributed object store by 30-60%.


measurement and modeling of computer systems | 2016

An Analysis of Load Imbalance in Scale-out Data Serving

Stanko Novakovic; Alexandros Daglis; Edouard Bugnion; Babak Falsafi; Boris Grot

Despite the natural parallelism across lookups, performance of distributed key-value stores is often limited due to load imbalance induced by heavy skew in the popularity distribution of the dataset. To avoid violating service level objectives expressed in terms of tail latency, systems tend to keep server utilization low and organize the data in micro-shards, which in turn provides units of migration and replication for the purpose of load balancing. These techniques reduce the skew, but incur additional monitoring, data replication and consistency maintenance overheads. This work shows that the trend towards extreme scale-out will further exacerbate the skew-induced load imbalance, and hence the overhead of migration and replication.


Operating Systems Review | 2018

Algorithm/Architecture Co-Design for Near-Memory Processing

Mario Drumond; Alexandros Daglis; Nooshin S. Mirzadeh; Dmitrii Ustiugov; Javier Picorel; Babak Falsafi; Boris Grot; Dionisios N. Pnevmatikatos

With mainstream technologies to couple logic tightly with memory on the horizon, near-memory processing has re-emerged as a promising approach to improving performance and energy for data-centric computing. DRAM, however, is primarily designed for density and low cost, with a rigid internal organization that favors coarse-grain streaming rather than byte-level random access. This paper makes the case that treating DRAM as a block-oriented streaming device yields significant efficiency and performance benefits, which motivate for algorithm/architecture co-design to favor streaming access patterns, even at the price of a higher order algorithmic complexity. We present the Mondrian Data Engine that drastically improves the runtime and energy efficiency of basic in-memory analytic operators, despite doing more work as compared to traditional CPU-optimized algorithms, which heavily rely on random accesses and deep cache hierarchies


Archive | 2015

Scale-out non-uniform memory access

Stanko Novakovic; Alexandros Daglis; Boris Grot; Edouard Bugnion; Babak Falsafi


arXiv: Hardware Architecture | 2018

Exploiting Errors for Efficiency: A Survey from Circuits to Algorithms

Phillip Stanley-Marbell; Armin Alaghi; Michael Carbin; Eva Darulova; Lara Dolecek; Andreas Gerstlauer; Ghayoor Gillani; Djordje Jevdjic; Thierry Moreau; Mattia Cacciotti; Alexandros Daglis; Natalie D. Enright Jerger; Babak Falsafi; Sasa Misailovic; Adrian Sampson; Damien Zufferey


arXiv: Hardware Architecture | 2018

Enabling Storage Class Memory as a DRAM Replacement for Datacenter Services.

Dmitrii Ustiugov; Alexandros Daglis; Javier Picorel; Mark Sutherland; Edouard Bugnion; Babak Falsafi; Dionisios N. Pnevmatikatos

Collaboration


Dive into the Alexandros Daglis's collaboration.

Top Co-Authors

Avatar

Babak Falsafi

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar

Edouard Bugnion

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar

Boris Grot

University of Edinburgh

View shared research outputs
Top Co-Authors

Avatar

Stanko Novakovic

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar

Dmitrii Ustiugov

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Javier Picorel

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar

Mario Drumond

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar

Nooshin S. Mirzadeh

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar

Djordje Jevdjic

École Polytechnique Fédérale de Lausanne

View shared research outputs
Researchain Logo
Decentralizing Knowledge