Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Mingdi Xue is active.

Publication


Featured researches published by Mingdi Xue.


ieee conference on mass storage systems and technologies | 2016

Fine-grained metadata journaling on NVM

Cheng Chen; Jun Yang; Qingsong Wei; Chundong Wang; Mingdi Xue

Journaling file systems have been widely used where data consistency must be assured. However, we observed that the overhead of journaling can cause up to 48.2% performance drop under certain kinds of workloads. On the other hand, the emerging high-performance, byte-addressable Non-volatile Memory (NVM) has the potential to minimize such overhead by being used as the journal device. The traditional journaling mechanism based on block devices is nevertheless unsuitable for NVM due to the write amplification of metadata journal we observed. In this paper, we propose a fine-grained metadata journal mechanism to fully utilize the low-latency byte-addressable NVM so that the overhead of journaling can be significantly reduced. Based on the observation that conventional block-based metadata journal contains up to 90% clean metadata that is unnecessary to be journalled, we design a fine-grained journal format for byte-addressable NVM which contains only modified metadata. Moreover, we redesign the process of transaction committing, checkpointing and recovery in journaling file systems utilizing the new journal format. Therefore, thanks to the reduced amount of ordered writes to NVM, the overhead of journaling can be reduced without compromising the file system consistency. Experimental results show that our NVM-based fine-grained metadata journaling is up to 15.8× faster than the traditional approach under FileBench workloads.


networking architecture and storages | 2015

How to be consistent with persistent memory? An evaluation approach

Chundong Wang; Qingsong Wei; Jun Yang; Cheng Chen; Mingdi Xue

The advent of the byte-addressable, non-volatile memory (NVM) has initiated the design of new data management strategies to utilize it as the persistent memory (PM). One way to manage the PM is via an in-memory file system. The consistency of the in-memory file system may nevertheless be compromised from directly exposing the PM to the CPU, because data are likely to be flushed from the CPU cache to the PM in an order that is different from the order in which they have been programed to be. As a result, in spite of classic consistency mechanisms, such as journaling and Copy-on-Write, file systems for the PM have to seek support of cacheline flush and memory fence instructions, e.g., clflush and sfence, to achieve ordered writes. On the other hand, manipulating the PM as a consistent block device with conventional file systems is also doable. The pros and cons of two approaches, however, have not been thoroughly investigated yet. We hence do so with extensive evaluations and detailed analyses. Our aim of this paper is to inspire how the PM shall be managed, especially from the performance perspective.


ieee international conference on high performance computing data and analytics | 2017

Transactional NVM cache with high performance and crash consistency

Qingsong Wei; Chundong Wang; Cheng Chen; Yechao Yang; Jun Yang; Mingdi Xue

The byte-addressable non-volatile memory (NVM) is new promising storage medium. Compared to NAND flash memory, the next-generation NVM not only preserves the durability of stored data but has much shorter access latencies. An architect can utilize the fast and persistent NVM as an external disk cache. Regarding the systems crash consistency, a prevalent journaling file system needs to run atop an NVM disk cache. However, the performance is severely impaired by redundant efforts in achieving crash consistency in both file system and disk cache. Therefore, we propose a new mechanism called transactional NVM disk cache (Tinca). In brief, Tinca jointly guarantees consistency of file system and disk cache and removes the performance penalty of file system journaling with a lightweight transaction scheme. Evaluations confirm that Tinca significantly outperforms state-of-the-art design by up to 2.5X in local and cluster tests without causing any inconsistency issue.


international conference on parallel and distributed systems | 2015

Accelerating Cloud Storage System with Byte-Addressable Non-Volatile Memory

Qingsong Wei; Mingdi Xue; Jun Yang; Chundong Wang; Chen Cheng

As building block for cloud storage, distributed file system uses underlying local file systems to manage objects. However, the underlying file system, which is limited by metadata and journaling I/O, significantly affects the performance of the distributed file system. This paper presents an NVM-based file system (referred to as NV-Booster) to accelerate object access for storage node. The NV-Booster leverages byte-addressability and persistency of nonvolatile memory (NVM) to speedup metadata accesses and file system journaling. With NV-Booster, metadata is kept in NVM and accessed in byte-addressable manner through memory bus, while object is stored on hard disk and accessed from I/O bus. In addition, proposed NV-Booster enables fast object search and mapping between object ID and on-disk location with an efficient in-memory namespace management. NV-Booster is implemented in kernel space with NVDIMM and has been extensively evaluated under various workloads. Our experiments show that NV-Booster improves Ceph performance up to 10X, compared to the Ceph with existing local file systems.


ACM Transactions on Storage | 2015

Z-MAP: A Zone-Based Flash Translation Layer with Workload Classification for Solid-State Drive

Qingsong Wei; Cheng Chen; Mingdi Xue; Jun Yang

Existing space management and address mapping schemes for flash-based Solid-State-Drive (SSD) operate either at page or block granularity, with inevitable limitations in terms of memory requirement, performance, garbage collection, and scalability. To overcome these limitations, we proposed a novel space management and address mapping scheme for flash referred to as Z-MAP, which manages flash space at granularity of Zone. Each Zone consists of multiple numbers of flash blocks. Leveraging workload classification, Z-MAP explores Page-mapping Zone (Page Zone) to store random data and handle a large number of partial updates, and Block-mapping Zone (Block Zone) to store sequential data and lower the overall mapping table. Zones are dynamically allocated and a mapping scheme for a Zone is determined only when it is allocated. Z-MAP uses a small part of Flash memory or phase change memory as a streaming Buffer Zone to log data sequentially and migrate data into Page Zone or Block Zone based on workload classification. A two-level address mapping is designed to reduce the overall mapping table and address translation latency. Z-MAP classifies data before it is permanently stored into Flash memory so that different workloads can be isolated and garbage collection overhead can be minimized. Z-MAP has been extensively evaluated by trace-driven simulation and a prototype implementation on OpenSSD. Our benchmark results conclusively demonstrate that Z-MAP can achieve up to 76% performance improvement, 81% mapping table reduction, and 88% garbage collection overhead reduction compared to existing Flash Translation Layer (FTL) schemes.


IEEE Transactions on Computers | 2018

Dynamic Scheduling with Service Curve for QoS Guarantee of Large-Scale Cloud Storage

Yu Zhang; Qingsong Wei; Cheng Chen; Mingdi Xue; Xinkun Yuan; Chundong Wang

With the growing popularity of cloud storage, more and more diverse applications with diverse service level agreements (SLAs) are being accommodated into it. The quality of service (QoS) support for applications in a shared cloud storage becomes important. However, performance isolation, diverse performance requirements, especially harsh latency guarantees and high system utilization, are all challenging and desirable for QoS design. In this paper, we propose a service curve-based QoS algorithm to support latency guarantee applications, IOPS guarantee applications and best-effort applications at the same storage system, which not only provides a QoS guarantee for applications, but also pursues better system utilization. Three priority queues are exploited and different service curves are applied for different types of applications. I/O requests from different applications are scheduled and dispatched among the three queues according to their service curves and I/O urgency status, so that QoS requirements of all applications can be guaranteed on the shared storage system. Our experimental results show that our algorithm not only simultaneously guarantees the QoS targets of latency and throughput (IOPS), but also improves the utilization of storage resources.


ACM Transactions on Storage | 2018

Persisting RB-Tree into NVM in a Consistency Perspective

Chundong Wang; Qingsong Wei; Lingkun Wu; Sibo Wang; Cheng Chen; Xiaokui Xiao; Jun Yang; Mingdi Xue; Yechao Yang

Byte-addressable non-volatile memory (NVM) is going to reshape conventional computer systems. With advantages of low latency, byte-addressability, and non-volatility, NVM can be directly put on the memory bus to replace DRAM. As a result, both system and application softwares have to be adjusted to perceive the fact that the persistent layer moves up to the memory. However, most of the current in-memory data structures will be problematic with consistency issues if not well tuned with NVM. This article places emphasis on an important in-memory structure that is widely used in computer systems, i.e., the Red/Black-tree (RB-tree). Since it has a long and complicated update process, the RB-tree is prone to inconsistency problems with NVM. This article presents an NVM-compatible consistent RB-tree with a new technique named cascade-versioning. The proposed RB-tree (i) is all-time consistent and scalable and (ii) needs no recovery procedure after system crashes. Experiment results show that the RB-tree for NVM not only achieves the aim of consistency with insignificant spatial overhead but also yields comparable performance to an ordinary volatile RB-tree.


international conference on cloud computing | 2016

NVM-Accelerated Metadata Management for Flash-Based SSDs

Mingdi Xue; Chundong Wang; Qingsong Wei; Jun Yang; Cheng Chen

Data storage system is a de facto building brick for the cloud framework. The advent of byte-addressable non-volatile memory (NVM) brings in new opportunities to optimize conventional storage systems. In this paper, we propose to utilize the NVM to separate and manage metadata of file systems for flash-based solid-state drives (SSDs). The SSD has made impressive strides in the persistent storage of enterprise servers, personal computers and mobile devices. However, it suffers fromhandling the file system metadata which are mostly small andfrequently used. We hence separate metadata from data and give the former a second chance to be cached in a NVM-based buffer cache. To closely collaborate with metadata management modules, the flash translation layer (FTL) of the SSD is also promoted to the NVM. Experimental results based on Linux and a real SSD board show that, our proposal is able to evidently boost the access performance of the SSD.


2015 IEEE Non-Volatile Memory System and Applications Symposium (NVMSA) | 2015

Data-centric garbage collection for NAND flash devices

Chundong Wang; Qingsong Wei; Mingdi Xue; Jun Yang; Cheng Chen

Garbage collection has been concerned for NAND flash devices for years. The ever-increasing utilization of flash device demands more effective and efficient garbage collection strategies. This paper proposes a novel approach, namely Data-centrIc Garbage collection (DIG). DIG online forecasts update intervals for data and clusters them accordingly into groups in a lightweight way. Data with similar update intervals form a group and are stored together. Obsolete data and valid data are hence prevented from being mixed. Moreover, DIG takes advantage of clustering to further separate data and select promising victims for reclamations. Experiments show that DIG can significantly reduce the overheads of garbage collection by 94.3% and 73.5% on average, respectively, compared to two state-of-the-art algorithms.


ACM Transactions on Storage | 2017

Optimizing File Systems with Fine-grained Metadata Journaling on Byte-addressable NVM

Cheng Chen; Jun Yang; Qingsong Wei; Chundong Wang; Mingdi Xue

Journaling file systems have been widely adopted to support applications that demand data consistency. However, we observed that the overhead of journaling can cause up to 48.2% performance drop under certain kinds of workloads. On the other hand, the emerging high-performance, byte-addressable Non-volatile Memory (NVM) has the potential to minimize such overhead by being used as the journal device. The traditional journaling mechanism based on block devices is nevertheless unsuitable for NVM due to the write amplification of metadata journal we observed. In this article, we propose a fine-grained metadata journal mechanism to fully utilize the low-latency byte-addressable NVM so that the overhead of journaling can be significantly reduced. Based on the observation that conventional block-based metadata journal contains up to 90% clean metadata that is unnecessary to be journalled, we design a fine-grained journal format for byte-addressable NVM which contains only modified metadata. Moreover, we redesign the process of transaction committing, checkpointing, and recovery in journaling file systems utilizing the new journal format. Therefore, thanks to the reduced amount of ordered writes for journals, the overhead of journaling can be reduced without compromising the file system consistency. To evaluate our fine-grained metadata journaling mechanism, we have implemented a journaling file system prototype based on Ext4 and JBD2 in Linux. Experimental results show that our NVM-based fine-grained metadata journaling is up to 15.8 × faster than the traditional approach under FileBench workloads.

Collaboration


Dive into the Mingdi Xue's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Cheng Chen

Data Storage Institute

View shared research outputs
Top Co-Authors

Avatar

Jun Yang

Data Storage Institute

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Yechao Yang

Data Storage Institute

View shared research outputs
Top Co-Authors

Avatar

Chen Cheng

Data Storage Institute

View shared research outputs
Top Co-Authors

Avatar

Lingkun Wu

Data Storage Institute

View shared research outputs
Top Co-Authors

Avatar

Sibo Wang

Nanyang Technological University

View shared research outputs
Top Co-Authors

Avatar

Xiaokui Xiao

Nanyang Technological University

View shared research outputs
Top Co-Authors

Avatar

Yu Zhang

Data Storage Institute

View shared research outputs
Researchain Logo
Decentralizing Knowledge