Vijay Balakrishnan
Samsung
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Vijay Balakrishnan.
acm international conference on systems and storage | 2015
Qiumin Xu; Huzefa Siyamwala; Mrinmoy Ghosh; Tameesh Suri; Manu Awasthi; Zvika Guz; Anahita Shayesteh; Vijay Balakrishnan
The storage subsystem has undergone tremendous innovation in order to keep up with the ever-increasing demand for throughput. Non Volatile Memory Express (NVMe) based solid state devices are the latest development in this domain, delivering unprecedented performance in terms of latency and peak bandwidth. NVMe drives are expected to be particularly beneficial for I/O intensive applications, with databases being one of the prominent use-cases. This paper provides the first, in-depth performance analysis of NVMe drives. Combining driver instrumentation with system monitoring tools, we present a breakdown of access times for I/O requests throughout the entire system. Furthermore, we present a detailed, quantitative analysis of all the factors contributing to the low-latency, high-throughput characteristics of NVMe drives, including the system software stack. Lastly, we characterize the performance of multiple cloud databases (both relational and NoSQL) on state-of-the-art NVMe drives, and compare that to their performance on enterprise-class SATA-based SSDs. We show that NVMe-backed database applications deliver up to 8× superior client-side performance over enterprise-class, SATA-based SSDs.
international performance computing and communications conference | 2016
Janki Bhimani; Jingpei Yang; Zhengyu Yang; Ningfang Mi; Qiumin Xu; Manu Awasthi; Rajinikanth Pandurangan; Vijay Balakrishnan
Our cloud-based IT world is founded on hyper-visors and containers. Containers are becoming an important cornerstone, which is increasingly used day-by-day. Among different available frameworks, docker has become one of the major adoptees to use containerized platform in data centers and enterprise servers, due to its ease of deploying and scaling. Further more, the performance benefits of a lightweight container platform can be leveraged even more with a fast back-end storage like high performance SSDs. However, increase in number of simultaneously operating docker containers may not guarantee an aggregated performance improvement due to saturation. Thus, understanding performance bottleneck in a multi-tenancy docker environment is critically important to maintain application level fairness and perform better resource management. In this paper, we characterize the performance of persistent storage option (through data volume) for I/O intensive, dockerized applications. Our work investigates the impact on performance with increasing number of simultaneous docker containers in different workload environments. We provide, first of its kind study of I/O intensive containerized applications operating with NVMe SSDs. We show that 1) a six times better application throughput can be obtained, just by wise selection of number of containerized instances compared to single instance; and 2) for multiple application containers running simultaneously, an application throughput may degrade upto 50% compared to a stand-alone applications throughput, if good choice of application and workload is not made. We then propose novel design guidelines for an optimal and fair operation of both homogeneous and heterogeneous environments mixed with different applications and workloads.
measurement and modeling of computer systems | 2015
Qiumin Xu; Huzefa Siyamwala; Mrinmoy Ghosh; Manu Awasthi; Tameesh Suri; Zvika Guz; Anahita Shayesteh; Vijay Balakrishnan
The storage subsystem has undergone tremendous innovation in order to keep up with the ever-increasing demand for throughput. NVMe based SSDs are the latest development in this domain, delivering unprecedented performance in terms of both latency and peak bandwidth. Given their superior performance, NVMe drives are expected to be particularly beneficial for I/O intensive applications in datacenter installations. In this paper we identify and analyze the different factors leading to the better performance of NVMe SSDs. Then, using databases as the prominent use-case, we show how these would translate into real-world benefits. We evaluate both a relational database (MySQL) and a NoSQL database (Cassandra) and demonstrate significant performance gains over best-in-class enterprise SATA SSDs: from 3.5x for TPC-C and up to 8.5x for Cassandra.
acm international conference on systems and storage | 2017
Zvika Guz; Harry (Huan) Li; Anahita Shayesteh; Vijay Balakrishnan
Storage disaggregation separates compute and storage to different nodes in order to allow for independent resource scaling and thus, better hardware resource utilization. While disaggregation of hard-drives storage is a common practice, NVMe-SSD (i.e., PCIe-based SSD) disaggregation is considered more challenging. This is because SSDs are significantly faster than hard drives, so the latency overheads (due to both network and CPU processing) as well as the extra compute cycles needed for the offloading stack become much more pronounced. In this work we characterize the overheads of NVMe-SSD disaggregation. We show that NVMe-over-Fabrics (NVMf) - a recently-released remote storage protocol specification - reduces the overheads of remote access to a bare minimum, thus greatly increasing the cost-efficiency of Flash disaggregation. Specifically, while recent work showed that SSD storage disaggregation via iSCSI degrades application-level throughput by 20%, we report on negligible performance degradation with NVMf - both when using stress-tests as well as with a more-realistic KV-store workload.
acm international conference on systems and storage | 2017
Jingpei Yang; Rajinikanth Pandurangan; Changho Choi; Vijay Balakrishnan
Multi-stream SSDs can isolate data with different life time to disparate erase blocks, thus reduce garbage collection overhead and improve overall SSD performance. Applications are responsible for management of these device-level steams such as stream open/close and data-to-stream mapping. This requires application changes, and the engineer deploying the solution needs to be able to individually identify the streams in their workload. Furthermore, when multiple applications are involved, such as in VM or containerized environments, stream management becomes more complex due to the limited number of streams a device can support, for example, allocating streams to applications or sharing streams across applications will cause additional overhead. To address these issues and reduce the overhead of stream management, this paper proposes automatic stream management algorithms that operate under the application layer. Our stream assignment techniques, called AutoStream, is based on run time workload detection and independent of the application(s). We implement our AutoStream prototype in NVMe Linux device driver and our performance evaluation shows up to 60% reduction on WAF(Write Amplification Factor) and up to 237% improvement on performance compared to a conventional SSD device.
international conference on performance engineering | 2015
Manu Awasthi; Tameesh Suri; Zvika Guz; Anahita Shayesteh; Mrinmoy Ghosh; Vijay Balakrishnan
IEEE Transactions on Multi-Scale Computing Systems | 2018
Janki Bhimani; Zhengyu Yang; Ningfang Mi; Jingpei Yang; Qiumin Xu; Manu Awasthi; Rajinikanth Pandurangan; Vijay Balakrishnan
international performance computing and communications conference | 2017
Janki Bhimani; Jingpei Yang; Zhengyu Yang; Ningfang Mi; N. H. V. Krishna Giri; Rajinikanth Pandurangan; Changho Choi; Vijay Balakrishnan
international conference on cloud computing | 2018
Janki Bhimani; Ningfang Mi; Zhengyu Yang; Jingpei Yang; Rajinikanth Pandurangan; Changho Choi; Vijay Balakrishnan
Archive | 2017
Narges Shahidi; Manu Awasthi; Tameesh Suri; Vijay Balakrishnan