Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Luis A. Lastras-Montano is active.

Publication


Featured researches published by Luis A. Lastras-Montano.


high-performance computer architecture | 2010

Improving read performance of Phase Change Memories via Write Cancellation and Write Pausing

Moinuddin K. Qureshi; Michele M. Franceschini; Luis A. Lastras-Montano

Phase Change Memory (PCM) is emerging as a promising technology to build large-scale main memory systems in a cost-effective manner. A characteristic of PCM is that it has write latency much higher than read latency. A higher write latency can typically be tolerated using buffers. However, once a write request is scheduled for service to a bank, it can still cause increased latency for later arriving read requests to the same bank. We show that for the baseline PCM system with read-priority scheduling, the write requests increase the effective read latency to 2.3x (on average), causing significant performance degradation. To reduce the read latency of PCM devices under such scenarios, we propose adaptive Write Cancellation policies. Such policies can abort the processing of a scheduled write requests if a read request arrives to the same bank within a predetermined period. We also propose Write Pausing, which exploits the iterative write algorithms used in PCM to pause at the end of each write iteration to service any pending reads. For the baseline system, the proposed technique removes 75% of the latency increase incurred by read requests and improves overall system performance by 46% (on average), while requiring negligible hardware and simple extensions to PCM controller.


international symposium on computer architecture | 2010

Morphable memory system: a robust architecture for exploiting multi-level phase change memories

Moinuddin K. Qureshi; Michele M. Franceschini; Luis A. Lastras-Montano; John P. Karidis

Phase Change Memory (PCM) is emerging as a scalable and power efficient technology to architect future main memory systems. The scalability of PCM is enhanced by the property that PCM devices can store multiple bits per cell. While such Multi-Level Cell (MLC) devices can offer high density, this benefit comes at the expense of increased read latency, which can cause significant performance degradation. This paper proposes Morphable Memory System (MMS), a robust architecture for efficiently incorporating MLC PCM devices in main memory. MMS is based on observation that memory requirement varies between workloads, and systems are typically over-provisioned in terms of memory capacity. So, during a phase of low memory usage, some of the MLC devices can be operated at fewer bits per cell to obtain lower latency. When the workload requires full memory capacity, these devices can be restored to high density MLC operation to have full main-memory capacity. We provide the runtime monitors, the hardware-OS interface, and the detailed mechanism for implementing MMS. Our evaluations on an 8-core 8GB MLC PCM-based system show that MMS provides, on average, low latency access for 95% of all memory requests, thereby improving overall system performance by 40%.


IEEE Transactions on Information Theory | 2009

On the Redundancy of Slepian–Wolf Coding

Dake He; Luis A. Lastras-Montano; En-hui Yang; Ashish Jagmohan; Jun Chen

In this paper, the redundancy of both variable and fixed rate Slepian-Wolf coding is considered. Given any jointly memoryless source-side information pair {(Xi, Yi)}<sub>i=1</sub> <sup>infin</sup> with finite alphabet, the redundancy R<sup>n</sup>(isin<sub>n</sub>) of variable rate Slepian-Wolf coding of X<sub>1</sub> <sup>n</sup> with decoder only side information Y<sub>1</sub> <sup>n</sup> depends on both the block length n and the decoding block error probability isin<sub>n</sub>, and is defined as the difference between the minimum average compression rate of order n variable rate Slepian-Wolf codes having the decoding block error probability less than or equal to isin<sub>n</sub>, and the conditional entropy H(X|Y), where H(X|Y) is the conditional entropy rate of the source given the side information. The redundancy of fixed rate Slepian-Wolf coding of X<sub>1</sub> <sup>n</sup> with decoder only side information Y<sub>1</sub> <sup>n</sup> is defined similarly and denoted by R<sub>F</sub> <sup>n</sup>(isin<sub>n</sub>). It is proved that under mild assumptions about isin<sub>n</sub>, R<sup>n</sup>(isin<sub>n</sub>) = d<sub>v</sub>radic-log isin<sub>n</sub>/n + (oradic-log isin<sub>n</sub>/n) and R<sub>F</sub> <sup>n</sup>(isin<sub>n</sub>) - d<sub>f</sub>radic-log isin<sub>n</sub>/n + o(radic-log isin<sub>n</sub>/n), where df and dnu are two constants completely determined by the joint distribution of the source-side information pair. Since d<sub>v</sub> is generally smaller than d<sub>f</sub>, our results show that variable rate Slepian-Wolf coding is indeed more efficient than fixed rate Slepian-Wolf coding.


data compression conference | 2005

Distributed source coding in dense sensor networks

Akshay Kashyap; Luis A. Lastras-Montano; Cathy H. Xia; Zhen Liu

We study the problem of the reconstruction of a Gaussian field defined in [0,1] using N sensors deployed at regular intervals. The goal is to quantify the total data rate required for the reconstruction of the field with a given mean square distortion. We consider a class of two-stage mechanisms which (a) send information to allow the reconstruction of the sensors samples within sufficient accuracy, and then (b) use these reconstructions to estimate the entire field. To implement the first stage, the heavy correlation between the sensor samples suggests the use of distributed coding schemes to reduce the total rate. Our main contribution is to demonstrate the existence of a distributed block coding scheme that achieves, for a given fidelity criterion for the sensors measurements, a total information rate that is within a constant, independent of N, of the minimum information rate required by an encoder that has access to all the sensor measurements simultaneously. The constant in general depends on the autocorrelation function of the field and the desired distortion criterion for the sensor samples.


Ibm Journal of Research and Development | 2012

IBM zEnterprise redundant array of independent memory subsystem

Patrick J. Meaney; Luis A. Lastras-Montano; Vesselina K. Papazova; Eldee Stephens; Judy S. Johnson; Luiz C. Alves; James A. O'Connor; William J. Clarke

The IBM zEnterprise® system introduced a new and innovative redundant array of independent memory (RAIM) subsystem design as a standard feature on all zEnterprise servers. It protects the server from single-channel errors such as sudden control, bus, buffer, and massive dynamic RAM (DRAM) failures, thus achieving the highest System z® memory availability. This system also introduced innovations such as DRAM and channel marking, as well as a novel dynamic cyclic redundancy code channel marking. This paper describes this RAIM subsystem and other reliability, availability, and serviceability features, including automatic channel error recovery; data and clock interface lane calibration, recovery, and repair; intermittent lane sparing; and specialty engines for maintenance, periodic calibration, power, and power-on controls.


international symposium on information theory | 2009

On the lifetime of multilevel memories

Luis A. Lastras-Montano; Michele M. Franceschini; Thomas Mittelholzer; John P. Karidis; Mark N. Wegman

We study memories capable of storing multiple bits per memory cell, with the property that certain state transitions “wear” the cell. We introduce a model that is relevant for Phase Change Memory, a promising emerging nonvolatile memory technology that exhibits limitations in the number of particular write actions that one may apply to a cell before rendering it unusable. We exploit the theory of Write Efficient Memories to derive a closed form expression for the storage capacity/lifetime fundamental tradeoff for this model. We then present families of codes specialized to distinct ranges for the target lifetimes, covering the full range from moderate redundancy to an arbitrarily large lifetime increase. These codes have low implementation complexity and remarkably good performance; for example in an 8 level cell we can increase the lifetime of a memory by a factor of ten while sacrificing only 2/3 of the uncoded storage capacity of the memory.


international symposium on information theory | 2010

Algorithms for memories with stuck cells

Luis A. Lastras-Montano; Ashish Jagmohan; Michele M. Franceschini

We present a class of algorithms for encoding data in memories with stuck cells. These algorithms rely on earlier code constructions termed cyclic Partitioned Linear Block Codes. For the corresponding q-ary BCH-like codes for u stucks in a codeword of length n, our encoding algorithm has complexity O((u logq n)2) Fq operations, which we will show compares favorably to a generic approach based on Gaussian elimination. The computational complexity improvements are realized by taking advantage of the algebraic structure of cyclic codes for stucks. The algorithms are also applicable to cyclic codes for both stucks and errors.


global communications conference | 2010

Adaptive endurance coding for NAND Flash

Ashish Jagmohan; Michele M. Franceschini; Luis A. Lastras-Montano; John P. Karidis

A fundamental constraint in the use of newer NAND Flash devices in the enterprise space is the low cycling endurance of such devices. As an example, the latest 2-bit MLC devices have a cycling endurance ranging from 3K to 10K program/erase cycles. Upcoming higher-density devices are expected to have even lower endurance. In this paper we propose a coding technique called Adaptive Endurance Coding (AEC) which increases the number of program/erase cycles that a Flash device can endure. The key insight leveraged by the proposed technique is the data-dependent nature of Flash cell-wear. Data-dependent wear implies that Flash chip/device lifetime can be significantly increased by converting data into bit-patterns, prior to programming, which cause minimal wear. AEC can be used to generate a capacity-wear trade-off; for compressible data, AEC can be adapted to data compressibility in order to maximize endurance gains with low system overhead costs. The technique can be implemented in the Flash device controller without requiring any hardware changes to the device itself. We present empirical results on SLC and MLC Flash chips demonstrating the improvements in retention and bit-error rate which can be obtained via this technique, and present device-level simulation results quantifying the gains achievable by the use of AEC.


IEEE Transactions on Information Theory | 2006

Near sufficiency of random coding for two descriptions

Luis A. Lastras-Montano; Vittorio Castelli

We give a single-letter outer bound for the two-descriptions problem for independent and identically distributed (i.i.d.) sources that is universally close to the El Gamal and Cover (EGC) inner bound. The gaps for the sum and individual rates using a quadratic distortion measure are upper-bounded by 1.5 and 0.5 bits/sample, respectively, and are universal with respect to the source being encoded and the desired distortion levels. Variants of our basic ideas are presented, including upper and lower bounds on the second channels rate when the first channels rate is arbitrarily close to the rate-distortion function; these bounds differ, in the limit as the code block length goes to infinity, by not more than 2 bits/sample. An interesting aspect of our methodology is the manner in which the matching single-letter outer bound is obtained, as we eschew common techniques for constructing single-letter bounds in favor of new ideas in the field of rate loss bounds. We expect these techniques to be generally applicable to other settings of interest.


IEEE Transactions on Information Theory | 2009

On the Linear Codebook-Level Duality Between Slepian–Wolf Coding and Channel Coding

Jun Chen; Dake He; Ashish Jagmohan; Luis A. Lastras-Montano; En-hui Yang

In this paper, it is shown that each Slepian-Wolf coding problem is related to a dual channel coding problem in the sense that the sphere packing exponents, random coding exponents, and correct decoding exponents in these two problems are mirror-symmetrical to each other. This mirror symmetry is interpreted as a manifestation of the linear codebook-level duality between Slepian-Wolf coding and channel coding. Furthermore, this duality, in conjunction with a systematic analysis of the expurgated exponents, reveals that nonlinear Slepian-Wolf codes can strictly outperform linear Slepian-Wolf codes in terms of rate-error tradeoff at high rates. The linear codebook-level duality is also established for general sources and channels.

Researchain Logo
Decentralizing Knowledge