Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Thomas J. E. Schwarz is active.

Publication


Featured researches published by Thomas J. E. Schwarz.


ieee conference on mass storage systems and technologies | 2003

Reliability mechanisms for very large storage systems

Qin Xin; Ethan L. Miller; Thomas J. E. Schwarz; Darrell D. E. Long; Scott A. Brandt; Witold Litwin

Reliability and availability are increasingly important in large-scale storage systems built from thousands of individual storage devices. Large systems must survive the failure of individual components; in systems with thousands of disks, even infrequent failures are likely in some device. We focus on two types of errors: nonrecoverable read errors and drive failures. We discuss mechanisms for detecting and recovering from such errors, introducing improved techniques for detecting errors in disk reads and fast recovery from disk failure. We show that simple RAID cannot guarantee sufficient reliability; our analysis examines the tradeoffs among other schemes between system availability and storage efficiency. Based on our data, we believe that two-way mirroring should be sufficient for most large storage systems. For those that need very high reliability, we recommend either three-way mirroring or mirroring combined with RAID.


modeling, analysis, and simulation on computer and telecommunication systems | 2004

Disk scrubbing in large archival storage systems

Thomas J. E. Schwarz; Qin Xin; Ethan L. Miller; Darrell D. E. Long; Andy Hospodor; Spencer W. Ng

Large archival storage systems experience long periods of idleness broken up by rare data accesses. In such systems, disks may remain powered off for long periods of time. These systems can lose data for a variety of reasons, including failures at both the device level and the block level. To deal with these failures, we must detect them early enough to be able to use the redundancy built into the storage system. We propose a process called disk scrubbing in a system in which drives are periodically accessed to detect drive failure. By scrubbing all of the data stored on all of the disks, we can detect block failures and compensate for them by rebuilding the affected blocks. Our research shows how the scheduling of disk scrubbing affects overall system reliability, and that opportunistic scrubbing, in which the system scrubs disks only when they are powered on for other reasons, performs very well without the need to power on disks solely to check them.


international conference on management of data | 2000

LH*RS: a high-availability scalable distributed data structure using Reed Solomon Codes

Witold Litwin; Thomas J. E. Schwarz

LH*RS is a new high-availability Scalable Distributed Data Structure (SDDS). The data storage scheme and the search performance of LH*RS are basically these of LH*. LH*RS manages in addition the parity information to tolerate the unavailability of k ⪈ 1 server sites. The value of k scales with the file, to prevent the reliability decline. The parity calculus uses the Reed -Solomon Codes. The storage and access performance overheads to provide the high-availability are about the smallest possible. The scheme should prove attractive to data-intensive applications.


symposium on large spatial databases | 2001

A Model-Based, Open Architecture for Mobile, Spatially Aware Applications

Daniela Nicklas; Matthias Großmann; Thomas J. E. Schwarz; Steffen Volz; Bernhard Mitschang

With the emerging availability of small and portable devices that are able to determine their position and to communicate wirelessly, mobile and spatially aware applications become feasible. These applications rely on information that is bound to locations. In this paper we present NEXUS, a platform for such applications, which is open for both new applications and new information providers, similar to the World Wide Web. Distributed servers provide location-based information, which is federated to an integrated view for the applications. To achieve this goal, we present the concept of the Augmented World Model, which is a common data model for location-based information. We give an example to show how applications can use this platform.


modeling, analysis, and simulation on computer and telecommunication systems | 2006

Providing High Reliability in a Minimum Redundancy Archival Storage System

Deepavali Bhagwat; Kristal T. Pollack; Darrell D. E. Long; Thomas J. E. Schwarz; Ethan L. Miller; Jehan-Francois Paris

Inter-file compression techniques store files as sets of references to data objects or chunks that can be shared among many files. While these techniques can achieve much better compression ratios than conventional intra-file compression methods such as Lempel-Ziv compression, they also reduce the reliability of the storage system because the loss of a few critical chunks can lead to the loss of many files. We show how to eliminate this problem by choosing for each chunk a replication level that is a function of the amount of data that would be lost if that chunk were lost. Experiments using actual archival data show that our technique can achieve significantly higher robustness than a conventional approach combining data mirroring and intra-file compression while requiring about half the storage space.


international conference on data engineering | 2004

Algebraic signatures for scalable distributed data structures

Witold Litwin; Thomas J. E. Schwarz

Signatures detect changes to data objects. Numerous schemes are in use, especially the cryptographically secure standards SHA-1. We propose a novel signature scheme which we call algebraic signatures. The scheme uses the Galois field calculations. Its major property is the sure detection of any changes up to a parameterized size. More precisely, we detect for sure any changes that do not exceed n-symbols for an n-symbol algebraic signature. This property is new for any known signature scheme. For larger changes, the collision probability is typically negligible, as for the other known schemes. We apply the algebraic signatures to the scalable distributed data structures (SDDS). We filter at the SDDS client node the updates that do not actually change the records. We also manage the concurrent updates to data stored in the SDDS RAM buckets at the server nodes. We further use the scheme for the fast disk backup of these buckets. We sign our objects with 4-byte signatures, instead of 20-byte standard SHA-1 signatures. Our algebraic calculus is then also about twice as fast.


ACM Transactions on Database Systems | 2005

LH* RS ---a highly-available scalable distributed data structure

Witold Litwin; Rim Moussa; Thomas J. E. Schwarz

LH*RS is a high-availability scalable distributed data structure (SDDS). An LH*RS file is hash partitioned over the distributed RAM of a multicomputer, for example, a network of PCs, and supports the unavailability of any k ≥ 1 of its server nodes. The value of k transparently grows with the file to offset the reliability decline. Only the number of the storage nodes potentially limits the file growth. The high-availability management uses a novel parity calculus that we have developed, based on Reed-Salomon erasure correcting coding. The resulting parity storage overhead is about the lowest possible. The parity encoding and decoding are faster than for any other candidate coding we are aware of. We present our scheme and its performance analysis, including experiments with a prototype implementation on Wintel PCs. The capabilities of LH*RS offer new perspectives to data intensive applications, including the emerging ones of grids and of P2P computing.


IEEE Transactions on Magnetics | 2011

Data Management and Layout for Shingled Magnetic Recording

Ahmed Amer; JoAnne Holliday; Darrell D. E. Long; Ethan L. Miller; Jehan-Francois Paris; Thomas J. E. Schwarz

Ultimately the performance and success of a shingled write disk (SWD) will be determined by more than the physical hardware realized, but will depend on the data layouts employed, the workloads experienced, and the architecture of the overall system, including the level of interfaces provided by the devices to higher levels of system software. While we discuss several alternative layouts for use with SWD, we also discuss the dramatic implications of observed workloads. Example data access traces demonstrate the surprising stability of written device blocks, with a small fraction requiring multiple updates (the problematic operation for a shingled-write device). Specifically, we discuss how general purpose workloads can show that more than 93% of device blocks can remain unchanged over a day, and that for more specialized workloads less than 0.5% of a shingled-write disks capacity would be needed to hold randomly updated blocks. We further demonstrate how different approaches to data layout can alternatively improve or reduce the performance of a shingled-write device in comparison to the performance of a traditional non-shingled device.


international conference on distributed computing systems | 1992

RAID organization and performance

Thomas J. E. Schwarz; W.A. Buckhard

A disk array architecture that generalizes the RAID (redundant arrays of inexpensive disks) level V data organization while providing excellent storage utilization, response times, and fault tolerance is discussed. A key feature of the approach is that reliability groups can contain several check data disks beyond the single parity disk. RAID response times for fault-free and failure recovery operations are presented.<<ETX>>


modeling, analysis, and simulation on computer and telecommunication systems | 2005

Disk infant mortality in large storage systems

Qin Xin; Thomas J. E. Schwarz; Ethan L. Miller

As disk drives have dropped in price relative to tape, the desire for the convenience and speed of online access to large data repositories has, led to the deployment of petabyte-scale disk farms with thousands of disks. Unfortunately, the very large size of these repositories renders them vulnerable to previously rare failure modes such as multiple, unrelated disk failures leading to data loss. While some business models, such as free email servers, may be able to tolerate some occurrence of data loss, others, including premium online services and storage of simulation results at a national laboratory, cannot. This paper describes the effect of infant mortality on long-term failure rates of systems that must preserve their data for decades. Our failure models incorporate the well-known bathtub curve, which reflects the higher failure rates of new disk drives, a lower, constant failure rate during the remainder of the design life span, and increased failure rates as components wear out. Large systems are vulnerable to the cohort effect that occurs when many disks are simultaneously replaced by new disks. Our more accurate disk models and simulations have yielded predictions of system lifetimes that are more pessimistic than existing models that assume a constant disk failure rate. Thus, larger system scale requires designers to take disk infant mortality into account.

Collaboration


Dive into the Thomas J. E. Schwarz's collaboration.

Top Co-Authors

Avatar

Witold Litwin

Paris Dauphine University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Ahmed Amer

Santa Clara University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Qin Xin

University of California

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge