Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Bruce G. Lindsay is active.

Publication


Featured researches published by Bruce G. Lindsay.


ACM Computing Surveys | 1981

The Recovery Manager of the System R Database Manager

Jim Gray; Paul R. McJones; Mike Blasgen; Bruce G. Lindsay; Raymond A. Lorie; Thomas G. Price; Franco Putzolu; Irving L. Traiger

The recovery subsystem of an experimental data management system is described and evaluated. The transactmn concept allows application programs to commit, abort, or partially undo their effects. The DO-UNDO-REDO protocol allows new recoverable types and operations to be added to the recovery system Apphcation programs can record data m the transaction log to facilitate application-specific recovery. Transaction undo and redo are based on records kept in a transaction log. The checkpoint mechanism is based on differential fries (shadows). The recovery log is recorded on disk rather than tape.


ACM Transactions on Database Systems | 1992

ARIES: a transaction recovery method supporting fine-granularity locking and partial rollbacks using write-ahead logging

C. Mohan; Donald J. Haderle; Bruce G. Lindsay; Hamid Pirahesh; Peter M. Schwarz

DB2<supscrpt>TM</supscrpt>, IMS, and Tandem<supscrpt>TM</supscrpt> systems. ARIES is applicable not only to database management systems but also to persistent object-oriented languages, recoverable file systems and transaction-based operating systems. ARIES has been implemented, to varying degrees, in IBMs OS/2<supscrpt>TM</supscrpt> Extended Edition Database Manager, DB2, Workstation Data Save Facility/VM, Starburst and QuickSilver, and in the University of Wisconsins EXODUS and Gamma database machine.


very large data bases | 2001

Efficiently publishing relational data as XML documents

Jayavel Shanmugasundaram; Eugene J. Shekita; Rimon Barr; Michael J. Carey; Bruce G. Lindsay; Hamid Pirahesh; Berthold Reinwald

Abstract. XML is rapidly emerging as a standard for exchanging business data on the World Wide Web. For the foreseeable future, however, most business data will continue to be stored in relational database systems. Consequently, if XML is to fulfill its potential, some mechanism is needed to publish relational data as XML documents. Towards that goal, one of the major challenges is finding a way to efficiently structure and tag data from one or more tables as a hierarchical XML document. Different alternatives are possible depending on when this processing takes place and how much of it is done inside the relational engine. In this paper, we characterize and study the performance of these alternatives. Among other things, we explore the use of new scalar and aggregate functions in SQL for constructing complex XML documents directly in the relational engine. We also explore different execution plans for generating the content of an XML document. The results of an experimental study show that constructing XML documents inside the relational engine can have a significant performance benefit. Our results also show the superiority of having the relational engine use what we call an “outer union plan” to generate the content of an XML document.


ACM Transactions on Database Systems | 1986

Transaction management in the R* distributed database management system

C. Mohan; Bruce G. Lindsay; Ron Obermarck

This paper deals with the transaction management aspects of the R* distributed database system. It concentrates primarily on the description of the R* commit protocols, Presumed Abort (PA) and Presumed Commit (PC). PA and PC are extensions of the well-known, two-phase (2P) commit protocol. PA is optimized for read-only transactions and a class of multisite update transactions, and PC is optimized for other classes of multisite update transactions. The optimizations result in reduced intersite message traffic and log writes, and, consequently, a better response time. The paper also discusses R*s approach toward distributed deadlock detection and resolution.


IEEE Transactions on Knowledge and Data Engineering | 1990

Starburst mid-flight: as the dust clears (database project)

Laura M. Haas; Walter Chang; Guy M. Lohman; John McPherson; Paul F. Wilms; George Lapis; Bruce G. Lindsay; Hamid Pirahesh; Michael J. Carey; Eugene J. Shekita

The purpose of the Starburst project is to improve the design of relational database management systems and enhance their performance, while building an extensible system to better support nontraditional applications and to serve as a testbed for future improvements in database technology. The design and implementation of the Starburst system to date are considered. Some key design decisions and how they affect the goal of improved structure and performance are examined. How well the goal of extensibility has been met is examined: what aspects of the system are extensible, how extensions can be done, and how easy it is to add extensions. Some actual extensions to the system, including the experiences of the first real customizers, are discussed. >


international conference on management of data | 1998

Approximate medians and other quantiles in one pass and with limited memory

Gurmeet Singh Manku; Sridhar Rajagopalan; Bruce G. Lindsay

We present new algorithms for computing approximate quantiles of large datasets in a single pass. The approximation guarantees are explicit, and apply for arbitrary value distributions and arrival distributions of the dataset. The main memory requirements are smaller than those reported earlier by an order of magnitude. We also discuss methods that couple the approximation algorithms with random sampling to further reduce memory requirements. With sampling, the approximation guarantees are explicit but probabilistic, i.e. they apply with respect to a (user controlled) confidence parameter. We present the algorithms, their theoretical analysis and simulation results on different datasets.


Communications of The ACM | 1981

A history and evaluation of System R

Donald D. Chamberlin; Morton M. Astrahan; Michael W. Blasgen; Jim Gray; W. Frank King; Bruce G. Lindsay; Raymond A. Lorie; James W. Mehl; Thomas G. Price; Franco Putzolu; Patricia G. Selinger; Mario Schkolnick; Donald R. Slutz; Irving L. Traiger; Bradford W. Wade; Robert A. Yost

System R, an experimental database system, was constructed to demonstrate that the usability advantages of the relational data model can be realized in a system with the complete function and high performance required for everyday production use. This paper describes the three principal phases of the System R project and discusses some of the lessons learned from System R about the design of relational systems and database systems in general.


international conference on management of data | 1999

Random sampling techniques for space efficient online computation of order statistics of large datasets

Gurmeet Singh Manku; Sridhar Rajagopalan; Bruce G. Lindsay

In a recent paper [MRL98], we had described a general framework for single pass approximate quantile finding algorithms. This framework included several known algorithms as special cases. We had identified a new algorithm, within the framework, which had a significantly smaller requirement for main memory than other known algorithms. In this paper, we address two issues left open in our earlier paper. First, all known and space efficient algorithms for approximate quantile finding require advance knowledge of the length of the input sequence. Many important database applications employing quantiles cannot provide this information. In this paper, we present a novel non-uniform random sampling scheme and an extension of our framework. Together, they form the basis of a new algorithm which computes approximate quantiles without knowing the input sequence length. Second, if the desired quantile is an extreme value (e.g., within the top 1% of the elements), the space requirements of currently known algorithms are overly pessimistic. We provide a simple algorithm which estimates extreme values using less space than required by the earlier more general technique for computing all quantiles. Our principal observation here is that random sampling is quantifiably better when estimating extreme values than is the case with the median.


ACM Transactions on Database Systems | 1982

Transactions and consistency in distributed database systems

Irving L. Traiger; Jim Gray; Cesare A. Galtieri; Bruce G. Lindsay

The concepts of transaction and of data consistency are defined for a distributed system. The cases of partitioned data, where fragments of a file are stored at multiple nodes, and replicated data, where a file is replicated at several nodes, are discussed. It is argued that the distribution and replication of data should be transparent to the programs which use the data. That is, the programming interface should provide location transparency, replica transparency, concurrency transparency, and failure transparency. Techniques for providing such transparencies are abstracted and discussed. By extending the notions of system schedule and system clock to handle multiple nodes, it is shown that a distributed system can be modeled as a single sequential execution sequence. This model is then used to discuss simple techniques for implementing the various forms of transparency.


international conference on management of data | 1986

A snapshot differential refresh algorithm

Bruce G. Lindsay; Laura M. Haas; C. Mohan; Hamid Pirahesh; Paul F. Wilms

This article presents an algorithm to refresh the contents of database snapshots. A database snapshot is a read-only table whose contents are extracted from other tables in the database. The snapshot contents can be periodically refreshed to reflect the current state of the database. Snapshots are useful in many applications as a cost effective substitute for replicated data in a distributed database system. When the snapshot contents are a simple restriction and projection of a single base table, differential refresh techniques can reduce the message and update costs of the snapshot refresh operation. The algorithm presented annotates the base table to detect the changes which must be applied to the snapshot table during snapshot refresh. The cost of maintaining the base table annotations is minimal and the amount of data transmitted during snapshot refresh is close to optimal in most circumstances.

Collaboration


Dive into the Bruce G. Lindsay's collaboration.

Researchain Logo
Decentralizing Knowledge