Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Brian F. Cooper is active.

Publication


Featured researches published by Brian F. Cooper.


ACM Transactions on Information Systems | 2002

Peer-to-peer data trading to preserve information

Brian F. Cooper; Hector Garcia-Molina

Data archiving systems rely on replication to preserve information. This paper discusses how a network of autonomous archiving sites can trade data to achieve the most reliable replication. A series of binary trades among sites produces a peer-to-peer archiving network. Two trading algorithms are examined, one based on trading collections (even if they are different sizes) and another based on trading equal sized blocks of space (which can then store collections). The concept of deeds is introduced; deeds track the blocks of space owned by one site at another. Policies for tuning these algorithms to provide the highest reliability, for example by changing the order in which sites are contacted and offered trades, are discussed. Finally, simulation results are presented that reveal which policies are best. The experiments indicate that a digital archive can achieve the best reliability by trading blocks of space (deeds), and that following certain policies will allow that site to maximize its reliability.


international conference on distributed computing systems | 2002

Bidding for storage space in a peer-to-peer data preservation system

Brian F. Cooper; Hector Garcia-Molina

Digital archives protect important data collections from failures by making multiple copies at other archives, so that there are always several good copies of a collection. In a cooperative replication network sites trade space, so that each site contributes storage resources to the system and uses storage resources at other sites. Here, we examine bid trading: a mechanism where sites conduct auctions to determine who to trade with. A local site wishing to make a copy of a collection announces how much remote space is needed, and accepts bids for how much of its own space the local site must pay to acquire that remote space. We examine the best policies for determining when to call auctions and how much to bid, as well as the effects of maverick sites that attempt to subvert the bidding system. Simulations of auction and trading sessions indicate that bid trading can allow sites to achieve higher reliability than the alternative: a system where sites trade equal amounts of space without bidding.


european conference on research and advanced technology for digital libraries | 2000

Implementing a Reliable Digital Object Archive

Brian F. Cooper; Arturo Crespo; Hector Garcia-Molina

An Archival Repository reliably stores digital objects for long periods of time (decades or centuries). The archival nature of the system requires new techniques for storing, indexing, and replicating digital objects. In this paper we discuss the specialized indexing needs of a write-once archive. We also present a reliability algorithm for effectively replicating sets of related objects. We describe a data import utility for archival repositories. Finally, we discuss and evaluate a prototype repository we have built, the Stanford Archival Vault (SAV).


international conference on management of data | 2003

Peer-to-peer research at Stanford

Mayank Bawa; Brian F. Cooper; Arturo Crespo; Neil Daswani; Prasanna Ganesan; Hector Garcia-Molina; Sepandar D. Kamvar; Sergio Marti; Mario T. Schlosser; Qi Sun; Patrick Vinograd; Beverly Yang

n this paper we present recent and ongoing research projects of the Peers research group at Stanford University.


acm/ieee joint conference on digital libraries | 2001

Creating trading networks of digital archives

Brian F. Cooper; Hector Garcia

Digital archives can best survive failures if they have made several c opies of their collections at remote sites. In this paper, we discuss how autonomous sites can cooperate to provide preservation by trading data. We examine the decisions that an archive must make when forming trading networks, such as the amount of storage space to provide and the best number of partner sites. We also deal with the fact that some sites may be more reliable than others. Experimental results from a data trading simulator illustrate which policies are most reliable. Our techniques focus on preserving the ``bits of digital collections; other services that focus on other archiving concerns (such as preserving meaningful metadata) can be built on top of the system we describe here.


international workshop on peer to peer systems | 2002

Peer-to-Peer Resource Trading in a Reliable Distributed System

Brian F. Cooper; Hector Garcia-Molina

Peer-to-peer architectures can be used to build a robust, fault tolerant infrastructure for important services. One example is a peer-to-peer data replication system, in which digital collections are protected from failure by being replicated at multiple peers. We argue that such community-based redundancy, in which multiple sites contribute resources to build a fault-tolerant system, is an important application of peer-to-peer networking. In such a system, there must be flexible, effective techniques for managing resource allocation. We propose data trading, a mechanism where a site acquires remote resources in the community by trading away its own local resources. We discuss the application of data trading to the data replication problem, and examine other applications of trading. A general trading infrastructure is a valuable part of a peer-to-peer, community-based redundancy system.


international workshop on peer-to-peer systems | 2003

Studying Search Networks with SIL

Brian F. Cooper; Hector Garcia-Molina

We present a general model, called the Search/Index Link (SIL) model, for studying peer-to-peer search networks. This model allows us to analyze and visualize existing network architectures. It also allows us to discover novel architectures that have desirable properties. Finally, it can be used as a starting point for developing new network construction techniques.


International Journal on Digital Libraries | 2005

InfoMonitor: unobtrusively archiving a World Wide Web server

Brian F. Cooper; Hector Garcia-Molina

It is important to provide long-term preservation of digital data even when those data are stored in an unreliable system such as a filesystem, a legacy database, or even the World Wide Web. In this paper we focus on the problem of archiving the contents of a Web site without disrupting users who maintain the site. We propose an archival storage system, the InfoMonitor, in which a reliable archive is integrated with an unmodified existing store. Implementing such a system presents various challenges related to the mismatch of features between the components such as differences in naming and data manipulation operations. We examine each of these issues as well as solutions for the conflicts that arise. We also discuss our experience using the InfoMonitor to archive the Stanford Database Group’s Web site.


acm symposium on applied computing | 2002

A parallel index for semistructured data

Brian F. Cooper; Neal Sample; Moshe Shadmon

Database systems are increasingly being used to manage semistructured data, which may not have a fixed structure or set of relationships between data items. Indexes which use tree structures to manage semistructured data become unbalanced and difficult to parallelize due to the complex nature of the data. We propose a mechanism by which an unbalanced vertical tree is managed in a balanced way by additional layers of horizontal index. Then, the vertical tree can be partitioned among parallel computing nodes in a balanced fashion. We discuss how to construct, search and update such a horizontal structure using the example of a Patricia trie index. We also present simulation results that demonstrate the speedup offered by such parallelism, for example, with three-way parallelism, our techniques can provide almost a factor of three speedup.


international workshop on research issues in data engineering | 2002

Extensible data management in the middle-tier

Brian F. Cooper; Neal Sample; Michael J. Franklin; J. Olshansky; Moshe Shadmon; L. Cohen

Current data management solutions are optimized for intra-enterprise, client-server applications. They depend on predictability, predefined structure and universal administrative control, and cannot easily cope with change and lack of structure. However, modern e-commerce applications are dynamic, unpredictable, organic, and decentralized, and require adaptability. eXtensible Data Management (XDM) is a new approach that enables rapid development and deployment of networked, data-intensive services by providing semantically-rich, high-performance middle-tier data management, and allows heterogeneous data from different sources to be accessed in a uniform manner. We discuss how middle tier extensible data management can benefit an enterprise, and present technical details and examples from the Index Fabric, an XDM engine we have implemented.

Collaboration


Dive into the Brian F. Cooper's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Antony Williams

United States Environmental Protection Agency

View shared research outputs
Researchain Logo
Decentralizing Knowledge