Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Honesty C. Young is active.

Publication


Featured researches published by Honesty C. Young.


ACM Transactions on Computer Systems | 2005

The automatic improvement of locality in storage systems

Windsor Wee Sun Hsu; Alan Jay Smith; Honesty C. Young

Disk I/O is increasingly the performance bottleneck in computer systems despite rapidly increasing disk data transfer rates. In this article, we propose Automatic Locality-Improving Storage (ALIS), an introspective storage system that automatically reorganizes selected disk blocks based on the dynamic reference stream to increase effective storage performance. ALIS is based on the observations that sequential data fetch is far more efficient than random access, that improving seek distances produces only marginal performance improvements, and that the increasingly powerful processors and large memories in storage systems have ample capacity to reorganize the data layout and redirect the accesses so as to take advantage of rapid sequential data transfer. Using trace-driven simulation with a large set of real workloads, we demonstrate that ALIS considerably outperforms prior techniques, improving the average read performance by up to 50% for server workloads and by about 15% for personal computer workloads. We also show that the performance improvement persists as disk technology evolves. Since disk performance in practice is increasing by only about 8% per year, the benefit of ALIS may correspond to as much as several years of technological progress.


ACM Transactions on Database Systems | 2001

I/O reference behavior of production database workloads and the TPC benchmarks—an analysis at the logical level

Windsor Wee Sun Hsu; Alan Jay Smith; Honesty C. Young

As improvements in processor performance continue to far outpace improvements in storage performance, I/O is increasingly the bottleneck in computer systems, especially in large database systems that manage huge amoungs of data. The key to achieving good I/O performance is to thoroughly understand its characteristics. In this article we present a comprehensive analysis of the logical I/O reference behavior of the peak productiondatabase workloads from ten of the worlds largest corporations. In particular, we focus on how these workloads respond to different techniques for caching, prefetching, and write buffering. Our findings include several broadly applicable rules of thumb that describe how effective the various I/O optimization techniques are for the production workloads. For instance, our results indicate that the buffer pool miss ratio tends to be related to the ratio of buffer pool size to data size by an inverse square root rule. A similar fourth root rule relates the write miss ratio and the ration of buffer pool size to data size. In addition, we characterize the reference characteristics of workloads similar to the Transaction Processing Performance Council (TPC) benchmarks C (TPC-C) and D(TPC-D), which are de facto standard performance measures for online transaction processing (OLTP) systems and decision support systems (DSS), respectively. Since benchmarks such as TPC-C and TPC-D can only be used effectively if their strengths and limitations are understood, a major focus of our analysis is to identify aspects of the benchmarks that stress the system differently than the production workloads. We discover that for the most part, the reference behavior of TPC-C and TPC-D fall within the range of behavior exhibited by the production workloads. However, there are some noteworthy exceptions that affect well-known I/O optimization techniques such as caching (LRU is further from the optimal for TPC-C, while there is little sharing of pages between transactions for TPC-D), prefetching (TPC-C exhibits no significant sequentiality), and write buffering (write buffering is lees effective for the TPC benchmarks). While the two TPC benchmarks generally complement one another in reflecting the characteristics of the production workloads, there remain aspects of the real workloads that are not represented by either of the benchmarks.


IEEE Transactions on Parallel and Distributed Systems | 1993

A symmetric fragment and replicate algorithm for distributed joins

James W. Stamos; Honesty C. Young

It is shown that the fragment and replicate (FR) distributed join algorithm is a special case of the symmetric fragment and replicate (SFR) algorithm, which improves the FR algorithm by reducing its communication. The SFR algorithm, like the FR algorithm, is applicable to N-way joins and nonequijoins and does tuple balancing automatically. The authors derive formulae that show how to minimize the communication in the SFR algorithm, discuss its performance on a parallel database prototype, and evaluate its practicality under various conditions. It is claimed that SFR improves the worst-case cost for a distributed join, but it will not displace specialized distributed join algorithms when the later are applicable. >


international conference on parallel and distributed systems | 2000

Projecting the performance of decision support workloads on systems with smart storage (SmartSTOR)

Windsor Wee Sun Hsu; Alan Jay Smith; Honesty C. Young

Recent developments in both hardware and software have made it worthwhile to consider embedding intelligence in storage to handle general-purpose processing that can be off-loaded from the hosts. In particular, low-cost processing power is now widely available and software can be made robust, secure and mobile. In this paper, we propose a general smart storage (SmartSTOR) architecture in which a processing unit that is coupled to one or more disks can be used to perform such off-loaded processing. A major part of the paper is devoted to understanding the performance potential of the SmartSTOR architecture for decision support workloads. Our analysis suggests that there is a definite performance advantage in using fewer but more powerful processors, a result that bolsters the case for sharing a powerful processor among multiple disks. As for software architecture, we find that the off-loading of database operations that involve only a single relation is not very promising. In order to achieve significant speed-up, we have to consider the off-loading of multiple-relation operations. In general, if embedding intelligence in storage is an inevitable architectural trend, we have to focus on developing parallel software systems that can effectively take advantage of the large number of processing units that will be in the system.


international conference on computer design | 1993

An intelligent I-cache prefetch mechanism

Honesty C. Young; Eugene J. Shekita

Modern, high-performance processors employ techniques such as superscalar execution and super-pipelining to increase their instruction issue rate. As the instruction issue rate of processors increases, however, the negative impact of branches on performance also increases. This paper describes an instruction cache (I-cache) prefetch mechanism to improve processor performance on the taken branches. Under perfect conditions, the mechanism allows the target of a taken branch to be prefetched early enough to completely hide the memory latency of an I-cache miss.<<ETX>>


Ibm Journal of Research and Development | 1991

Exploiting database parallelism in a message-passing multiprocessor

Raymond A. Lorie; Jean-Jacques Daudenarde; James W. Stamos; Honesty C. Young

Parallel processing may well be the only means of satisfying the long-term performance requirements for database systems: an increase in throughput for transactions and a drastic decrease in response time for complex queries, in this paper, we review various alternatives, and then focus entirely on exploiting parallel-processing configurations in which general-purpose processors communicate only via message passing, in our configuration, the database Is partitioned among the processors. This approach iooi<s promising but offers challenging problems. The paper reports on our solutions to some of them: how to express strategies for efficiently executing complex queries, how to minimize overhead In operations such as parallel Joins and sorts, and how to deal with transaction management in a highly distributed system. The paper ends with a discussion of the lessons we learned from exercising a prototype developed in IBIM Research. introduction Parallel processing may be harnessed to increase throughput (using n times as many computers to do n times as much work in the same amount of time) or to reduce response time (using n times as many computers to solve a problem in 1/nth the time). Parallel processing is applicable to database systems, but several factors make the problem challenging. First, the nature of the workload is unknown when the database system is deployed, since interactive users can construct arbitrary queries and execute them at any time. Second, the characteristics of the subset of data manipulated by a query are in general not known until run time, because it depends on both the predicates in the query and the state of the database. Third, shared-database systems permit concurrent access, which means that a parallel database system must support parallelism and multiprogramming simultaneously. These three challenges emphasize the need for run-time flexibility, load balancing, and effective scheduling.


european conference on parallel processing | 2000

Logging RAID - An Approach to Fast, Reliable, and Low-Cost Disk Arrays

Ying Chen; Windsor Wee Sun Hsu; Honesty C. Young

Parity-based disk arrays provide high reliability and high performance for read and large write accesses at low storage cost. However, small writes are notoriously slow due to the well-known read-modify-write problem. This paper presents logging RAID, a disk array architecture that adopts data logging techniques to overcome the small-write problem in parity-based disk arrays. Logging RAID achieves high performance for a wide variety of I/O access patterns with very small disk space overhead. We show this through trace-driven simulations.


database and expert systems applications | 1997

A General Multidimensional Data Allocation Method for Multicomputer Database Systems

Yu-lung Lo; Kien A. Hua; Honesty C. Young

Several studies have demonstrated that both the performance and scalability of a shared-nothing parallel database system depend on the physical layout of data across the processing nodes of the system. Today, data is allocated in these systems using horizontal partitioning strategies. This approach has a number of drawbacks. In recent years, several multidimensional data declustering techniques have been proposed to address these problems. However, these schemes are too restrictive, or optimized for a certain type of queries. In this paper, we introduce a new technique which is flexible, and performs well for general queries.


IEEE Transactions on Knowledge and Data Engineering | 1995

Applying segmented right-deep trees to pipelining multiple hash joins

Ming-Syan Chen; Ming-Ling Lo; Philip S. Yu; Honesty C. Young

The pipelined execution of multijoin queries in a multiprocessor-based database system is explored in this paper. Using hash-based joins, multiple joins can be pipelined so that the early results from a join, before the whole join is completed, are sent to the next join for processing. The execution of a query is usually denoted by a query execution tree. To improve the execution of pipelined hash joins, an innovative approach to query execution tree selection is proposed to exploit segmented right-deep trees, which are bushy trees of right-deep subtrees. We first derive an analytical model for the execution of a pipeline segment, and then, in the light of the model, we develop heuristic schemes to determine the query execution plan based on a segmented right-deep tree so that the query can be efficiently executed. As shown by our simulation, the proposed approach, without incurring additional overhead on plan execution, possesses more flexibility in query plan generation, and can lead to query plans of better performance than those achievable by the previous schemes using right-deep trees. >


very large data bases | 1993

Considering data skew factor in multi-way join query optimization for parallel execution

Kien A. Hua; Yo Lung Lo; Honesty C. Young

A consensus on parallel architecture for very large database management has emerged. This architecture is based on a shared-nothing hardware organization. The computation model is very sensitive to skew in tuple distribution, however. Recently, several parallel join algorithms with dynamic load balancing capabilities have been proposed to address this issue, but none of them consider multi-way join problems. In this article we propose a dynamic load balancing technique for multi-way joins, and investigate the effect of load balancing on query optimization. In particular, we present a join-ordering strategy that takes load-balancing issues into consideration. Our performance study indicates that the proposed query optimization technique can provide very impressive performance improvement over conventional approaches.

Collaboration


Dive into the Honesty C. Young's collaboration.

Researchain Logo
Decentralizing Knowledge