Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Erich M. Nahum is active.

Publication


Featured researches published by Erich M. Nahum.


international world wide web conferences | 2004

A method for transparent admission control and request scheduling in e-commerce web sites

Sameh Elnikety; Erich M. Nahum; John M. Tracey; Willy Zwaenepoel

This paper presents a method for admission control and request scheduling for multiply-tiered e-commerce Web sites, achieving both stable behavior during overload and improved response times. Our method externally observes execution costs of requests online, distinguishing different request types, and performs overload protection and preferential scheduling using relatively simple measurements and a straight forward control mechanism. Unlike previous proposals, which require extensive changes to the server or operating system, our method requires no modifications to the host O.S., Web server, application server or database. Since our method is external, it can be implemented in a proxy. We present such an implementation, called Gatekeeper, using it with standard software components on the Linux operating system. We evaluate the proxy using the industry standard TPC-W workload generator in a typical three-tiered e-commerce environment. We show consistent performance during overload and throughput increases of up to 10 percent. Response time improves by up to a factor of 14, with only a 15 percent penalty to large jobs.


international workshop on quality of service | 2004

Yaksha: a self-tuning controller for managing the performance of 3-tiered Web sites

Abhinav Kamra; Vishal Misra; Erich M. Nahum

Managing the performance of multiple-tiered Web sites under high client loads is a critical problem with the advent of dynamic content and database-driven servers on the Internet. This paper presents a control-theoretic approach for admission control in multitiered Web sites that both prevents overload and enforces absolute client response times, while still maintaining high throughput under load. We use classical control theoretic techniques to design a proportional integral (PI) controller for admission control of client HTTP requests. In addition, we present a processor-sharing model that is used to make the controller self-tuning, so that no parameter setting is required beyond a target response time. Our controller is implemented as a proxy, called Yaksha, which operates by taking simple external measurements of the client response times. Our design is noninvasive and requires minimal operator intervention. We evaluate our techniques experimentally using a 3-tiered dynamic content Web site as a testbed. Using the industry standard TPC-W client workload generator, we study the performance of the PI admission controller with extensive experiments. We show that the controller effectively bounds the response times of requests for dynamic content while still maintaining high throughput levels, even when the client request rate is many times that of the servers maximum processing rate. We demonstrate the effectiveness of our self-tuning mechanism, showing that it responds and adapts smoothly to changes in the workload.


internet measurement conference | 2013

A measurement-based study of MultiPath TCP performance over wireless networks

Yung-Chih Chen; Yeon-sup Lim; Richard J. Gibbens; Erich M. Nahum; Ramin Khalili; Donald F. Towsley

With the popularity of mobile devices and the pervasive use of cellular technology, there is widespread interest in hybrid networks and on how to achieve robustness and good performance from them. As most smart phones and mobile devices are equipped with dual interfaces (WiFi and 3G/4G), a promising approach is through the use of multi-path TCP, which leverages path diversity to improve performance and provide robust data transfers. In this paper we explore the performance of multi-path TCP in the wild, focusing on simple 2-path multi-path TCP scenarios. We seek to answer the following questions: How much can a user benefit from using multi-path TCP over cellular and WiFi relative to using the either interface alone? What is the impact of flow size on average latency? What is the effect of the rate/route control algorithm on performance? We are especially interested in understanding how application level performance is affected when path characteristics (e.g., round trip times and loss rates) are diverse. We address these questions by conducting measurements using one commercial Internet service provider and three major cellular carriers in the US.


international conference on data engineering | 2006

How to Determine a Good Multi-Programming Level for External Scheduling

Bianca Schroeder; Mor Harchol-Balter; Arun Iyengar; Erich M. Nahum; Adam Wierman

Scheduling/prioritization of DBMS transactions is important for many applications that rely on database backends. A convenient way to achieve scheduling is to limit the number of transactions within the database, maintaining most of the transactions in an external queue, which can be ordered as desired by the application. While external scheduling has many advantages in that it doesn’t require changes to internal resources, it is also difficult to get right in that its performance depends critically on the particular multiprogramming limit used (the MPL), i.e. the number of transactions allowed into the database. If the MPL is too low, throughput will suffer, since not all DBMS resources will be utilized. On the other hand, if the MPL is too high, there is insufficient control on scheduling. The question of how to adjust theMPL to achieve both goals simultaneously is an open problem, not just for databases but in system design in general. Herein we study this problem in the context of transactional workloads, both via extensive experimentation and queueing theoretic analysis. We find that the two most critical factors in adjusting the MPL are the number of resources that the workload utilizes and the variability of the transactions’ service demands. We develop a feedback based controller, augmented by queueing theoretic models for automatically adjusting the MPL. Finally, we apply our methods to the specific problem of external prioritization of transactions. We find that external prioritization can be nearly as effective as internal prioritization, without any negative consequences, when the MPL is set appropriately.


measurement and modeling of computer systems | 2001

The effects of wide-area conditions on WWW server performance

Erich M. Nahum; Marcel-Catalin Rosu; Srinivasan Seshan; Jussara M. Almeida

WWW workload generators are used to evaluate web server performance, and thus have a large impact on what performance optimizations are applied to servers. However, current benchmarks ignore a crucial component: how these servers perform in the environment in which they are intended to be used, namely the wide-area Internet.This paper shows how WAN conditions can affect WWW server performance. We examine these effects using an experimental test-bed which emulates WAN characteristics in a live setting, by introducing factors such as delay and packet loss in a controlled and reproducible fashion. We study how these factors interact with the host TCP implementation and what influence they have on web server performance. We demonstrate that when more realistic wide-area conditions are introduced, servers exhibit very different performance properties and scaling behaviors, which are not exposed by existing benchmarks running on LANs. We show that observed throughputs can give misleading information about server performance, and thus find that maximum throughput, or capacity, is a more useful metric. We find that packet losses can reduce server capacity by as much as 50 percent and increase response time as seen by the client. We show that using TCP SACK can reduce client response time, without reducing server capacity.


IEEE Network | 2006

A study of Internet instant messaging and chat protocols

Raymond B. Jennings; Erich M. Nahum; David P. Olshefski; Debanjan Saha; Zon-Yin Shae; Chris Waters

Instant messaging (IM) and network chat communication have seen an enormous rise in popularity over the last several years. However, since many of these systems are proprietary, little has been described about the network technology behind them. This analysis helps bridge this gap by providing an overview of the available features, functions, system architectures, and protocol specifications of the three most popular network IM protocols: AOL Instant Messenger, Yahoo! Messenger, and Microsoft Messenger. We describe common features across these systems and highlight distinctions between them. Where possible, we discuss the advantages and disadvantages of different technical approaches used in these systems to support different features and functions. We also briefly discuss ongoing efforts to standardize IM and chat-based protocols in IETF and other standards bodies


measurement and modeling of computer systems | 2007

Evaluating SIP server performance

Erich M. Nahum; John M. Tracey; Charles P. Wright

SIP is a protocol of growing importance, with uses for VoIP, instant messaging, presence, and more. However, its performance is not well-studied or understood. In this extended abstract we overview our experimental evaluation of common SIP server scenarios using open-source SIP software such as OpenSER and SIP prunning on Linux. We show performance varies greatly depending on the server scenario and how the protocol is used. Depending on the configuration, through put can vary from hundreds to thousands of operations per second. For example, we observe that the choice of stateless vs. stateful proxying, using TCP rather than UDP, or including MD5-based authentication can each can affect performance by a factor of 2-4. We also provide kernel and application profiles using Oprofile that help explain and illustrate processing costs. Finally, we provide a simple fix for transaction-stateful proxying that improves performance by a factor of 10. Full details can be found in our accompanying technical report.


internet measurement conference | 2007

Cryptographic strength of ssl/tls servers: current and recent practices

Homin K. Lee; Tal Malkin; Erich M. Nahum

The Secure Socket Layer (SSL) and its variant, Transport Layer Security (TLS), are used toward ensuring server security. In this paper, we characterize the cryptographic strength of public servers running SSL/TLS. We present a tool developed for this purpose, the Probing SSL Security Tool (PSST), and evaluate over 19,000 servers. We expose the great diversity in the levels of cryptographic strength that is supported on the Internet. Some of our discouraging results show that most sites still support the insecure SSL 2.0, weak export-level grades of encryption ciphers, or weak RSA key strengths. We also observe encouraging behavior such as sensible default choices by servers when presented with multiple options, the quick adoption of AES (more than half the servers support strong key AES as their default choice), and the use of strong RSA key sizes of 1024 bits and above. Comparing results of running our tool over the last two years points to a positive trend that is moving in the right direction, though perhaps not as quickly as it should.


acm special interest group on data communication | 2014

How green is multipath TCP for mobile devices

Yeon-sup Lim; Yung-Chih Chen; Erich M. Nahum; Donald F. Towsley; Richard J. Gibbens

Multipath TCP is a new transport protocol that enables systems to exploit available paths through multiple network interfaces. MPTCP is particularly useful for mobile devices, which frequently have multiple wireless interfaces. However, these devices have limited power capacity and thus judicious use of these interfaces is required. In this work, we develop a model for MPTCP energy consumption derived from experimental measurements using MPTCP on a mobile device with both cellular and WiFi interfaces. Using our MPTCP energy model, we identify the operating region where MPTCP can be more power efficient than either standard TCP or MPTCP. Based on our findings, we also design and implement an improved energy-efficient MPTCP that reduces power consumption by up to 8% in our experiments, while preserving the availability and robustness benefits of MPTCP.


measurement and modeling of computer systems | 1999

Performance issues in WWW servers

Erich M. Nahum; Tsipora P. Barzilai; Dilip D. Kandlur

This paper evaluates techniques for improving operating system and network protocol software support for high-performance World Wide Web servers. We study approaches in three categories: i.e., new socket functions, per-byte optimizations, and per-connection optimizations. We examine two proposed socket functions, i.e., acceptex( ) and send_file( ), comparing send_file( )s effectiveness with a combination of mmap( ) and writev( ). We show how send_file( ) provides the necessary semantic support to eliminate copies and checksums in the kernel, and quantify the benefit of the functions header and close options. We also present mechanisms to reduce the number of packets exchanged in an HTTP transaction, both increasing server performance and reducing network utilization, without compromising interoperability. Results using WebStone show that our combination of mechanisms can improve server throughput by up to 64%, and can eliminate up to 33% of the packets in an HTTP exchange. Results with SURGE show an aggregate increase in server throughput of 25%.

Collaboration


Dive into the Erich M. Nahum's collaboration.

Top Co-Authors

Avatar

Arun Iyengar

Huazhong University of Science and Technology

View shared research outputs
Top Co-Authors

Avatar

Donald F. Towsley

University of Massachusetts Amherst

View shared research outputs
Researchain Logo
Decentralizing Knowledge