Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Gabriela Jacques-Silva is active.

Publication


Featured researches published by Gabriela Jacques-Silva.


very large data bases | 2013

Streaming algorithms for k-core decomposition

Ahmet Erdem Sariyüce; Bugra Gedik; Gabriela Jacques-Silva; Kun-Lung Wu

A k-core of a graph is a maximal connected subgraph in which every vertex is connected to at least k vertices in the subgraph. k-core decomposition is often used in large-scale network analysis, such as community detection, protein function prediction, visualization, and solving NP-Hard problems on real networks efficiently, like maximal clique finding. In many real-world applications, networks change over time. As a result, it is essential to develop efficient incremental algorithms for streaming graph data. In this paper, we propose the first incremental k-core decomposition algorithms for streaming graph data. These algorithms locate a small subgraph that is guaranteed to contain the list of vertices whose maximum k-core values have to be updated, and efficiently process this subgraph to update the k-core decomposition. Our results show a significant reduction in run-time compared to non-incremental alternatives. We show the efficiency of our algorithms on different types of real and synthetic graphs, at different scales. For a graph of 16 million vertices, we observe speedups reaching a million times, relative to the non-incremental algorithms.


international conference on autonomic computing | 2007

Towards Autonomic Fault Recovery in System-S

Gabriela Jacques-Silva; Jim Challenger; Lou Degenaro; James R. Giles; Rohit Wagle

System-S is a stream processing infrastructure which enables program fragments to be distributed and connected to form complex applications. There may be potentially tens of thousands of interdependent and heterogeneous program fragments running across thousands of nodes. While the scale and interconnection imply the need for automation to manage the program fragments, the need is intensified because the applications operate on live streaming data and thus need to be highly available. System-S has been designed with components that autonomically manage the program fragments, but the system components themselves are also susceptible to failures which can jeopardize the system and its applications. The work we present addresses the self healing nature of these management components in System-S. In particular, we show how one key component of System-S, the job management orchestrator, can be abruptly terminated and then recover without interrupting any of the running program fragments by reconciling with other autonomous system components. We also describe techniques that we have developed to validate that the system is able to autonomically respond to a wide variety of error conditions including the abrupt termination and recovery of key system components. Finally, we show the performance of the job management orchestrator recovery for a variety of workloads.


dependable systems and networks | 2009

Language level checkpointing support for stream processing applications

Gabriela Jacques-Silva; Bugra Gedik; Henrique Andrade; Kun-Lung Wu

Many streaming applications demand continuous processing of live data with little or no downtime, therefore, making high-availability a crucial operational requirement. Fault tolerance techniques are generally expensive and when directly applied to streaming systems with stringent throughput and latency requirements, they might incur a prohibitive performance overhead. This paper describes a flexible, light-weight fault tolerance solution in the context of the SPADE language and the System S distributed stream processing engine. We devised language extensions so users can define and parameterize check-point policies easily. This configurable fault tolerance solution is implemented through code generation in SPADE, which reduces the overall application fault tolerance costs by incurring them only for the parts of the application that require it. In this paper we focus on the overall design of our checkpoint mechanism and we also describe an incremental checkpointing algorithm that is suitable for on-the-fly processing of high-rate data streams.


distributed event-based systems | 2011

Fault injection-based assessment of partial fault tolerance in stream processing applications

Gabriela Jacques-Silva; Bugra Gedik; Henrique Andrade; Kun Lung Wu; Ravishankar K. Iyer

This paper describes an experimental methodology used to evaluate the effectiveness of partial fault tolerance (PFT) techniques in data stream processing applications. Without a clear understanding of the impact of faults on the quality of the application output, applying PFT techniques in practice is not viable. We assess the impact of PFT by injecting faults into a synthetic financial engineering application running on top of IBMs stream processing middleware, System S. The application output quality degradation is evaluated via an application-specific output score function. In addition, we propose four metrics that are aimed at assessing the impact of faults in different stream operators of the application flow graph with respect to predictability and availability. These metrics help the developer to decide where in the application he should place redundant resources. We show that PFT is indeed viable, which opens the way for considerably reducing the resource consumption when compared to fully consistent replicas.


Knowledge and Information Systems | 2015

Sliding windows over uncertain data streams

Michele Dallachiesa; Gabriela Jacques-Silva; Bugra Gedik; Kun-Lung Wu; Themis Palpanas

Uncertain data streams can have tuples with both value and existential uncertainty. A tuple has value uncertainty when it can assume multiple possible values. A tuple is existentially uncertain when the sum of the probabilities of its possible values is


pacific rim international symposium on dependable computing | 2008

Error Behavior Comparison of Multiple Computing Systems: A Case Study Using Linux on Pentium, Solaris on SPARC, and AIX on POWER

Daniel Chen; Gabriela Jacques-Silva; Zbigniew Kalbarczyk; Ravishankar K. Iyer; Bruce Mealey


dependable systems and networks | 2011

Modeling stream processing applications for dependability evaluation

Gabriela Jacques-Silva; Zbigniew Kalbarczyk; Bugra Gedik; Henrique Andrade; Kun Lung Wu; Ravishankar K. Iyer

<


very large data bases | 2012

Building user-defined runtime adaptation routines for stream processing applications

Gabriela Jacques-Silva; Bugra Gedik; Rohit Wagle; Kun-Lung Wu; Vibhore Kumar


very large data bases | 2016

Incremental k-core decomposition: algorithms and evaluation

Ahmet Erdem Sariyüce; Bugra Gedik; Gabriela Jacques-Silva; Kun Lung Wu

<1. A situation where existential uncertainty can arise is when applying relational operators to streams with value uncertainty. Several prior works have focused on querying and mining data streams with both value and existential uncertainty. However, none of them have studied, in depth, the implications of existential uncertainty on sliding window processing, even though it naturally arises when processing uncertain data. In this work, we study the challenges arising from existential uncertainty, more specifically the management of count-based sliding windows, which are a basic building block of stream processing applications. We extend the semantics of sliding window to define the novel concept of uncertain sliding windows and provide both exact and approximate algorithms for managing windows under existential uncertainty. We also show how current state-of-the-art techniques for answering similarity join queries can be easily adapted to be used with uncertain sliding windows. We evaluate our proposed techniques under a variety of configurations using real data. The results show that the algorithms used to maintain uncertain sliding windows can efficiently operate while providing a high-quality approximation in query answering. In addition, we show that sort-based similarity join algorithms can perform better than index-based techniques (on 17 real datasets) when the number of possible values per tuple is low, as in many real-world applications.


very large data bases | 2016

Consistent regions: guaranteed tuple processing in IBM streams

Gabriela Jacques-Silva; Fang Zheng; Daniel J. Debrunner; Kun-Lung Wu; Victor Dogaru; Eric A. Johnson; Michael John Elvery Spicer; Ahmet Erdem Sariyüce

This paper presents an approach to conducting experimental studies for the characterization and comparison of the error behavior in different computing systems. The proposed approach is applied to characterize and compare the error behavior of three commercial systems (Linux 2.6 on Pentium 4, Solaris 10 on UltraSPARC IIIi, and AIX 5.3 on POWER 5) under hardware transient faults. The data is obtained by conducting extensive fault injection into kernel code, kernel stack, and system registers with the NFTAPE framework while running the Apache Web server as a workload. The error behavior comparison shows that the Linux system has the highest average crash latency, the Solaris system has the highest hang rate, and the AIX system has the lowest error sensitivity and the least amount of crashes in the more severe categories.

Collaboration


Dive into the Gabriela Jacques-Silva's collaboration.

Researchain Logo
Decentralizing Knowledge