Reuven Lax
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Reuven Lax.
very large data bases | 2013
Tyler Akidau; Alex Balikov; Kaya Bekiroğlu; Slava Chernyak; Josh Haberman; Reuven Lax; Sam McVeety; Daniel Mills; Paul Nordstrom; Sam Whittle
MillWheel is a framework for building low-latency data-processing applications that is widely used at Google. Users specify a directed computation graph and application code for individual nodes, and the system manages persistent state and the continuous flow of records, all within the envelope of the frameworks fault-tolerance guarantees. This paper describes MillWheels programming model as well as its implementation. The case study of a continuous anomaly detector in use at Google serves to motivate how many of MillWheels features are used. MillWheels programming model provides a notion of logical time, making it simple to write time-based aggregations. MillWheel was designed from the outset with fault tolerance and scalability in mind. In practice, we find that MillWheels unique combination of scalability, fault tolerance, and a versatile programming model lends itself to a wide variety of problems at Google.
very large data bases | 2015
Tyler Akidau; Robert Bradshaw; Craig D. Chambers; Slava Chernyak; Rafael J. Fernández-Moctezuma; Reuven Lax; Sam McVeety; Daniel Mills; Frances Perry; Eric Schmidt; Sam Whittle
Unbounded, unordered, global-scale datasets are increasingly common in day-to-day business (e.g. Web logs, mobile usage statistics, and sensor networks). At the same time, consumers of these datasets have evolved sophisticated requirements, such as event-time ordering and windowing by features of the data themselves, in addition to an insatiable hunger for faster answers. Meanwhile, practicality dictates that one can never fully optimize along all dimensions of correctness, latency, and cost for these types of input. As a result, data processing practitioners are left with the quandary of how to reconcile the tensions between these seemingly competing propositions, often resulting in disparate implementations and systems. We propose that a fundamental shift of approach is necessary to deal with these evolved requirements in modern data processing. We as a field must stop trying to groom unbounded datasets into finite pools of information that eventually become complete, and instead live and breathe under the assumption that we will never know if or when we have seen all of our data, only that new data will arrive, old data may be retracted, and the only way to make this problem tractable is via principled abstractions that allow the practitioner the choice of appropriate tradeoffs along the axes of interest: correctness, latency, and cost. In this paper, we present one such approach, the Dataflow Model, along with a detailed examination of the semantics it enables, an overview of the core principles that guided its design, and a validation of the model itself via the real-world experiences that led to its development.
Archive | 2007
Reuven Lax; Poorva Arankalle; Shamim Samadi; Rajas Moonka
Archive | 2007
Poorva Arankalle; Brienne M. Finger; Lin Liao; Manish Gupta; Rajas Moonka; Reuven Lax; Jill A. Huchital
Archive | 2008
Reuven Lax; Chao Cai
Archive | 2017
Tyler Akidau; Robert Bradshaw; Ben Chambers; Craig D. Chambers; Reuven Lax; Daniel Mills; Frances Perry
Archive | 2017
Robert Bradshaw; Rafael de Jesús Fernández Moctezuma; Daniel Mills; Samuel Green Mcveety; Samuel Carl Whittle; Andrei Maksimenka; Cosmin Ionel Arad; Mark Brian Shields; Harris Samuel Nover; Manuel Fähndrich; Jeffrey Paul Gardner; Mikhail Shmulyan; Reuven Lax; Ahmet Altay; Craig D. Chambers
Archive | 2017
Poorva Arankalle; Brienne M. Finger; Lin Liao; Manish Gupta; Rajas Moonka; Reuven Lax; Jill A. Huchital
Archive | 2008
Reuven Lax; Poorva Arankalle; Shamim Samadi; Rajas Moonka
Archive | 2008
Poorva Arankalle; Reuven Lax; Rajas Moonka; Shamim Samadi