F. van Lingen
California Institute of Technology
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by F. van Lingen.
IEEE Transactions on Nuclear Science | 2005
Julia Andreeva; Ashiq Anjum; Ta Barrass; D. Bonacorsi; J. Bunn; Paolo Capiluppi; Marco Corvo; N. Darmenov; N. De Filippis; F. Donno; G. Donvito; Giulio Eulisse; A. Fanfani; F. Fanzago; A. Filine; C. Grandi; Jose M Hernandez; V. Innocente; A. Jan; S. Lacaprara; I. Legrand; S. Metson; H. B. Newman; D. M. Newbold; A. Pierro; Lucia Silvestris; Conrad Steenberg; Heinz Stockinger; L. Taylor; M. Thomas
The CMS experiment is currently developing a computing system capable of serving, processing and archiving the large number of events that will be generated when the CMS detector starts taking data. During 2004 CMS undertook a large scale data challenge to demonstrate the ability of the CMS computing system to cope with a sustained data-taking rate equivalent to 25% of startup rate. Its goals were: to run CMS event reconstruction at CERN for a sustained period at 25 Hz input rate; to distribute the data to several regional centers; and enable data access at those centers for analysis. Grid middleware was utilized to help complete all aspects of the challenge. To continue to provide scalable access from anywhere in the world to the data, CMS is developing a layer of software that uses Grid tools to gain access to data and resources, and that aims to provide physicists with a user friendly interface for submitting their analysis jobs. This paper describes the data challenge experience with Grid infrastructure and the current development of the CMS analysis system.
international conference on parallel processing | 2005
F. van Lingen; Conrad Steenberg; M. Thomas; Ashiq Anjum; Tahir Azim; F. Khan; Harvey B Newman; Arshad Ali; J. Bunn; I. Legrand
Large scientific collaborations are moving towards service oriented architectures for implementation and deployment of globally distributed systems. Clarens is a high performance, easy to deploy Web service framework that supports the construction of such globally distributed systems. This paper discusses some of the core functionality of Clarens that the authors believe is important for building distributed systems based on Web services that support scientific analysis.
Archive | 2005
F. van Lingen; M Thomas; Tahir Azim; Ashiq Anjum; J Bunn; C Steenberg; I Legrand; H Newman
This paper describes progress made in the development of the Clarens Web Services Framework, including a second Java-based server implementation, improved performance, a global lookup and discovery service leveraging of the MonALISA monitoring system, and adapting the framework to a secure message-based transport protocol.
broadband communications, networks and systems | 2006
A. Bobyshev; M. Crawford; P. DeMar; V. Grigaliunas; M. Grigoriev; Alexander Moibenko; D. Petravick; Ron Rechenmacher; Harvey B Newman; J. Bunn; F. van Lingen; Dan Nae; Sylvain Ravot; Conrad Steenberg; Xun Su; M. Thomas; Yang Xia
Lambda Station is an ongoing project of Fermi National Accelerator Laboratory and the California Institute of Technology. The goal of this project is to design, develop and deploy network services for path selection, admission control and flow based forwarding of traffic among data- intensive Grid applications such as are used in High Energy Physics and other communities. Lambda Station deals with the last-mile problem in local area networks, connecting production clusters through a rich array of wide area networks. Selective forwarding of traffic is controlled dynamically at the demand of applications. This paper introduces the motivation of this project, design principles and current status. Integration of Lambda Station client API with the essential Grid middleware such as the dCache/SRM Storage Resource Manager is also described. Finally, the results of applying Lambda Station services to development and production clusters at Fermilab and Caltech over, advanced networks such as DOEs UltraScience Net and NSFs UltraLight is covered.
IEEE Transactions on Nuclear Science | 2011
K Hasham; Antonio Delgado Peris; Ashiq Anjum; D Evans; S Gowdy; José M. Hernández; Eduardo Huedo; D. Hufnagel; F. van Lingen; Richard McClatchey; S. Metson
Complex scientific workflows can process large amounts of data using thousands of tasks. The turnaround times of these workflows are often affected by various latencies such as the resource discovery, scheduling and data access latencies for the individual workflow processes or actors. Minimizing these latencies will improve the overall execution time of a workflow and thus lead to a more efficient and robust processing environment. In this paper, we propose a pilot job concept that has intelligent data reuse and job execution strategies to minimize the scheduling, queuing, execution and data access latencies. The results have shown that significant improvements in the overall turnaround time of a workflow can be achieved with this approach. The proposed approach has been evaluated, first using the CMS Tier0 data processing workflow, and then simulating the workflows to evaluate its effectiveness in a controlled environment.
international conference on parallel processing | 2005
Arshad Ali; Ashiq Anjum; Tahir Azim; J. Bunn; A. Mehmood; Richard McClatchey; Harvey B Newman; W. ur Rehman; Conrad Steenberg; M. Thomas; F. van Lingen; Ian Willers; M.A. Zafar
Selecting optimal resources for submitting jobs on a computational grid or accessing data from a data grid is one of the most important tasks of any grid middleware. Most modern grid software today satisfies this responsibility and gives a best-effort performance to solve this problem. Almost all decisions regarding scheduling and data access are made by the software automatically, giving users little or no control over the entire process. To solve this problem, a more interactive set of services and middleware is desired that provides users more information about grid weather, and gives them more control over the decision making process. This paper presents a set of services that have been developed to provide more interactive resource management capabilities within the grid analysis environment (GAE) being developed collaboratively by Caltech, NUST and several other institutes. These include a steering service, a job monitoring service and an estimator service that have been designed and written using a common grid-enabled Web services framework named Clarens. The paper also presents a performance analysis of the developed services to show that they have indeed resulted in a more interactive and powerful system for user-centric grid-enabled physics analysis.
Archive | 2005
F. van Lingen; M Thomas; Tahir Azim; I Chitnis; Ashiq Anjum; D Bourilkov; M Kulkarni; C Steenberg; Richard Cavanaugh; J Bunn; J Ukin; P Avery; I Legrand; H Newman
The Grid Analysis Environment (GAE), which is a continuation of the CAIGEE project [5], is an effort to develop, integrate and deploy a system for distributed analysis. The current focus within the GAE is on the CMS experiment [1] however the GAE design abstracts from any specific scientific experiment and focuses on scientific analysis in general. The GAE project does not intend to reinvent services, but rather to integrate existing services into a collaborative system of web services.
Journal of Physics: Conference Series | 2008
S. Metson; S. Belforte; Brian Bockelman; K Dziedziniewicz; R. Egeland; P. Elmer; Giulio Eulisse; D Evans; A. Fanfani; Derek Feichtinger; C. Kavka; V. E. Kuznetsov; F. van Lingen; Dave M Newbold; L. Tuura; S. Wakefield
We describe a relatively new effort within CMS to converge on a set of web based tools, using state of the art industry techniques, to engage with the CMS offline computing system. CMS collaborators require tools to monitor various components of the computing system and interact with the system itself. The current state of the various CMS web tools is described along side current planned developments. The CMS collaboration comprises of nearly 3000 people from all over the world. As well as its collaborators, its computing resources are spread all over globe and are accessed via the LHC grid to run analysis, large scale production and data transfer tasks. Due to the distributed nature of collaborators effective provision of collaborative tools is essential to maximise physics exploitation of the CMS experiment, especially when the size of the CMS data set is considered. CMS has chosen to provide such tools over the world wide web as a top level service, enabling all members of the collaboration to interact with the various offline computing components. Traditionally web interfaces have been added in HEP experiments as an afterthought. In the CMS offline we have decided to put web interfaces, and the development of a common CMS web framework, on an equal footing with the rest of the offline development. Tools exist within CMS to transfer and catalogue data (PhEDEx and DBS/DLS), run Monte Carlo production (ProdAgent) and submit analysis (CRAB). Effective human interfaces to these systems are required for users with different agendas and practical knowledge of the systems to effectively use the CMS computing system. The CMS web tools project aims to provide a consistent interface to all these tools.
IEEE Symposium Conference Record Nuclear Science 2004. | 2004
Julia Andreeva; Ashiq Anjum; T. Barrass; D. Bonacorsi; J. Bunn; M. Corvo; N. Darmenov; N. De Filippis; F. Donno; G. Donvito; Giulio Eulisse; A. Fanfani; F. Fanzago; A. Filine; C. Grandi; José M. Hernández; V. Innocente; A. Jan; S. Lacaprara; I. Legrand; S. Metson; Harvey B Newman; A. Pierro; L. Silvestris; C. Steenberg; H. Stockinger; L. Taylor; M. Thomas; L. Tuura; T. Wildish
In order to prepare the Physics Technical Design Report, due by end of 2005, the CMS experiment needs to simulate, reconstruct and analyse about 100 million events, corresponding to more than 200 TB of data. The data will be distributed to several Computing Centres. In order to provide access to the whole data sample to all the world-wide dispersed physicists, CMS is developing a layer of software that uses the Grid tools provided by the LCG project to gain access to data and resources and that aims to provide a user friendly interface to the physicists submitting the analysis jobs. To achieve these aims CMS will use Grid tools from both the LCG-2 release and those being developed in the framework of the ARDA project. This work describes the current status and the future developments of the CMS analysis system.
international conference on parallel processing | 2005
Arshad Ali; Ashiq Anjum; Tahir Azim; J. Bunn; S. Iqbal; Richard McClatchey; Harvey B Newman; S.Y. Shah; Tony Solomonides; Conrad Steenberg; M. Thomas; F. van Lingen; Ian Willers
Grid based systems require a database access mechanism that can provide seamless homogeneous access to the requested data through a virtual data access system, i.e. a system which can take care of tracking the data that is stored in geographically distributed heterogeneous databases. This system should provide an integrated view of the data that is stored in the different repositories by using a virtual data access mechanism, i.e. a mechanism which can hide the heterogeneity of the backend databases from the client applications. This paper focuses on accessing data stored in disparate relational databases through a Web service interface, and exploits the features of a data warehouse and data marts. We present a middleware that enables applications to access data stored in geographically distributed relational databases without being aware of their physical locations and underlying schema. A Web service interface is provided to enable applications to access this middleware in a language and platform independent way. A prototype implementation was created based on Clarens (C. Steenberg et. al., 2003), Unity (R. Lawrence and K. Barker, 2000) and POOL (http://pool.cern.ch). This ability to access the data stored in the distributed relational databases transparently is likely to be a very powerful one for grid users, especially the scientific community wishing to collate and analyze data distributed over the grid.