Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Petr Zejdl is active.

Publication


Featured researches published by Petr Zejdl.


Journal of Instrumentation | 2013

10 Gbps TCP/IP streams from the FPGA for the CMS DAQ eventbuilder network

G. Bauer; Tomasz Bawej; Ulf Behrens; J. G. Branson; Olivier Chaze; Sergio Cittolin; Jose Antonio Coarasa; G-L Darlea; Christian Deldicque; M. Dobson; Aymeric Dupont; S. Erhan; Dominique Gigi; F. Glege; G. Gomez-Ceballos; Robert Gomez-Reino; C. Hartl; Jeroen Hegeman; A. Holzner; L. Masetti; F. Meijers; E. Meschi; R. Mommsen; S. Morovic; Carlos Nunez-Barranco-Fernandez; V. O'Dell; Luciano Orsini; Wojciech Ozga; C. Paus; Andrea Petrucci

For the upgrade of the DAQ of the CMS experiment in 2013/2014 an interface between the custom detector Front End Drivers (FEDs) and the new DAQ eventbuilder network has to be designed. For a loss-less data collection from more then 600 FEDs a new FPGA based card implementing the TCP/IP protocol suite over 10Gbps Ethernet has been developed. We present the hardware challenges and protocol modifications made to TCP in order to simplify its FPGA implementation together with a set of performance measurements which were carried out with the current prototype.


Journal of Physics: Conference Series | 2014

The new CMS DAQ system for LHC operation after 2014 (DAQ2)

Gerry Bauer; Tomasz Bawej; Ulf Behrens; James G Branson; Olivier Chaze; Sergio Cittolin; Jose Antonio Coarasa; Georgiana-Lavinia Darlea; Christian Deldicque; M. Dobson; Aymeric Dupont; S. Erhan; Dominique Gigi; F. Glege; G. Gomez-Ceballos; Robert Gomez-Reino; C. Hartl; Jeroen Hegeman; A. Holzner; L. Masetti; F. Meijers; E. Meschi; Remigius K. Mommsen; S. Morovic; Carlos Nunez-Barranco-Fernandez; V. O'Dell; Luciano Orsini; Wojciech Ozga; Christoph Paus; Andrea Petrucci

The Data Acquisition system of the Compact Muon Solenoid experiment at CERN assembles events at a rate of 100 kHz, transporting event data at an aggregate throughput of 100 GByte/s. We are presenting the design of the 2nd generation DAQ system, including studies of the event builder based on advanced networking technologies such as 10 and 40 Gbit/s Ethernet and 56 Gbit/s FDR Infiniband and exploitation of multicore CPU architectures. By the time the LHC restarts after the 2013/14 shutdown, the current compute nodes, networking, and storage infrastructure will have reached the end of their lifetime. In order to handle higher LHC luminosities and event pileup, a number of sub-detectors will be upgraded, increase the number of readout channels and replace the off-detector readout electronics with a μTCA implementation. The second generation DAQ system, foreseen for 2014, will need to accommodate the readout of both existing and new off-detector electronics and provide an increased throughput capacity. Advances in storage technology could make it feasible to write the output of the event builder to (RAM or SSD) disks and implement the HLT processing entirely file based.


Journal of Physics: Conference Series | 2014

10 Gbps TCP/IP streams from the FPGA for high energy physics

Gerry Bauer; Tomasz Bawej; Ulf Behrens; James G Branson; Olivier Chaze; Sergio Cittolin; Jose Antonio Coarasa; Georgiana-Lavinia Darlea; Christian Deldicque; M. Dobson; Aymeric Dupont; S. Erhan; Dominique Gigi; F. Glege; G. Gomez-Ceballos; Robert Gomez-Reino; C. Hartl; Jeroen Hegeman; A. Holzner; L. Masetti; F. Meijers; E. Meschi; Remigius K. Mommsen; S. Morovic; Carlos Nunez-Barranco-Fernandez; V. O'Dell; Luciano Orsini; Wojciech Ozga; Christoph Paus; Andrea Petrucci

The DAQ system of the CMS experiment at CERN collects data from more than 600 custom detector Front-End Drivers (FEDs). During 2013 and 2014 the CMS DAQ system will undergo a major upgrade to address the obsolescence of current hardware and the requirements posed by the upgrade of the LHC accelerator and various detector components. For a loss-less data collection from the FEDs a new FPGA based card implementing the TCP/IP protocol suite over 10Gbps Ethernet has been developed. To limit the TCP hardware implementation complexity the DAQ group developed a simplified and unidirectional but RFC 793 compliant version of the TCP protocol. This allows to use a PC with the standard Linux TCP/IP stack as a receiver. We present the challenges and protocol modifications made to TCP in order to simplify its FPGA implementation. We also describe the interaction between the simplified TCP and Linux TCP/IP stack including the performance measurements.


nuclear science symposium and medical imaging conference | 2015

The CMS Timing and Control Distribution System

Jeroen Hegeman; Jean-Marc Andre; Ulf Behrens; James G Branson; Olivier Chaze; Sergio Cittolin; Georgiana-Lavinia Darlea; Christian Deldicque; Z. Demiragli; M. Dobson; S. Erhan; J. Fulcher; Dominique Gigi; F. Glege; G. Gomez-Ceballos; Magnus Hansen; A. Holzner; Raul Jimenez-Estupiñán; L. Masetti; F. Meijers; E. Meschi; Remigius K. Mommsen; S. Morovic; V. O'Dell; Luciano Orsini; Christoph Paus; M. Pieri; Attila Racz; H. Sakulin; C. Schwick

The Compact Muon Solenoid (CMS) experiment operating at the CERN (European Laboratory for Nuclear Physics) Large Hadron Collider (LHC) is in the process of upgrading several of its detector systems. Adding more individual detector components brings the need to test and commission those components separately from existing ones so as not to compromise physics data-taking. The CMS Trigger, Timing and Control (TTC) system had reached its limits in terms of the number of separate elements (partitions) that could be supported. A new Timing and Control Distribution System (TCDS) has been designed, built and commissioned in order to overcome this limit. It also brings additional functionality to facilitate parallel commissioning of new detector elements. The new TCDS system and its components will be described and results from the first operational experience with the TCDS in CMS will be shown.


ieee-npss real-time conference | 2014

Achieving High Performance With TCP Over 40 GbE on NUMA Architectures for CMS Data Acquisition

Tomasz Bawej; Ulf Behrens; James G Branson; Olivier Chaze; Sergio Cittolin; Georgiana-Lavinia Darlea; Christian Deldicque; M. Dobson; Aymeric Dupont; S. Erhan; Andrew Forrest; Dominique Gigi; F. Glege; G. Gomez-Ceballos; Robert Gomez-Reino; Jeroen Hegeman; A. Holzner; L. Masetti; F. Meijers; E. Meschi; Remigius K. Mommsen; S. Morovic; Carlos Nunez-Barranco-Fernandez; V. O'Dell; Luciano Orsini; Christoph Paus; Andrea Petrucci; M. Pieri; Attila Racz; H. Sakulin

TCP and the socket abstraction have barely changed over the last two decades, but at the network layer there has been a giant leap from a few megabits to 100 gigabits in bandwidth. At the same time, CPU architectures have evolved into the multi-core era and applications are expected to make full use of all available resources. Applications in the data acquisition domain based on the standard socket library running in a Non-Uniform Memory Access (NUMA) architecture are unable to reach full efficiency and scalability without the software being adequately aware about the IRQ (Interrupt Request), CPU and memory affinities. During the first long shutdown of LHC, the CMS DAQ system is going to be upgraded for operation from 2015 onwards and a new software component has been designed and developed in the CMS online framework for transferring data with sockets. This software attempts to wrap the low-level socket library to ease higher-level programming with an API based on an asynchronous event driven model similar to the DAT uDAPL API. It is an event-based application with NUMA optimizations, that allows for a high throughput of data across a large distributed system. This paper describes the architecture, the technologies involved and the performance measurements of the software in the context of the CMS distributed event building.


Proceedings of Topical Workshop on Electronics for Particle Physics — PoS(TWEPP-17) | 2018

The FEROL40, a microTCA card interfacing custom point-to-point links and standard TCP/IP

Dominique Gigi; Petia Petrova; Attila Racz; Samuel Johan Orn; T. Reis; Christian Deldicque; Michail Vougioukas; Michael Lettrich; Cristian Contescu; E. Meschi; Ioannis Papakrivopoulos; M. Dobson; V. O'Dell; F. Glege; Maciej Gladki; Dainius Simelevicius; James G Branson; A. Holzner; H. Sakulin; Sergio Cittolin; Andrea Petrucci; F. Meijers; Remigius K. Mommsen; Mindaugas Janulis; M. Pieri; Ulf Behrens; Nicolas Doualot; C. Schwick; J. Fulcher; Jeroen Hegeman

In order to accommodate new back-end electronics of upgraded CMS sub-detectors, a new FEROL40 card in the microTCA standard has been developed. The main function of the FEROL40 is to acquire event data over multiple point-to-point serial optical links, provide buffering, perform protocol conversion, and transmit multiple TCP/IP streams (4x10Gbps) to the Ethernet network of the aggregation layer of the CMS DAQ (data acquisition) event builder. This contribution discusses the design of the FEROL40 and experience from operation


Proceedings of Topical Workshop on Electronics for Particle Physics — PoS(TWEPP-17) | 2018

CMS DAQ current and future hardware upgrades up to post Long Shutdown 3 (LS3) times

Attila Racz; Petia Petrova; Dominique Gigi; Samuel Johan Orn; T. Reis; Christian Deldicque; Michail Vougioukas; Michael Lettrich; Cristian Contescu; E. Meschi; Ioannis Papakrivopoulos; M. Dobson; V. O'Dell; F. Glege; Maciej Gladki; Dainius Simelevicius; James G Branson; A. Holzner; H. Sakulin; Sergio Cittolin; Andrea Petrucci; F. Meijers; Remigius K. Mommsen; Mindaugas Janulis; M. Pieri; Ulf Behrens; Nicolas Doualot; C. Schwick; J. Fulcher; Jeroen Hegeman

Following the first LHC collisions seen and recorded by CMS in 2009, the DAQ hardware went through a major upgrade during LS1 (2013-2014) and new detectors have been connected during 2015-2016 and 2016-2017 winter shutdowns. Now, LS2 (2019-2020) and LS3 (2024-mid 2026) are actively being prepared. This paper shows how CMS DAQ hardware has evolved from the beginning and will continue to evolve in order to meet the future challenges posed by High Luminosity LHC (HL-LHC) and the CMS detector evolution. In particular, post LS3 DAQ architectures are focused upon.


Journal of Physics: Conference Series | 2018

DAQExpert - An expert system to increase CMS data-taking efficiency

Jean-Marc Andre; Petia Petrova; D Gigi; Attila Racz; Samuel Johan Orn; A. Holzner; T. Reis; Christian Deldicque; Michail Vougioukas; Michael Lettrich; Cristian Contescu; E. Meschi; Ioannis Papakrivopoulos; F Meijers; M. Dobson; V. O'Dell; F. Glege; Dainius Simelevicius; Georgiana Lavinia Darlea; Christoph Paus; Z. Demiragli; H. Sakulin; D. Rabady; Jeroen Hegeman; Andrea Petrucci; Remigius K. Mommsen; Mindaugas Janulis; M. Pieri; Ulf Behrens; Nicolas Doualot

The efficiency of the Data Acquisition (DAQ) of the Compact Muon Solenoid (CMS) experiment for LHC Run 2 is constantly being improved. A significant factor affecting the data taking efficiency is the experience of the DAQ operator. One of the main responsibilities of the DAQ operator is to carry out the proper recovery procedure in case of failure of datataking. At the start of Run 2, understanding the problem and finding the right remedy could take a considerable amount of time (up to many minutes). Operators heavily relied on the support of on-call experts, also outside working hours. Wrong decisions due to time pressure sometimes lead to an additional overhead in recovery time. To increase the efficiency of CMS data-taking we developed a new expert system, the DAQExpert, which provides shifters with optimal recovery suggestions instantly when a failure occurs. DAQExpert is a web application analyzing frequently updating monitoring data from all DAQ components and identifying problems based on expert knowledge expressed in small, independent logic-modules written in Java. Its results are presented in real-time in the control room via a web-based GUI and a sound-system in a form of short description of the current failure, and steps to recover.


Journal of Physics: Conference Series | 2017

The CMS Data Acquisition - Architectures for the Phase-2 Upgrade

Jean-Marc Andre; Petia Petrova; Dominique Gigi; Attila Racz; A. Holzner; T. Reis; Christian Deldicque; Cristian Contescu; E. Meschi; Philipp Maximilian Brummer; F Meijers; M. Dobson; Raul Jimenez Estupinan; F. Glege; Dainius Simelevicius; James G Branson; Christoph Paus; Z. Demiragli; H. Sakulin; Jeroen Hegeman; Jonathan F Fulcheri; V. O'Dell; Remigius K. Mommsen; Mindaugas Janulis; M. Pieri; Ulf Behrens; Nicolas Doualot; C. Schwick; adki; Luciano Orsini

The upgraded High Luminosity LHC, after the third Long Shutdown (LS3), will provide an instantaneous luminosity of 7.5 × 10 cm−2s−1 (levelled), at the price of extreme pileup of up to 200 interactions per crossing. In LS3, the CMS Detector will also undergo a major upgrade to prepare for the phase-2 of the LHC physics program, starting around 2025. The upgraded detector will be read out at an unprecedented data rate of up to 50 Tb/s and an event rate of 750 kHz. Complete events will be analysed by software algorithms running on standard processing nodes, and selected events will be stored permanently at a rate of up to 10 kHz for offline processing and analysis. In this paper we discuss the baseline design of the DAQ and HLT systems for the phase-2, taking into account the projected evolution of high speed network fabrics for event building and distribution, and the anticipated performance of general purpose CPU. Implications on hardware and infrastructure requirements for the DAQ ”data center” are analysed. Emerging technologies for data reduction are considered. Novel possible approaches to event building and online processing, inspired by trending developments in other areas of computing dealing with large masses of data, are also examined. We conclude by discussing the opportunities offered by reading out and processing parts of the detector, wherever the front-end electronics allows, at the machine clock rate (40 MHz). Presented at CHEP 2016 22nd International Conference on Computing in High Energy and Nuclear Physics 1) DESY, Hamburg, Germany 2) CERN, Geneva, Switzerland 3) University of California, Los Angeles, Los Angeles, California, USA 4) University of California, San Diego, San Diego, California, USA 5) FNAL, Chicago, Illinois, USA 6) Massachusetts Institute of Technology, Cambridge, Massachusetts, USA The CMS Data Acquisition Architectures for the Phase-2 Upgrade J-M Andre5, U Behrens1, J Branson4, P Brummer2, O Chaze2, S Cittolin4, C Contescu5, B G Craigs2, G-L Darlea6, C Deldicque2, Z Demiragli6, M Dobson2, N Doualot5, S Erhan3, J F Fulcher2, D Gigi2, M G ladki2, F Glege2, G Gomez-Ceballos6, J Hegeman2, A Holzner4, M Janulis2a, R Jimenez-Estupiñán2, L Masetti2, F Meijers2, E Meschi2, R K Mommsen5, S Morovic2, V O’Dell5, L Orsini2, C Paus6, P Petrova2, M Pieri4, A Racz2, T Reis2, H Sakulin2, C Schwick2, D Simelevicius2a, and P Zejdl5b 1 DESY, Hamburg, Germany 2 CERN, Geneva, Switzerland 3 University of California, Los Angeles, California, USA 4 University of California, San Diego, California, USA 5 FNAL, Batavia, Illinois, USA 6 Massachusetts Institute of Technology, Cambridge, Massachusetts, USA a also at Vilnius University, Vilnius, Lithuania b also at CERN, Geneva, Switzerland E-mail: [email protected] Abstract. The upgraded High Luminosity LHC, after the third Long Shutdown (LS3), will provide an instantaneous luminosity of 7.5 × 10 cm−2s−1 (levelled), at the price of extreme pileup of up to 200 interactions per crossing. In LS3, the CMS Detector will also undergo a major upgrade to prepare for the phase-2 of the LHC physics program, starting around 2025. The upgraded detector will be read out at an unprecedented data rate of up to 50 Tb/s and an event rate of 750 kHz. Complete events will be analysed by software algorithms running on standard processing nodes, and selected events will be stored permanently at a rate of up to 10 kHz for offline processing and analysis. In this paper we discuss the baseline design of the DAQ and HLT systems for the phase-2, taking into account the projected evolution of high speed network fabrics for event building and distribution, and the anticipated performance of general purpose CPU. Implications on hardware and infrastructure requirements for the DAQ ”data center” are analysed. Emerging technologies for data reduction are considered. Novel possible approaches to event building and online processing, inspired by trending developments in other areas of computing dealing with large masses of data, are also examined. We conclude by discussing the opportunities offered by reading out and processing parts of the detector, wherever the front-end electronics allows, at the machine clock rate (40 MHz). This idea presents interesting challenges and its physics potential should be studied. The upgraded High Luminosity LHC, after the third Long Shutdown (LS3), will provide an instantaneous luminosity of 7.5 × 10 cm−2s−1 (levelled), at the price of extreme pileup of up to 200 interactions per crossing. In LS3, the CMS Detector will also undergo a major upgrade to prepare for the phase-2 of the LHC physics program, starting around 2025. The upgraded detector will be read out at an unprecedented data rate of up to 50 Tb/s and an event rate of 750 kHz. Complete events will be analysed by software algorithms running on standard processing nodes, and selected events will be stored permanently at a rate of up to 10 kHz for offline processing and analysis. In this paper we discuss the baseline design of the DAQ and HLT systems for the phase-2, taking into account the projected evolution of high speed network fabrics for event building and distribution, and the anticipated performance of general purpose CPU. Implications on hardware and infrastructure requirements for the DAQ ”data center” are analysed. Emerging technologies for data reduction are considered. Novel possible approaches to event building and online processing, inspired by trending developments in other areas of computing dealing with large masses of data, are also examined. We conclude by discussing the opportunities offered by reading out and processing parts of the detector, wherever the front-end electronics allows, at the machine clock rate (40 MHz). This idea presents interesting challenges and its physics potential should be studied.


Journal of Physics: Conference Series | 2017

Performance of the CMS Event Builder

Jean-Marc Andre; Petia Petrova; Dominique Gigi; Attila Racz; A. Holzner; T. Reis; Lithuania; Christian Deldicque; Illinois; Cristian Contescu; E. Meschi; Philipp Maximilian Brummer; F Meijers; M. Dobson; Raul Jimenez Estupinan; F. Glege; Dainius Simelevicius; James G Branson; Christoph Paus; Z. Demiragli; H. Sakulin; Hamburg; Jeroen Hegeman; V. O'Dell; Remigius K. Mommsen; Chicago; Mindaugas Janulis; M. Pieri; Ulf Behrens; Nicolas Doualot

The data acquisition system (DAQ) of the CMS experiment at the CERN Large Hadron Collider (LHC) assembles events at a rate of 100 kHz. It transports event data at an aggregate throughput of 100 GB/s to the high-level trigger (HLT) farm. The CMS DAQ system has been completely rebuilt during the first long shutdown of the LHC in 2013/14. The new DAQ architecture is based on state-ofthe-art network technologies for the event building. For the data concentration, 10/40 Gb/s Ethernet technologies are used together with a reduced TCP/IP protocol implemented in FPGA for a reliable transport between custom electronics and commercial computing hardware. A 56 Gb/s Infiniband FDR CLOS network has been chosen for the event builder. We report on the performance of the event builder system and the steps taken to exploit the full potential of the network technologies. Presented at CHEP 2016 22nd International Conference on Computing in High Energy and Nuclear Physics Performance of the CMS Event Builder J-M Andre5, U Behrens1, J Branson4, P Brummer2, O Chaze2, S Cittolin4, C Contescu5, B G Craigs2, G-L Darlea6, C Deldicque2, Z Demiragli6, M Dobson2, N Doualot5, S Erhan3, J F Fulcher2, D Gigi2, M G ladki2, F Glege2, G Gomez-Ceballos6, J Hegeman2, A Holzner4, M Janulis2a, R Jimenez-Estupiñán2, L Masetti2, F Meijers2, E Meschi2, R K Mommsen5, S Morovic2, V O’Dell5, L Orsini2, C Paus6, P Petrova2, M Pieri4, A Racz2, T Reis2, H Sakulin2, C Schwick2, D Simelevicius2a and P Zejdl5b 1 DESY, Hamburg, Germany 2 CERN, Geneva, Switzerland 3 University of California, Los Angeles, California, USA 4 University of California, San Diego, California, USA 5 FNAL, Chicago, Illinois, USA 6 Massachusetts Institute of Technology, Cambridge, Massachusetts, USA a also at Vilnius University, Vilnius, Lithuania b also at CERN, Geneva, Switzerland E-mail: [email protected] Abstract. The data acquisition system (DAQ) of the CMS experiment at the CERN Large Hadron Collider assembles events at a rate of 100 kHz, transporting event data at an aggregate throughput of O(100 GB/s) to the high-level trigger farm. The DAQ architecture is based on state-of-the-art network technologies for the event building. For the data concentration, 10/40 Gbit/s Ethernet technologies are used together with a reduced TCP/IP protocol implemented in FPGA for a reliable transport between custom electronics and commercial computing hardware. A 56 Gbit/s Infiniband FDR Clos network has been chosen for the event builder. This paper presents the implementation and performance of the event-building system. The data acquisition system (DAQ) of the CMS experiment at the CERN Large Hadron Collider assembles events at a rate of 100 kHz, transporting event data at an aggregate throughput of O(100 GB/s) to the high-level trigger farm. The DAQ architecture is based on state-of-the-art network technologies for the event building. For the data concentration, 10/40 Gbit/s Ethernet technologies are used together with a reduced TCP/IP protocol implemented in FPGA for a reliable transport between custom electronics and commercial computing hardware. A 56 Gbit/s Infiniband FDR Clos network has been chosen for the event builder. This paper presents the implementation and performance of the event-building system.

Collaboration


Dive into the Petr Zejdl's collaboration.

Top Co-Authors

Avatar

A. Holzner

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

M. Pieri

University of California

View shared research outputs
Researchain Logo
Decentralizing Knowledge