Featured Researches

Data Analysis Statistics And Probability

Comparative visualization of epidemiological data during various stages of a pandemic

After COVID-19 was first reported in China at the end of 2019, it took only a few months for this local crisis to turn into a global pandemic with unprecedented disruptions of everyday life. However, at any moment in time the situation in different parts of the world is far from uniform and each country follows its own epidemiological trajectory. In order to keep track of the course of the pandemic in many different places at the same time, it is vital to develop comparative visualizations that facilitate the recognition of common trends and divergent behaviors. Similarly, it is important to always focus on the information that is most relevant at any given point in time. In this study we look at exactly one year of daily numbers of new cases and deaths and present data visualizations that compare many different countries and are adapted to the overall stage of the pandemic. During the early stage when cases and deaths still rise we focus on the time lag relative to the current epicenter of the pandemic and the doubling times. Later we monitor the rise and fall of the daily numbers via wave detection plots. The transition between these two stages takes place when the daily numbers stop rising for the first time.

Read more
Data Analysis Statistics And Probability

Comparison of different ML methods applied to the classification of events with ttbar in the final state at the ATLAS experiment

This contribution describes the experience with the application of different Machine Learning (ML) techniques to a physics analysis case. The use case chosen is the classification of top-antitop events coming from BSM or from SM using data from a repository of simulated events. The features of these events are represented by their kinematic observables. The initial objective was to compare different ML methods in order to see whether they can lead to an improvement in the classification, but the work has also helped us to test many variations in the methods by changing hyper-parameters, using different optimisers, ensembles, etc. With this information we have been able to conduct a comparative study that is useful for ensuring as complete control as possible of the methodology.

Read more
Data Analysis Statistics And Probability

Comparison of unfolding methods using RooFitUnfold

In this paper we describe RooFitUnfold, an extension of the RooFit statistical software package to treat unfolding problems, and which includes most of the unfolding methods that commonly used in particle physics. The package provides a common interface to these algorithms as well as common uniform methods to evaluate their performance in terms of bias, variance and coverage. In this paper we exploit this common interface of RooFitUnfold to compare the performance of unfolding with the Richardson-Lucy, Iterative Dynamically Stabilized, Tikhonov, Gaussian Process, Bin-by-bin and inversion methods on several example problems.

Read more
Data Analysis Statistics And Probability

Compressive Sensing for Dynamic XRF Scanning

X-Ray Fluorescence (XRF) scanning is a widespread technique of high importance and impact since it provides chemical composition maps crucial for several scientific investigations. There are continuous requirements for larger, faster and highly resolved acquisitions in order to study complex structures. Among the scientific applications that benefit from it, some of them, such as wide scale brain imaging, are prohibitively difficult due to time constraints. However, typically the overall XRF imaging performance is improving through technological progress on XRF detectors and X-ray sources. This paper suggests an additional approach where XRF scanning is performed in a sparse way by skipping specific points or by varying dynamically acquisition time or other scan settings in a conditional manner. This paves the way for Compressive Sensing in XRF scans where data are acquired in a reduced manner allowing for challenging experiments, currently not feasible with the traditional scanning strategies. A series of different compressive sensing strategies for dynamic scans are presented here. A proof of principle experiment was performed at the TwinMic beamline of Elettra synchrotron. The outcome demonstrates the potential of Compressive Sensing for dynamic scans, suggesting its use in challenging scientific experiments while proposing a technical solution for beamline acquisition software.

Read more
Data Analysis Statistics And Probability

Computational Techniques for the Analysis of Small Signals in High-Statistics Neutrino Oscillation Experiments

The current and upcoming generation of Very Large Volume Neutrino Telescopes---collecting unprecedented quantities of neutrino events---can be used to explore subtle effects in oscillation physics, such as (but not restricted to) the neutrino mass ordering. The sensitivity of an experiment to these effects can be estimated from Monte Carlo simulations. With the high number of events that will be collected, there is a trade-off between the computational expense of running such simulations and the inherent statistical uncertainty in the determined values. In such a scenario, it becomes impractical to produce and use adequately-sized sets of simulated events with traditional methods, such as Monte Carlo weighting. In this work we present a staged approach to the generation of binned event distributions in order to overcome these challenges. By combining multiple integration and smoothing techniques which address limited statistics from simulation it arrives at reliable analysis results using modest computational resources.

Read more
Data Analysis Statistics And Probability

Conception and software implementation of a nuclear data evaluation pipeline

We discuss the design and software implementation of a nuclear data evaluation pipeline applied for a fully reproducible evaluation of neutron-induced cross sections of 56 Fe above the resolved resonance region using the nuclear model code TALYS combined with relevant experimental data. The emphasis is on the mathematical and technical aspects of the pipeline and not on the evaluation of 56 Fe, which is tentative. The mathematical building blocks combined and employed in the pipeline are discussed in detail. A unified representation of experimental data, systematic and statistical errors, model parameters and defects enables the application of the Generalized Least Squares (GLS) and its natural extension, the Levenberg-Marquardt (LM) algorithm, on a large collection of experimental data. The LM algorithm tailored to nuclear data evaluation accounts for the exact non-linear physics model to determine best estimates of nuclear quantities. Associated uncertainty information is derived from a Taylor expansion at the maximum of the posterior distribution. We also discuss the pipeline in terms of its IT (=information technology) building blocks, such as those to efficiently manage and retrieve experimental data of the EXFOR library and to distribute computations on a scientific cluster. Relying on the mathematical and IT building blocks, we elaborate on the sequence of steps in the pipeline to perform the evaluation, such as the retrieval of experimental data, the correction of experimental uncertainties using marginal likelihood optimization (MLO) and after a screening of thousand TALYS parameters -- including Gaussian process priors on energy dependent parameters -- the fitting of about 150 parameters using the LM algorithm. The code of the pipeline including a manual and a Dockerfile for a simplified installation is available at this http URL.

Read more
Data Analysis Statistics And Probability

Conservation Laws and Spin System Modeling through Principal Component Analysis

This paper examines several applications of principal component analysis (PCA) to physical systems. The first of these demonstrates that the principal components in a basis of appropriate system variables can be employed to identify physically conserved quantities. That is, if the general form of a physical symmetry law is known, the PCA can identify an algebraic expression for the symmetry from the observed system trajectories. Secondly, the eigenvalue spectrum of the principal component spectrum for homogeneous periodic spin systems is found to reflect the geometric shape of the boundary. Finally, the PCA is employed to generate synthetic spin realizations with probability distributions in energy-magnetization space that closely resemble that of the input realizations although statistical quantities are inaccurately reproduced.

Read more
Data Analysis Statistics And Probability

Contamination Source Detection in Water Distribution Networks using Belief Propagation

We present a Bayesian approach for the Contamination Source Detection problem in Water Distribution Networks. Given an observation of contaminants in one or more nodes in the network, we try to give probable explanation for it assuming that contamination is a rare event. We introduce extra variables to characterize the place and pattern of the first contamination event. Then we write down the posterior distribution for these extra variables given the observation obtained by the sensors. Our method relies on Belief Propagation for the evaluation of the marginals of this posterior distribution and the determination of the most likely origin. The method is implemented on a simplified binary forward-in-time dynamics. Simulations on data coming from the realistic simulation software EPANET on two networks show that the simplified model is nevertheless flexible enough to capture crucial information about contaminant sources.

Read more
Data Analysis Statistics And Probability

Continuous cascades in the wavelet space as models for synthetic turbulence

We introduce a wide family of stochastic processes that are obtained as sums of self-similar localized "waveforms" with multiplicative intensity in the spirit of the Richardson cascade picture of turbulence. We establish the convergence and the minimum regularity of our construction. We show that its continuous wavelet transform is characterized by stochastic self-similarity and multifractal scaling properties. This model constitutes a stationary, "grid free", extension of W -cascades introduced in the past by Arneodo, Bacry and Muzy using wavelet orthogonal basis. Moreover our approach generically provides multifractal random functions that are not invariant by time reversal and therefore is able to account for skewed multifractal models and for the so-called "leverage effect". In that respect, it can be well suited to providing synthetic turbulence models or to reproducing the main observed features of asset price fluctuations in financial markets.

Read more
Data Analysis Statistics And Probability

Convolutional neural network for self-mixing interferometric displacement sensing

Self mixing interferometry is a well established interferometric measurement technique. In spite of the robustness and simplicity of the concept, interpreting the self-mixing signal is often complicated in practice, which is detrimental to measurement availability. Here we discuss the use of a convolutional neural network to reconstruct the displacement of a target from the self mixing signal in a semiconductor laser. The network, once trained on periodic displacement patterns, can reconstruct arbitrarily complex displacement in different alignment conditions and setups. The approach validated here is amenable to generalization to modulated schemes or even to totally different self mixing sensing tasks.

Read more

Ready to get started?

Join us today