Featured Researches

Data Analysis Statistics And Probability

A fluctuation theorem for time-series of signal-response models with the backward transfer entropy

The irreversibility of trajectories in stochastic dynamical systems is linked to the structure of their causal representation in terms of Bayesian networks. We consider stochastic maps resulting from a time discretization with interval \tau of signal-response models, and we find an integral fluctuation theorem that sets the backward transfer entropy as a lower bound to the conditional entropy production. We apply this to a linear signal-response model providing analytical solutions, and to a nonlinear model of receptor-ligand systems. We show that the observational time \tau has to be fine-tuned for an efficient detection of the irreversibility in time-series.

Read more
Data Analysis Statistics And Probability

A general procedure for detector-response correction of higher order cumulants

We propose a general procedure for the detector-response correction (including efficiency correction) of higher order cumulants observed by the event-by-event analysis in heavy-ion collisions. This method makes use of the moments of the response matrix characterizing the property of a detector, and is applicable to a wide variety of response matrices such as those having non-binomial responses and including the effects of ghost tracks. A procedure to carry out the detector-response correction of realistic detectors is discussed. In test analyses, we show that this method can successfully re- construct the cumulants of true distribution for various response matrices including the one having multiplicity-dependent efficiency.

Read more
Data Analysis Statistics And Probability

A generalized permutation entropy for random processes

Permutation entropy measures the complexity of deterministic time series via a data symbolic quantization consisting of rank vectors called ordinal patterns or just permutations. The reasons for the increasing popularity of this entropy in time series analysis include that (i) it converges to the Kolmogorov-Sinai entropy of the underlying dynamics in the limit of ever longer permutations, and (ii) its computation dispenses with generating and ad hoc partitions. However, permutation entropy diverges when the number of allowed permutations grows super-exponentially with their length, as is usually the case when time series are output by random processes. In this Letter we propose a generalized permutation entropy that is finite for random processes, including discrete-time dynamical systems with observational or dynamical noise.

Read more
Data Analysis Statistics And Probability

A geometric approach to the transport of discontinuous densities

Different observations of a relation between inputs ("sources") and outputs ("targets") are often reported in terms of histograms (discretizations of the source and the target densities). Transporting these densities to each other provides insight regarding the underlying relation. In (forward) uncertainty quantification, one typically studies how the distribution of inputs to a system affects the distribution of the system responses. Here, we focus on the identification of the system (the transport map) itself, once the input and output distributions are determined, and suggest a modification of current practice by including data from what we call "an observation process". We hypothesize that there exists a smooth manifold underlying the relation; the sources and the targets are then partial observations (possibly projections) of this manifold. Knowledge of such a manifold implies knowledge of the relation, and thus of "the right" transport between source and target observations. When the source-target observations are not bijective (when the manifold is not the graph of a function over both observation spaces, either because folds over them give rise to density singularities, or because it marginalizes over several observables), recovery of the manifold is obscured. Using ideas from attractor reconstruction in dynamical systems, we demonstrate how additional information in the form of short histories of an observation process can help us recover the underlying manifold. The types of additional information employed and the relation to optimal transport based solely on density observations is illustrated and discussed, along with limitations in the recovery of the true underlying relation.

Read more
Data Analysis Statistics And Probability

A machine learning framework for computationally expensive transient models

The promise of machine learning has been explored in a variety of scientific disciplines in the last few years, however, its application on first-principles based computationally expensive tools is still in nascent stage. Even with the advances in computational resources and power, transient simulations of large-scale dynamic systems using a variety of the first-principles based computational tools are still limited. In this work, we propose an ensemble approach where we combine one such computationally expensive tool, called discrete element method (DEM), with a time-series forecasting method called auto-regressive integrated moving average (ARIMA) and machine-learning methods to significantly reduce the computational burden while retaining model accuracy and performance. The developed machine-learning model shows good predictability and agreement with the literature, demonstrating its tremendous potential in scientific computing.

Read more
Data Analysis Statistics And Probability

A machine learning route between band mapping and band structure

The electronic band structure (BS) of solid state materials imprints the multidimensional and multi-valued functional relations between energy and momenta of periodically confined electrons. Photoemission spectroscopy is a powerful tool for its comprehensive characterization. A common task in photoemission band mapping is to recover the underlying quasiparticle dispersion, which we call band structure reconstruction. Traditional methods often focus on specific regions of interests yet require extensive human oversight. To cope with the growing size and scale of photoemission data, we develop a generic machine-learning approach leveraging the information within electronic structure calculations for this task. We demonstrate its capability by reconstructing all fourteen valence bands of tungsten diselenide and validate the accuracy on various synthetic data. The reconstruction uncovers previously inaccessible momentum-space structural information on both global and local scales in conjunction with theory, while realizing a path towards integrating band mapping data into materials science databases.

Read more
Data Analysis Statistics And Probability

A multi-sensor data-driven methodology for all-sky passive microwave inundation retrieval

We present a multi-sensor Bayesian passive microwave retrieval algorithm for flood inundation mapping at high spatial and temporal resolutions. The algorithm takes advantage of observations from multiple sensors in optical, short-infrared, and microwave bands, thereby allowing for detection and mapping of the sub-pixel fraction of inundated areas under almost all-sky conditions. The method relies on a nearest-neighbor search and a modern sparsity-promoting inversion method that make use of an a priori dataset in the form of two joint dictionaries. These dictionaries contain almost overlapping observations by the Special Sensor Microwave Imager and Sounder (SSMIS) on board the Defense Meteorological Satellite Program (DMSP) F17 satellite and the Moderate Resolution Imaging Spectroradiometer (MODIS) on board the Aqua and Terra satellites. Evaluation of the retrieval algorithm over the Mekong Delta shows that it is capable of capturing to a good degree the inundation diurnal variability due to localized convective precipitation. At longer timescales, the results demonstrate consistency with the ground-based water level observations, denoting that the method is properly capturing inundation seasonal patterns in response to regional monsoonal rain. The calculated Euclidean distance, rank-correlation, and also copula quantile analysis demonstrate a good agreement between the outputs of the algorithm and the observed water levels at monthly and daily timescales. The current inundation products are at a resolution of 12.5 km and taken twice per day, but a higher resolution (order of 5 km and every 3 h) can be achieved using the same algorithm with the dictionary populated by the Global Precipitation Mission (GPM) Microwave Imager (GMI) products.

Read more
Data Analysis Statistics And Probability

A new "3D Calorimetry" of hot nuclei

In the domain of Fermi energy, it is extremely complex to isolate experimentally fragments and particles issued from the cooling of a hot nucleus produced during a heavy ion collision. This paper presents a new method to characterize more precisely hot Quasi-Projectiles. It tries to take into account as accurately as possible the distortions generated by all the other potential participants in the nuclear reaction. It is quantitatively shown that this method is a major improvement respect to classic calorimetries used with a 4 π detector array. By detailing and deconvolving the different steps of the reconstitution of the hot nucleus, this study shows also the respective role played by the experimental device and the event selection criteria on the quality of the determination of QP characteristics.

Read more
Data Analysis Statistics And Probability

A new Monte Carlo-based fitting method

We present a new fitting technique based on the parametric bootstrap method, which relies on the idea to produce artificial measurements using the estimated probability distribution of the experimental data. In order to investigate the main properties of this technique, we develop a toy model and we analyze several fitting conditions with a comparison of our results to the ones obtained using both the standard χ 2 minimization procedure and a Bayesian approach. Furthermore, we investigate the effect of the data systematic uncertainties both on the probability distribution of the fit parameters and on the shape of the expected goodness-of-fit distribution. Our conclusion is that, when systematic uncertainties are included in the analysis, only the bootstrap procedure is able to provide reliable confidence intervals and p-values, thus improving the results given by the standard χ 2 minimization approach. Our technique is then applied to an actual physics process, the real Compton scattering off the proton, thus confirming both the portability and the validity of the bootstrap-based fit method.

Read more
Data Analysis Statistics And Probability

A new approach on estimating the fluid temperature in a multiphase flow system using particle filter method

Fluid temperature is important for the analysis of the heat transfers in thermal hydraulics. An accurate measurement or estimation of the fluid temperature in multiphase flows is challenging. This is due to that the thermocouple signal that mixes with temperature signals for each phase and non-negligible noises. This study provides a new approach to estimate the local fluid temperature in multiphase flows using experimental time-series temperature signal. The thermocouple signal is considered to be a sequence with Markov property and the particle filter method is utilized in the new method to extract the fluid temperature. A complete description of the new method is presented in this article.

Read more

Ready to get started?

Join us today