Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Scott Vander Wiel is active.

Publication


Featured researches published by Scott Vander Wiel.


Statistical Science | 2006

Monitoring Networked Applications With Incremental Quantile Estimation

John M. Chambers; David A. James; Diane Lambert; Scott Vander Wiel

Networked applications have software components that reside on different computers. Email, for example, has database, processing, and user interface components that can be distributed across a network and shared by users in different locations or work groups. End-to-end performance and reliability metrics describe the software quality experienced by these groups of users, taking into account all the software components in the pipeline. Each user produces only some of the data needed to understand the quality of the application for the group, so group performance metrics are obtained by combining summary statistics that each end computer periodically (and automatically) sends to a central server. The group quality metrics usually focus on medians and tail quantiles rather than on averages. Distributed quantile estimation is challenging, though, especially when passing large amounts of data around the network solely to compute quality metrics is undesirable. This paper describes an Incremental Quantile (IQ) estimation method that is designed for performance monitoring at arbitrary levels of network aggregation and time resolution when only a limited amount of data can be transferred. Applications to both real and simulated data are provided.


IEEE Transactions on Power Systems | 2016

Line Outage Localization Using Phasor Measurement Data in Transient State

Manuel J. Garcia; Thomas A. Catanach; Scott Vander Wiel; Russell Bent; Earl Lawrence

This paper introduces a statistical classifier that quickly locates line outages in a power system utilizing only time series phasor measurement data sampled during the systems transient response to the outage. The presented classifier is a linear multinomial regression model that is trained by solving a maximum likelihood optimization problem using synthetic data. The synthetic data is produced through dynamic simulations which are initialized by random samples of a forecast load/generation distribution. Real time computation of the proposed classifier is minimal and therefore the classifier is capable of locating a line outage before steady state is reached, allowing for quick corrective action in response to an outage. In addition, the output of the classifier fits into a statistical framework that is easily accessible. Specific line outages are identified as being difficult to localize and future improvements to the classifier are proposed.


The Annals of Applied Statistics | 2014

Stochastic identification of malware with dynamic traces

Curtis B. Storlie; Blake Anderson; Scott Vander Wiel; Daniel Quist; Curtis Hash; Nathan Brown

A novel approach to malware classification is introduced based on analysis of instruction traces that are collected dynamically from the program in question. The method has been implemented online in a sandbox environment (i.e., a security mechanism for separating running programs) at Los Alamos National Laboratory, and is intended for eventual host-based use, provided the issue of sampling the instructions executed by a given process without disruption to the user can be satisfactorily addressed. The procedure represents an instruction trace with a Markov chain structure in which the transition matrix,


The Astronomical Journal | 2017

The Nonhomogeneous Poisson Process for Fast Radio Burst Rates

Earl Lawrence; Scott Vander Wiel; Casey J. Law; Sarah Burke Spolaor; Geoffrey C. Bower

\mathbf {P}


Journal of Computational and Graphical Statistics | 2012

Developing Systems for Real-Time Streaming Analysis

Sarah Michalak; Andrew J. DuBois; David H. DuBois; Scott Vander Wiel; John Hogden

, has rows modeled as Dirichlet vectors. The malware class (malicious or benign) is modeled using a flexible spline logistic regression model with variable selection on the elements of


Technometrics | 2013

Model Bank State Estimation for Power Grids Using Importance Sampling

Earl Lawrence; Scott Vander Wiel; Russell Bent

\mathbf {P}


Siam Journal on Optimization | 2007

Statistical Quasi-Newton: A New Look at Least Change

Chuanhai Liu; Scott Vander Wiel

, which are observed with error. The utility of the method is illustrated on a sample of traces from malware and nonmalware programs, and the results are compared to other leading detection schemes (both signature and classification based). This article also has supplementary materials available online.


Technometrics | 2011

A Random Onset Model for Degradation of High-Reliability Systems

Scott Vander Wiel; Alyson G. Wilson; Todd L. Graves; C. Shane Reese

This paper presents the nonhomogeneous Poisson process (NHPP) for modeling the rate of fast radio bursts (FRBs) and other infrequently observed astronomical events. The NHPP, well-known in statistics, can model the dependence of the rate on both astronomical features and the details of an observing campaign. This is particularly helpful for rare events like FRBs because the NHPP can combine information across surveys, making the most of all available information. The goal of the paper is two-fold. First, it is intended to be a tutorial on the use of the NHPP. Second, we build an NHPP model that incorporates beam patterns and a power-law flux distribution for the rate of FRBs. Using information from 12 surveys including 15 detections, we find an all-sky FRB rate of 587 events per sky per day above a flux of 1 Jy (95% CI: 272, 924) and a flux power-law index of (95% CI: 0.57, 1.25). Our rate is lower than other published rates, but consistent with the rate given in Champion et al.


Statistical Science | 2010

Parameter Expansion and Efficient Inference

Andrew Lewandowski; Chuanhai Liu; Scott Vander Wiel

Sources of streaming data are proliferating and so are the demands to analyze and mine such data in real time. Statistical methods frequently form the core of real-time analysis, and therefore, statisticians increasingly encounter the challenges and implicit requirements of real-time systems. This work recommends a comprehensive strategy for development and implementation of streaming algorithms, beginning with exploratory data analysis in a flexible computing environment, leading to specification of a computational algorithm for the streaming setting and its initial implementation, and culminating in successive improvements to computational efficiency and throughput. This sequential development relies on a collaboration between statisticians, domain scientists, and the computer engineers developing the real-time system. This article illustrates the process in the context of a radio astronomy challenge to mitigate adverse impacts of radio frequency interference (noise) in searches for high-energy impulses from distant sources. The radio astronomy application motivates discussion of system design, code optimization, and the use of hardware accelerators such as graphics processing units, field-programmable gate arrays, and IBM Cell processors. Supplementary materials, available online, detail the computing systems typically used for streaming systems with real-time constraints and the process of optimizing code for high efficiency and throughput.


Proceedings of the Institution of Mechanical Engineers, Part O: Journal of Risk and Reliability | 2017

The negative log-gamma prior distribution for Bayesian assessment of system reliability:

Roger S. Zoh; Alyson G. Wilson; Scott Vander Wiel; Earl Lawrence

Power grid operators decide where and how much power to generate based on the current topology and demands of the network. The topology can change as safety devices trigger (connecting or disconnecting parts of the network) or as lines go down. Often, the operator cannot observe these events directly, but instead has contemporary measurements and historical information about a subset of the line flows and bus (node) properties. This information can be used in conjunction with a computational model to infer the topology of the network. We present a Bayesian approach to topological inference that considers a bank of possible topologies. The solution provides a probability for each member in the model bank. The approach has two important features. First, we build a statistical approximation, or emulator, to the computational model, which is too computationally expensive to run a large number of times. Second, we use the emulator in an importance sampling scheme to estimate the probabilities. The resulting algorithm is fast enough to use in real time and very accurate. This article has online supplementary materials.

Collaboration


Dive into the Scott Vander Wiel's collaboration.

Top Co-Authors

Avatar

Earl Lawrence

Los Alamos National Laboratory

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Brian Weaver

Los Alamos National Laboratory

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Russell Bent

Los Alamos National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Alyson G. Wilson

North Carolina State University

View shared research outputs
Top Co-Authors

Avatar

Curtis B. Storlie

Los Alamos National Laboratory

View shared research outputs
Top Co-Authors

Avatar

David E. Sigeti

Los Alamos National Laboratory

View shared research outputs
Researchain Logo
Decentralizing Knowledge