Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Mark A. Przybocki is active.

Publication


Featured researches published by Mark A. Przybocki.


IEEE Computer | 2000

An introduction evaluating biometric systems

P J. Phillips; Alvin F. Martin; Charles L. Wilson; Mark A. Przybocki

On the basis of media hype alone, you might conclude that biometric passwords will soon replace their alphanumeric counterparts with versions that cannot be stolen, forgotten, lost, or given to another person. But what if the actual performance of these systems falls short of the estimates? The authors designed this article to provide sufficient information to know what questions to ask when evaluating a biometric system, and to assist in determining whether performance levels meet the requirements of an application. For example, a low-performance biometric is probably sufficient for reducing-as opposed to eliminating-fraud. Likewise, completely replacing an existing security system with a biometric-based one may require a high-performance biometric system, or the required performance may be beyond what current technology can provide. Of the biometrics that give the user some control over data acquisition, voice, face, and fingerprint systems have undergone the most study and testing-and therefore occupy the bulk of this discussion. This article also covers the tools and techniques of biometric testing.


Archive | 2005

The NIST speaker recognition evaluation program

Alvin F. Martin; Mark A. Przybocki; Joseph P. Campbell

The National Institute of Standards and Technology (NIST) has coordinated annual scientific evaluations of text-independent speaker recognition since 1996. These evaluations aim to provide important contributions to the direction of research efforts and the calibration of technical capabilities. They are intended to be of interest to all researchers working on the general problem of text-independent speaker recognition. To this end, the evaluations are designed to be simple, fully supported, accessible and focused on core technology issues. The evaluations have focused primarily on speaker detection in the context of conversational telephone speech. More recent evaluations have also included related tasks, such as speaker segmentation, and have used data in addition to conversational telephone speech. The evaluations are designed to foster research progress, with the objectives of:


Digital Signal Processing | 2000

The NIST 1999 Speaker Recognition Evaluation An Overview

Alvin F. Martin; Mark A. Przybocki

Martin, Alvin, and Przybocki, Mark, The NIST 1999 Speaker Recognition Evaluation?An Overview, Digital Signal Processing10(2000), 1?18.This article summarizes the 1999 NIST Speaker Recognition Evaluation. It discusses the overall research objectives, the three task definitions, the development and evaluation data sets, the specified performance measures and their manner of presentation, the overall quality of the results. More than a dozen sites from the United States, Europe, and Asia participated in this evaluation. There were three primary tasks for which automatic systems could be designed: one-speaker detection, two-speaker detection, and speaker tracking. All three tasks were performed in the context of mu-law encoded conversational telephone speech. The one-speaker detection task used single channel data, while the other two tasks used summed two-channel data. About 500 target speakers were specified, with 2 min of training speech data provided for each. Both multiple and single speaker test segments were selected from about 2000 conversations that were not used for training material. The duration of the multiple speaker test data was nominally 1 min, while the duration of the single speaker test segments varied from near zero up to 60 s. For each task, systems had to make independent decisions for selected combinations of a test segment and a hypothesized target speaker. The data sets for each task were designed to be large enough to provide statistically meaningful results on test subsets of interest. Results were analyzed with respect to various conditions including duration, pitch differences, and handset types.


Computer Speech & Language | 2006

NIST and NFI-TNO evaluations of automatic speaker recognition

David A. van Leeuwen; Alvin F. Martin; Mark A. Przybocki; Jos S. Bouten

In the past years, several text-independent speaker recognition evaluation campaigns have taken place. This paper reports on results of the NIST evaluation of 2004 and the NFI-TNO forensic speaker recognition evaluation held in 2003, and reflects on the history of the evaluation campaigns. The effects of speech duration, training handsets, transmission type, and gender mix show expected behaviour on the DET curves. New results on the influence of language show an interesting dependence of the DET curves on the accent of speakers. We also report on a number of statistical analysis techniques that have recently been introduced in the speaker recognition community, as well as a new application of the analysis of deviance analysis. These techniques are used to determine that the two evaluations held in 2003, by NIST and NFI-TNO, are of statistically different difficulty to the speaker recognition systems.


IEEE Transactions on Audio, Speech, and Language Processing | 2007

NIST Speaker Recognition Evaluations Utilizing the Mixer Corpora—2004, 2005, 2006

Mark A. Przybocki; Alvin F. Martin; Audrey N. Le

NIST has coordinated annual evaluations of text-independent speaker recognition from 1996 to 2006. This paper discusses the last three of these, which utilized conversational speech data from the Mixer Corpora recently collected by the Linguistic Data Consortium. We review the evaluation procedures, the matrix of test conditions included, and the performance trends observed. While most of the data is collected over telephone channels, one multichannel test condition utilizes a subset of Mixer conversations recorded simultaneously over multiple microphone channels and a telephone line. The corpus also includes some non-English conversations involving bilingual speakers, allowing an examination of the effect of language on performance results. On the various test conditions involving English language conversational telephone data, considerable performance gains are observed over the past three years.


2006 IEEE Odyssey - The Speaker and Language Recognition Workshop | 2006

NIST Speaker Recognition Evaluation Chronicles - Part 2

Mark A. Przybocki; Alvin F. Martin; Audrey N. Le

NIST has coordinated annual evaluations of text-independent speaker recognition since 1996. This update to an Odyssey 2004 paper concentrates on the past two years of the NIST evaluations. We discuss in particular the results of the 2004 and 2005 evaluations, and how they compare to earlier evaluation results. We also discuss the preparation and planning for the 2006 evaluation, which concludes with the evaluation workshop in San Juan, Puerto Rico, in June 2006


Machine Translation | 2009

The NIST 2008 Metrics for machine translation challenge--overview, methodology, metrics, and results

Mark A. Przybocki; Kay Peterson; Sebastien Bronsart; Gregory A. Sanders

This paper discusses the evaluation of automated metrics developed for the purpose of evaluating machine translation (MT) technology. A general discussion of the usefulness of automated metrics is offered. The NIST MetricsMATR evaluation of MT metrology is described, including its objectives, protocols, participants, and test data. The methodology employed to evaluate the submitted metrics is reviewed. A summary is provided for the general classes of evaluated metrics. Overall results of this evaluation are presented, primarily by means of correlation statistics, showing the degree of agreement between the automated metric scores and the scores of human judgments. Metrics are analyzed at the sentence, document, and system level with results conditioned by various properties of the test data. This paper concludes with some perspective on the improvements that should be incorporated into future evaluations of metrics for MT evaluation.


ieee international conference on data science and advanced analytics | 2015

The NIST data science initiative

Bonnie J. Dorr; Craig S. Greenberg; Peter C. Fontana; Mark A. Przybocki; Marion Le Bras; Cathryn A. Ploehn; Oleg Aulov; Martial Michel; E. Jim Golden; Wo Chang

We examine foundational issues in data science including current challenges, basic research questions, and expected advances, as the basis for a new Data Science Initiative and evaluation series, introduced by the National Institute of Standards and Technology (NIST) in the fall of 2015. The evaluations will facilitate research efforts, collaboration, leverage shared infrastructure, and effectively address cross-cutting challenges faced by diverse data science communities. The evaluations will have multiple research tracks championed by members of the data science community, and will enable rigorous comparison of approaches through common tasks, datasets, metrics, and shared research challenges. The tracks will measure several different data science technologies in a wide range of fields, starting with a pre-pilot. In addition to developing data science evaluation methods and metrics, it will address computing infrastructure, standards for an interoperability framework, and domain-specific examples.


Machine Translation | 2009

Introduction to the special issue on Automated Metrics for Machine Translation Evaluation

Alon Lavie; Mark A. Przybocki

The development of fully-automated programs that can effectively assess the translation quality performance of Machine Translation (MT) systems has blossomed into a vibrant research area in recent years. The broad interest in such “automated metrics for MT evaluation” is an outcome of the growing recognition of the role such metrics play in the development of high-performance MT systems. While human assessments of various forms serve as the ultimate arbitrators of translation quality, the training and development of modern MT systems critically rely on automated metrics. The increasingly dominant data-driven machine-learning-based paradigm for Machine Translation is inherently dependent on automated translation quality measures which can serve as target functions for optimizing MT system performance during training. It has been increasingly recognized in recent years that perhaps the single most important enabling factor of modern MT research progress has been the advent and broad adoption of IBM’s BLEU metric. BLEU, however, provides only a rough approximation of translation quality. As MT systems improve, there is an increasing need for better metrics that exhibit high-levels of correlation with human judgments of translation quality. This has resulted in the development of an expanding number of alternative metrics in recent years. Evaluating the performance, efficacy and utility of these various automated metrics has consequently become a significant challenge of its own.


Journal of data science | 2016

A new data science research program: evaluation, metrology, standards, and community outreach

Bonnie J. Dorr; Craig S. Greenberg; Peter C. Fontana; Mark A. Przybocki; Marion Le Bras; Cathryn A. Ploehn; Oleg Aulov; Martial Michel; E. Jim Golden; Wo Chang

This article examines foundational issues in data science including current challenges, basic research questions, and expected advances, as the basis for a new data science research program (DSRP) and associated data science evaluation (DSE) series, introduced by the National Institute of Standards and Technology (NIST) in the fall of 2015. The DSRP is designed to facilitate and accelerate research progress in the field of data science and consists of four components: evaluation and metrology, standards, compute infrastructure, and community outreach. A key part of the evaluation and measurement component is the DSE. The DSE series aims to address logistical and evaluation design challenges while providing rigorous measurement methods and an emphasis on generalizability rather than domain- and application-specific approaches. Toward that end, each year the DSE will consist of multiple research tracks and will encourage the application of tasks that span these tracks. The evaluations are intended to facilitate research efforts and collaboration, leverage shared infrastructure, and effectively address crosscutting challenges faced by diverse data science communities. Multiple research tracks will be championed by members of the data science community with the goal of enabling rigorous comparison of approaches through common tasks, datasets, metrics, and shared research challenges. The tracks will permit us to measure several different data science technologies in a wide range of fields and will address computing infrastructure, standards for an interoperability framework, and domain-specific examples. This article also summarizes lessons learned from the data science evaluation series pre-pilot that was held in fall of 2015.

Collaboration


Dive into the Mark A. Przybocki's collaboration.

Top Co-Authors

Avatar

Alvin F. Martin

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar

Craig S. Greenberg

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar

David S. Pallett

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar

Jonathan G. Fiscus

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar

Marion Le Bras

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar

Peter C. Fontana

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar

Cathryn A. Ploehn

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar

Oleg Aulov

University of Maryland

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Audrey N. Le

National Institute of Standards and Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge