Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Samuel Bayer is active.

Publication


Featured researches published by Samuel Bayer.


International Journal of Medical Informatics | 2010

The MITRE Identification Scrubber Toolkit: Design, training, and assessment

John S. Aberdeen; Samuel Bayer; Reyyan Yeniterzi; Benjamin Wellner; Cheryl Clark; David A. Hanauer; Bradley Malin; Lynette Hirschman

PURPOSE Medical records must often be stripped of patient identifiers, or de-identified, before being shared. De-identification by humans is time-consuming, and existing software is limited in its generality. The open source MITRE Identification Scrubber Toolkit (MIST) provides an environment to support rapid tailoring of automated de-identification to different document types, using automatically learned classifiers to de-identify and protect sensitive information. METHODS MIST was evaluated with four classes of patient records from the Vanderbilt University Medical Center: discharge summaries, laboratory reports, letters, and order summaries. We trained and tested MIST on each class of record separately, as well as on pooled sets of records. We measured precision, recall, F-measure and accuracy at the word level for the detection of patient identifiers as designated by the HIPAA Safe Harbor Rule. RESULTS MIST was applied to medical records that differed in the amounts and types of protected health information (PHI): lab reports contained only two types of PHI (dates, names) compared to discharge summaries, which were much richer. Performance of the de-identification tool depended on record class; F-measure results were 0.996 for order summaries, 0.996 for discharge summaries, 0.943 for letters and 0.934 for laboratory reports. Experiments suggest the tool requires several hundred training exemplars to reach an F-measure of at least 0.9. CONCLUSIONS The MIST toolkit makes possible the rapid tailoring of automated de-identification to particular document types and supports the transition of the de-identification software to medical end users, avoiding the need for developers to have access to original medical records. We are making the MIST toolkit available under an open source license to encourage its application to diverse data sets at multiple institutions.


international conference on spoken language processing | 1996

Embedding speech in Web interfaces

Samuel Bayer

We describe work in progress at the MITRE Corporation on embedding speech enabled interfaces in Web browsers. This research is part of our work to establish the infrastructure to create Web hosted versions of prototype multimodal interfaces, both intelligent and otherwise. Like many others, we believe that the Web is the best potential delivery and distribution vehicle for complex software applications, and that the functionality of these Web hosted applications should match the functionality available in standalone applications. We discuss our approach to several aspects of this goal.


International Journal of Medical Informatics | 2013

Bootstrapping a de-identification system for narrative patient records: Cost-performance tradeoffs

David A. Hanauer; John S. Aberdeen; Samuel Bayer; Benjamin Wellner; Cheryl Clark; Kai Zheng; Lynette Hirschman

PURPOSE We describe an experiment to build a de-identification system for clinical records using the open source MITRE Identification Scrubber Toolkit (MIST). We quantify the human annotation effort needed to produce a system that de-identifies at high accuracy. METHODS Using two types of clinical records (history and physical notes, and social work notes), we iteratively built statistical de-identification models by annotating 10 notes, training a model, applying the model to another 10 notes, correcting the models output, and training from the resulting larger set of annotated notes. This was repeated for 20 rounds of 10 notes each, and then an additional 6 rounds of 20 notes each, and a final round of 40 notes. At each stage, we measured precision, recall, and F-score, and compared these to the amount of annotation time needed to complete the round. RESULTS After the initial 10-note round (33min of annotation time) we achieved an F-score of 0.89. After just over 8h of annotation time (round 21) we achieved an F-score of 0.95. Number of annotation actions needed, as well as time needed, decreased in later rounds as model performance improved. Accuracy on history and physical notes exceeded that of social work notes, suggesting that the wider variety and contexts for protected health information (PHI) in social work notes is more difficult to model. CONCLUSIONS It is possible, with modest effort, to build a functioning de-identification system de novo using the MIST framework. The resulting system achieved performance comparable to other high-performing de-identification systems.


Intelligence\/sigart Bulletin | 1991

The relation-based knowledge representation of King Kong

Samuel Bayer; Marc B. Vilain

This paper presents an overview of the knowledge representation facilities of King Kong a transportable natural language system. The thrust of the paper is towards demonstrating how the particulars of Kongs representation language support the processing of key phenomena of natural language. To this extent, we cover Kongs terminological hierarchies; the logical language in which Kong encodes utterance meanings; and the query evaluator that interfaces Kong to expert system back ends. We focus particularly on King Kongs innovative treatment of relations, as this treatment provides the system with much of its language analysis strengths.


ACM Computing Surveys | 1999

The MITRE Multi-Modal Logger: its use in evaluation of collaborative systems

Samuel Bayer; Laurie E. Damianos; Robyn Kozierok; James Mokwa

Collaborative environments have grown quite sophisticated over the years. They have evolved from simple text conferencing tools [2] to full suites of integrated multi-modal tools such as Habanero [9] or MITREs room-based Collaborative Virtual Workspace [3, 4]. As a result, evaluating multi-modal collaborative systems is an order of magnitude more complex than single-user, single-application HCI evaluation. In the context of the DARPA Intelligent Collaboration &Visualization (IC&V) program, the MITRE HCI group has adapted its logging technology to the task of evaluating collaborative systems. The MITRE Multi-Modal Logger (MML) is a system for recording, retrieving, annotating, and visualizing data. In this paper, we describe the capabilities of the MML and its application to research including the IC&V Evaluation Working Groups work [5, 6, 7, 8, 13].


data integration in the life sciences | 2012

Validating candidate gene-mutation relations in MEDLINE abstracts via crowdsourcing

John D. Burger; Emily Doughty; Samuel Bayer; David Tresner-Kirsch; Ben Wellner; John S. Aberdeen; Kyungjoon Lee; Maricel G. Kann; Lynette Hirschman

We describe an experiment to elicit judgments on the validity of gene-mutation relations in MEDLINE abstracts via crowdsourcing. The biomedical literature contains rich information on such relations, but the correct pairings are difficult to extract automatically because a single abstract may mention multiple genes and mutations. We ran an experiment presenting candidate gene-mutation relations as Amazon Mechanical Turk HITs (human intelligence tasks). We extracted candidate mutations from a corpus of 250 MEDLINE abstracts using EMU combined with curated gene lists from NCBI. The resulting document-level annotations were projected into the abstract text to highlight mentions of genes and mutations for review. Reviewers returned results within 36 hours. Initial weighted results evaluated against a gold standard of expert curated gene-mutation relations achieved 85% accuracy, with the best reviewer achieving 91% accuracy. We expect performance to increase with further experimentation, providing a scalable approach for rapid manual curation of important biological relations.


north american chapter of the association for computational linguistics | 2004

MiTAP for SARS detection

Laurie E. Damianos; Samuel Bayer; Michael Chisholm; John C. Henderson; Lynette Hirschman; William T. Morgan; Marc Ubaldino; Guido Zarrella; James M. Wilson; Marat G. Polyak

The MiTAP prototype for SARS detection uses human language technology for detecting, monitoring, and analyzing potential indicators of infectious disease outbreaks and reasoning for issuing warnings and alerts. MiTAP focuses on providing timely, multilingual information access to analysts, domain experts, and decision-makers worldwide. Data sources are captured, filtered, translated, summarized, and categorized by content. Critical information is automatically extracted and tagged to facilitate browsing, searching, and scanning, and to provide key terms at a glance. The processed articles are made available through an easy-to-use news server and cross-language information retrieval system for access and analysis anywhere, any time. Specialized newsgroups and customizable filters or searches on incoming stories allow users to create their own view into the data while a variety of tools summarize, indicate trends, and provide alerts to potentially relevant spikes of activity.


international conference on human language technology research | 2001

Exploring speech-enabled dialogue with the Galaxy Communicator infrastructure

Samuel Bayer; Christine Doran; Bryan George

This demonstration will motivate some of the significant properties of the Galaxy Communicator Software Infrastructure and show how they support the goals of the DARPA Communicator program.


international conference on human language technology research | 2001

Dialogue interaction with the DARPA communicator infrastructure: the development of useful software

Samuel Bayer; Christine Doran; Bryan George

To support engaging human users in robust, mixed-initiative speech dialogue interactions which reach beyond current capabilities in dialogue systems, the DARPA Communicator program [1] is funding the development of a distributed message-passing infrastructure for dialogue systems which all Communicator participants are using. In this presentation, we describe the features of and requirements for a genuinely useful software infrastructure for this purpose.


international conference on machine learning | 2005

Evaluating semantic evaluations: how RTE measures up

Samuel Bayer; John D. Burger; Lisa Ferro; John C. Henderson; Lynette Hirschman; Alexander S. Yeh

In this paper, we discuss paradigms for evaluating open-domain semantic interpretation as they apply to the PASCAL Recognizing Textual Entailment (RTE) evaluation (Dagan et al. 2005). We focus on three aspects critical to a successful evaluation: creation of large quantities of reasonably good training data, analysis of inter-annotator agreement, and joint analysis of test item difficulty and test-taker proficiency (Rasch analysis). We found that although RTE does not correspond to a “real” or naturally occurring language processing task, it nonetheless provides clear and simple metrics, a tolerable cost of corpus development, good annotator reliability (with the potential to exploit the remaining variability), and the possibility of finding noisy but plentiful training material.

Collaboration


Dive into the Samuel Bayer's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge