Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Lynetta Sacherek is active.

Publication


Featured researches published by Lynetta Sacherek.


Bulletin of The Medical Library Association | 2002

Factors associated with successful answering of clinical questions using an information retrieval system

William R. Hersh; M. Katherine Crabtree; David H. Hickam; Lynetta Sacherek; Charles P. Friedman; Patricia Tidmarsh; Craig Mosbaek; Dale F. Kraemer

OBJECTIVES This study sought to assess the ability of medical and nurse practitioner students to use MEDLINE to obtain evidence for answering clinical questions and to identify factors associated with the successful answering of questions. METHODS A convenience sample of medical and nurse practitioner students was recruited. After completing instruments measuring demographic variables, computer and searching attitudes and experience, and cognitive traits, the subjects were given a brief orientation to MEDLINE searching and the techniques of evidence-based medicine. The subjects were then given 5 questions (from a pool of 20) to answer in two sessions using the Ovid MEDLINE system and the Oregon Health & Science University library collection. Each question was answered using three possible responses that reflected the quality of the evidence. All actions capable of being logged by the Ovid system were captured. Statistical analysis was performed using a model based on generalized estimating equations. The relevance-based measures of recall and precision were measured by defining end queries and having relevance judgments made by physicians who were not associated with the study. RESULTS Forty-five medical and 21 nurse practitioner students provided usable answers to 324 questions. The rate of correctness increased from 32.3 to 51.6 percent for medical students and from 31.7 to 34.7 percent for nurse practitioner students. Ability to answer questions correctly was most strongly associated with correctness of the answer before searching, user experience with MEDLINE features, the evidence-based medicine question type, and the spatial visualization score. The spatial visualization score showed multi-colinearity with student type (medical vs. nurse practitioner). Medical and nurse practitioner students obtained comparable recall and precision, neither of which was associated with correctness of the answer. CONCLUSIONS Medical and nurse practitioner students in this study were at best moderately successful at answering clinical questions correctly with the assistance of literature searching. The results confirm the importance of evaluating both search ability and the ability to use the resulting information to accomplish a clinical task.


international acm sigir conference on research and development in information retrieval | 2000

Do batch and user evaluations give the same results

William R. Hersh; Andrew Turpin; Susan Price; Benjamin Chan; Dale Kramer; Lynetta Sacherek; Daniel Olson

Do improvements in system performance demonstrated by batch evaluations confer the same benefit for real users? We carried out experiments designed to investigate this question. After identifying a weighting scheme that gave maximum improvement over the baseline in a non-interactive evaluation, we used it with real users searching on an instance recall task. Our results showed the weighting scheme giving beneficial results in batch studies did not do so with real users. Further analysis did identify other factors predictive of instance recall, including number of documents saved by the user, document recall, and number of documents seen by the user.


text retrieval conference | 2001

Challenging conventional assumptions of automated information retrieval with real users: Boolean searching and batch retrieval evaluations

William R. Hersh; Andrew Turpin; Susan Price; Dale F. Kraemer; Daniel Olson; Benjamin Chan; Lynetta Sacherek

Two common assumptions held by information retrieval researchers are that searching using Boolean operators is inferior to natural language searching and that results from batch-style retrieval evaluations are generalizable to the real-world searching. We challenged these assumptions in the Text Retrieval Conference (TREC) interactive track, with real users following a consensus protocol to search for an instance recall task. Our results showed that Boolean and natural language searching achieved comparable results and that the results from batch evaluations were not comparable to those obtained in experiments with real users.


JAMA | 1998

Applicability and Quality of Information for Answering Clinical Questions on the Web

William R. Hersh; Paul N. Gorman; Lynetta Sacherek


Journal of the American Medical Informatics Association | 2002

Research Paper: Factors Associated with Success in Searching MEDLINE and Applying Evidence to Answer Clinical Questions.

William R. Hersh; M. Katherine Crabtree; David H. Hickam; Lynetta Sacherek; Charles P. Friedman; Patricia Tidmarsh; Craig Mosbaek; Dale F. Kraemer


text retrieval conference | 1998

A Large-Scale Comparison of Boolean vs. Natural Language Searching for the TREC-7 Interactive Track.

William R. Hersh; Susan Price; Dale F. Kraemer; Benjamin Chan; Lynetta Sacherek; Daniel Olson


text retrieval conference | 1999

Do Batch and User Evaluations Give the Same Results? An Analysis from the TREC-8 Interactive Track

William R. Hersh; Andrew Turpin; Susan Price; Dale F. Kraemer; Benjamin Chan; Lynetta Sacherek; Daniel Olson


american medical informatics association annual symposium | 1999

Maintaining a catalog of manually-indexed, clinically-oriented World Wide Web content.

William R. Hersh; Andrea Ball; Bikram Day; Mary Masterson; Li Zhang; Lynetta Sacherek


text retrieval conference | 2000

Further Analysis of Whether Batch and User Evaluations Give the Same Results with a Question-Answering Task.

William R. Hersh; Andrew Turpin; Lynetta Sacherek; Daniel Olson; Susan Price; Benjamin Chan; Dale F. Kraemer


american medical informatics association annual symposium | 2002

Professional's Information Link (PiL): a web-based asynchronous consultation service.

William R. Hersh; Robin Miller; Daniel Olson; Lynetta Sacherek; Ping Cross

Collaboration


Dive into the Lynetta Sacherek's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Susan Price

Portland State University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge