Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Penny Whiting is active.

Publication


Featured researches published by Penny Whiting.


BMC Medical Research Methodology | 2003

The development of QUADAS: a tool for the quality assessment of studies of diagnostic accuracy included in systematic reviews

Penny Whiting; Anne Wilhelmina Saskia Rutjes; Johannes B. Reitsma; Patrick M. Bossuyt; Jos Kleijnen

BackgroundIn the era of evidence based medicine, with systematic reviews as its cornerstone, adequate quality assessment tools should be available. There is currently a lack of a systematically developed and evaluated tool for the assessment of diagnostic accuracy studies. The aim of this project was to combine empirical evidence and expert opinion in a formal consensus method to develop a tool to be used in systematic reviews to assess the quality of primary studies of diagnostic accuracy.MethodsWe conducted a Delphi procedure to develop the quality assessment tool by refining an initial list of items. Members of the Delphi panel were experts in the area of diagnostic research. The results of three previously conducted reviews of the diagnostic literature were used to generate a list of potential items for inclusion in the tool and to provide an evidence base upon which to develop the tool.ResultsA total of nine experts in the field of diagnostics took part in the Delphi procedure. The Delphi procedure consisted of four rounds, after which agreement was reached on the items to be included in the tool which we have called QUADAS. The initial list of 28 items was reduced to fourteen items in the final tool. Items included covered patient spectrum, reference standard, disease progression bias, verification bias, review bias, clinical review bias, incorporation bias, test execution, study withdrawals, and indeterminate results. The QUADAS tool is presented together with guidelines for scoring each of the items included in the tool.ConclusionsThis project has produced an evidence based quality assessment tool to be used in systematic reviews of diagnostic accuracy studies. Further work to determine the usability and validity of the tool continues.


Annals of Internal Medicine | 2004

Sources of variation and bias in studies of diagnostic accuracy: a systematic review.

Penny Whiting; Anne W S Rutjes; Johannes B Reitsma; Afina S Glas; Patrick M. M. Bossuyt; Jos Kleijnen

Diagnostic tests are of crucial importance in health care. They are performed to reduce uncertainty concerning whether a patient has a condition of interest. A thorough evaluation of diagnostic tests is necessary to ensure that only accurate tests are used in practice. Diagnostic accuracy studies are a vital step in this evaluation process. Diagnostic accuracy studies aim to investigate how well the results from a test being evaluated (index test) agree with the results of the reference standard. The reference standard is considered the best available method to establish the presence or absence of a condition (target condition). In a classic diagnostic accuracy study, a consecutive series of patients who are suspected of having the target condition undergo the index test; then, all patients are verified by the same reference standard. The index test and reference standard are then read by persons blinded to the results of each, and various measures of agreement are calculated (for example, sensitivity, specificity, likelihood ratios, and diagnostic odds ratios). This classic design has many variations, including differences in the way patients are selected for the study, in test protocol, in the verification of patients, and in the way the index test and reference standard are read. Some of these differences may bias the results of a study, whereas others may limit the applicability of results. Bias is said to be present in a study if distortion is introduced as a consequence of defects in the design or conduct of a study. Therefore, a biased diagnostic accuracy study will produce estimates of test performance that differ from the true performance of the test. In contrast, variability arises from differences among studies, for example, in terms of population, setting, test protocol, or definition of the target disorder (1). Although variability does not lead to biased estimates of test performance, it may limit the applicability of results and thus is an important consideration when evaluating studies of diagnostic accuracy. The distinction between bias and variation is not always straightforward, and the use of different definitions in the literature further complicates this issue. For example, when a diagnostic study starts by including patients who have already received a diagnosis of the target condition and uses a group of healthy volunteers as the control group, it is likely that both sensitivity and specificity will be higher than they would be in a study made up of patients only suspected of having the target condition. This feature has been described as spectrum bias. However, strictly speaking, one could argue that it is a form of variability; sensitivity and specificity have been measured correctly within the study and thus there is no bias; however, the results cannot be applied to the clinical setting. In other words, they lack generalizability (2). Others have argued that when the goal of a study is to measure the accuracy of a test in the clinical setting, an error in the method of patient selection is made that will lead to biased estimates of test performance. They use a broader definition of bias and take into account the underlying research question when deciding whether results are biased. In this paper, we use a more restricted definition of bias. Our goal is to classify the various sources of variation and bias, describe their effects on test results, and provide a summary of the available evidence that supports each source of bias and variation (Table 1). For this purpose, we conducted a systematic review of all studies in which the main focus was examine the effects of one or more sources of bias or variation on estimates of test performance. Table 1. Description of Sources of Bias and Variation Methods Literature Searches We searched MEDLINE, EMBASE, BIOSIS and the methodologic databases of the Centre for Reviews and Dissemination and the Cochrane Collaboration from database inception to 2001. Search terms included sensitivit*, mass-screening, diagnostic-test, laboratory-diagnosis, false positive*, false negative*, specificit*, screening, accuracy, predictive value*, reference value*, likelihood ratio, sroc, and receiver operat* characteristic*. We also identified papers that had cited the key papers. Complete details of the search strategy are provided elsewhere (3). We contacted methodologic experts and groups conducting work in this field. Reference lists of retrieved articles were screened for additional studies. Inclusion Criteria All studies with the main objective of addressing bias or variation in the results of diagnostic accuracy studies were eligible for inclusion. Studies of any design, including reviews, and any topic area were eligible. Studies had to investigate the effects of bias or variation on measures of test performance, such as sensitivity, specificity, predictive values, likelihood ratios, and diagnostic odds ratios, and indicate how a particular feature may distort these measures. Inclusion was assessed by one reviewer and checked by a second reviewer; discrepancies were resolved through discussion. Data Extraction One reviewer extracted data and a second reviewer checked data on the following parameters: study design, objective, sources of bias or variation investigated, and the results for each source. Discrepancies were resolved by consensus or consultation with a third reviewer. Data Synthesis We divided the different sources of bias and variation into groups (Table 1). Table 1 provides a brief description of each source of bias and variation; more detailed descriptions are available elsewhere (3). Results were stratified according to the source of bias or variation. Studies were grouped according to study design. We classified studies that used actual data from one or more clinical studies to demonstrate the effect of a particular study feature as experimental studies, diagnostic accuracy studies, or systematic reviews. Experimental studies were defined as studies specifically designed to test a hypothesis about the effect of a certain feature, for example, rereading sets of radiographs while controlling (manipulating) the overall prevalence of abnormalities. Studies that used models to simulate how certain types of biases may affect estimates of diagnostic test performance were classified as modeling studies. These studies were considered to provide theoretical evidence of bias or variation. Role of the Funding Source The funding source was not involved in the design, conduct, or reporting of the study or in the decision to submit the manuscript for publication. Data Synthesis The literature searches identified a total of 8663 references. Of these, 569 studies were considered potentially relevant and were assessed for inclusion; 55, published from 1963 to 2000, met inclusion criteria. Nine studies were systematic reviews, 16 studies used an experimental design, 22 studies were diagnostic accuracy studies, and 8 studies used modeling to investigate the theoretical effects of bias or variation. Population Demographic Features Ten studies assessed the effects of demographic features on test performance (Table 2) (4, 5, 7, 9, 11, 14, 15, 20, 22, 24). Eight studies were diagnostic accuracy studies, and 2 were systematic reviews. All but one study (22) found an association between the features investigated and overall accuracy. The study that did not find an association investigated whether estimates of exercise testing performance differed between men and women; after correction for the effects of verification bias, no significant differences were found (22). Table 2. Population In general, the studies found associations between the demographic factors investigated and sensitivity; the reported effect on specificity was less strong. Four studies found that various factors, including sex, were associated with sensitivity but showed no association with specificity (4, 5, 11, 20). The index tests investigated in these studies were exercise testing (5, 11, 20) to diagnose heart disease and body mass index to test for obesity (4). Two additional studies of exercise testing also reported an association with sensitivity, but the effects on specificity differed. One found that factors that lead to increased sensitivity also lead to a decrease in specificity (14); the second reported higher sensitivity and specificity in men than in women (16). A study of the diagnostic accuracy of an alcohol screening questionnaire found that overall accuracy was increased in certain ethnic groups (24). Sex was the most commonly investigated variable. Three studies found no association between test performance and sex, 9 found significant effects on sensitivity, and 4 found significant effects on specificity. Other variables shown to have significant effects on test performance were age, race, and smoking status. Disease Severity Six studies looked at the effects of disease severity on test performance (Table 2) (5, 11, 14, 19, 23, 25). Three studies were diagnostic accuracy studies, 2 were reviews, and one used modeling to investigate the effects of differences in disease severity. The modeling study also included an example from a diagnostic accuracy study of tests for the diagnosis of ovarian cancer (25). Three studies investigated tests for heart disease (5, 11, 14), one examined ventilationperfusion lung scans for diagnosing pulmonary embolism (23), and one investigated 2 different laboratory tests (one for cancer and the other for bacterial infections) (19). All 6 studies found increased sensitivity with more severe disease; 5 found no effect on specificity (5, 11, 14, 19, 23), and one did not comment on the effects on specificity (25). Disease Prevalence Six studies looked at the effects of increased disease prevalence on test performance (Table 2) (8, 10, 13, 17, 21, 26). One study used an experimental design (8); the other studies were all diagnostic accuracy studies. The te


web science | 2000

Systematic review of water fluoridation

Marian S McDonagh; Penny Whiting; Paul Wilson; Alex J. Sutton; Ivor Gordon Chestnutt; Jan Cooper; Kate Misso; Matthew Bradley; Elizabeth Tulip Treasure; Jos Kleijnen

Abstract Objective: To review the safety and efficacy of fluoridation of drinking water. Design: Search of 25 electronic databases and world wide web. Relevant journals hand searched; further information requested from authors. Inclusion criteria were a predefined hierarchy of evidence and objectives. Study validity was assessed with checklists. Two reviewers independently screened sources, extracted data, and assessed validity. Main outcome measures: Decayed, missing, and filled primary/permanent teeth. Proportion of children without caries. Measure of effect was the difference in change in prevalence of caries from baseline to final examination in fluoridated compared with control areas. For potential adverse effects, all outcomes reported were used. Results: 214 studies were included. The quality of studies was low to moderate. Water fluoridation was associated with an increased proportion of children without caries and a reduction in the number of teeth affected by caries. The range (median) of mean differences in the proportion of children without caries was −5.0% to 64% (14.6%). The range (median) of mean change in decayed, missing, and filled primary/permanent teeth was 0.5 to 4.4 (2.25) teeth. A dose-dependent increase in dental fluorosis was found. At a fluoride level of 1 ppm an estimated 12.5% (95% confidence interval 7.0% to 21.5%) of exposed people would have fluorosis that they would find aesthetically concerning. Conclusions: The evidence of a beneficial reduction in caries should be considered together with the increased prevalence of dental fluorosis. There was no clear evidence of other potential adverse effects.


BMC Medical Research Methodology | 2005

How does study quality affect the results of a diagnostic meta-analysis?

Marie Westwood; Penny Whiting; Jos Kleijnen

BackgroundThe use of systematic literature review to inform evidence based practice in diagnostics is rapidly expanding. Although the primary diagnostic literature is extensive, studies are often of low methodological quality or poorly reported. There has been no rigorously evaluated, evidence based tool to assess the methodological quality of diagnostic studies.The primary objective of this study was to determine the extent to which variations in the quality of primary studies impact the results of a diagnostic meta-analysis and whether this differs with diagnostic test type. A secondary objective was to contribute to the evaluation of QUADAS, an evidence-based tool for the assessment of quality in diagnostic accuracy studies.MethodsThis study was conducted as part of large systematic review of tests used in the diagnosis and further investigation of urinary tract infection (UTI) in children. All studies included in this review were assessed using QUADAS, an evidence-based tool for the assessment of quality in systematic reviews of diagnostic accuracy studies. The impact of individual components of QUADAS on a summary measure of diagnostic accuracy was investigated using regression analysis. The review divided the diagnosis and further investigation of UTI into the following three clinical stages: diagnosis of UTI, localisation of infection, and further investigation of the UTI. Each stage used different types of diagnostic test, which were considered to involve different quality concerns.ResultsMany of the studies included in our review were poorly reported. The proportion of QUADAS items fulfilled was similar for studies in different sections of the review. However, as might be expected, the individual items fulfilled differed between the three clinical stages. Regression analysis found that different items showed a strong association with test performance for the different tests evaluated. These differences were observed both within and between the three clinical stages assessed by the review. The results of regression analyses were also affected by whether or not a weighting (by sample size) was applied. Our analysis was severely limited by the completeness of reporting and the differences between the index tests evaluated and the reference standards used to confirm diagnoses in the primary studies. Few tests were evaluated by sufficient studies to allow meaningful use of meta-analytic pooling and investigation of heterogeneity. This meant that further analysis to investigate heterogeneity could only be undertaken using a subset of studies, and that the findings are open to various interpretations.ConclusionFurther work is needed to investigate the influence of methodological quality on the results of diagnostic meta-analyses. Large data sets of well-reported primary studies are needed to address this question. Without significant improvements in the completeness of reporting of primary studies, progress in this area will be limited.


Quality & Safety in Health Care | 2002

Interventions for the treatment and management of chronic fatigue syndrome/myalgic encephalomyelitis

Anne-Marie Bagnall; Penny Whiting; Richardson R; Amanda Sowden

The research evidence on the effectiveness of interventions for the treatment and management of chronic fatigue syndrome/myalgic encephalomyelitis (CFS/ME) published in a recent issue of Effective Health Care is reviewed.


BMC Public Health | 2001

Association of Down's syndrome and water fluoride level: a systematic review of the evidence

Penny Whiting; Marian S McDonagh; Jos Kleijnen

BackgroundA review of the safety and efficacy of drinking water fluoridation was commissioned by the UK Department of Health to investigate whether the evidence supported a beneficial effect of water fluoridation and whether there was any evidence of adverse effects. Downs syndrome was one of the adverse effects reported. The aim of this review is to examine the evidence for an association between water fluoride level and Downs syndrome.MethodsA systematic review of research. Studies were identified through a comprehensive literature search, scanning citations and online requests for papers. Studies in all languages which investigated the incidence of Downs syndrome in areas with different levels of fluoride in their water supplies were included. Study inclusion and quality was assessed independently by 2 reviewers. A qualitative analysis was conducted.ResultsSix studies were included. All were ecological in design and scored poorly on the validity assessment. The estimates of the crude relative risk ranged from 0.84 to 3.0. Four studies showed no significant associations between the incidence of Downs syndrome and water fluoride level and two studies by the same author found a significant (p < 0.05) positive association (increased Downs syndrome incidence with increased water fluoride level). Only two of the studies controlled for confounding factors and only one of these presented summary outcome measures.ConclusionsThe evidence of an association between water fluoride level and Downs syndrome incidence is inconclusive.


JAMA | 2001

Interventions for the Treatment and Management of Chronic Fatigue Syndrome: A Systematic Review

Penny Whiting; Anne Marie Bagnall; Amanda Sowden; John E. Cornell; Cynthia D. Mulrow; Gilbert Ramirez


Archive | 2017

Interventions for the Treatment and Management of Chronic Fatigue Syndrome

Penny Whiting; Anne-Marie Bagnall; Amanda Sowden; John E. Cornell; Cynthia D. Mulrow


Archive | 2015

Cost-effectiveness acceptability curves and incremental cost-effectiveness planes for the base-case analyses

Marie Westwood; Bram Ramaekers; Penny Whiting; Florian Tomini; Manuela A. Joore; Nigel Armstrong; Steve Ryder; Johan Severens; Jos Kleijnen


Archive | 2015

Subgroup analyses (base case)

Marie Westwood; Thea van Asselt; Bram Ramaekers; Penny Whiting; Praveen Thokala; Manuela A. Joore; Nigel Armstrong; Janine Ross; Johan Severens; Jos Kleijnen

Collaboration


Dive into the Penny Whiting's collaboration.

Top Co-Authors

Avatar

Marie Westwood

University of Southampton

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Thea van Asselt

Maastricht University Medical Centre

View shared research outputs
Top Co-Authors

Avatar

Maiwenn Al

Erasmus University Rotterdam

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge