6
STATISTICS IN MEDICINE Statist. Med. 2002; 21:1635–1640 (DOI: 10.1002/sim.1190) Identication of randomized controlled trials in systematic reviews: accuracy and reliability of screening records Phil Edwards 1; ; , Mike Clarke 2 , Carolyn DiGuiseppi 3 , Sarah Pratap 4 , Ian Roberts 1 and Reinhard Wentz 4 1 Public Health Intervention Research Unit; Department of Epidemiology and Population Health; London School of Hygiene & Tropical Medicine; 49–51 Bedford Square; London WC1B 3DP; U.K. 2 The UK Cochrane Centre; NHS R&D Programme; Summertown Pavilion; Middle Way; Oxford OX2 7LG; U.K. 3 Department of Preventive Medicine & Biometrics; University of Colorado Health Sciences Center; Campus Box C245; 4200 East Ninth Avenue; Denver; CO 80262; U.S.A. 4 Cochrane Injuries Group; Public Health Intervention Research Unit; Department of Epidemiology and Population Health; London School of Hygiene & Tropical Medicine; 49–51 Bedford Square; London WC1B 3DP; U.K. SUMMARY A study was conducted to estimate the accuracy and reliability of reviewers when screening records for relevant trials for a systematic review. A sensitive search of ten electronic bibliographic databases yielded 22 571 records of potentially relevant trials. Records were allocated to four reviewers such that two reviewers examined each record and so that identication of trials by each reviewer could be com- pared with those identied by each of the other reviewers. Agreement between reviewers was assessed using Cohen’s kappa statistic. Ascertainment intersection methods were used to estimate the likely num- ber of trials missed by reviewers. Full copies of reports were obtained and assessed independently by two researchers for eligibility for the review. Eligible reports formed the ‘gold standard’ against which an assessment was made about the accuracy of screening by reviewers. After screening, 301 of 22 571 records were identied by at least one reviewer as potentially relevant. Agreement was ‘almost perfect’ (¿0:8) within two pairs, ‘substantial’ (¿0:6) within three pairs and ‘moderate’ (¿0:4) within one pair. Of the 301 records selected, 273 complete reports were available. When pairs of reviewers agreed on the potential relevance of records, 81 per cent were eligible (range 69 to 91 per cent). If reviewers disagreed, 22 per cent were eligible (range 12 to 45 per cent). Single reviewers missed on average 8 per cent of eligible reports (range 0 to 24 per cent), whereas pairs of reviewers did not miss any (range 0 to 1 per cent). The use of two reviewers to screen records increased the number of randomized trials identied by an average of 9 per cent (range 0 to 32 per cent). Reviewers can reliably identify potentially relevant records when screening thousands of records for eligibility. Two reviewers should screen records for eligibility, whenever possible, in order to maximize ascertainment of relevant trials. Copyright ? 2002 John Wiley & Sons, Ltd. KEY WORDS: systematic reviews; screening; inter-observer reliability; ascertainment intersection Correspondence to: Phil Edwards, Public Health Intervention Research Unit, Department of Epidemiology and Popu- lation Health, London School of Hygiene & Tropical Medicine, 49–51 Bedford Square, London WC1B 3DP, U.K. E-mail: [email protected] Contract=grant sponsor: The BUPA Foundation Copyright ? 2002 John Wiley & Sons, Ltd.

Identification of randomized controlled trials in systematic reviews: accuracy and reliability of screening records

Embed Size (px)

Citation preview

Page 1: Identification of randomized controlled trials in systematic reviews: accuracy and reliability of screening records

STATISTICS IN MEDICINEStatist. Med. 2002; 21:1635–1640 (DOI: 10.1002/sim.1190)

Identi�cation of randomized controlled trials in systematicreviews: accuracy and reliability of screening records

Phil Edwards1;∗;†, Mike Clarke2, Carolyn DiGuiseppi3, Sarah Pratap4,Ian Roberts1 and Reinhard Wentz4

1Public Health Intervention Research Unit; Department of Epidemiology and Population Health;London School of Hygiene & Tropical Medicine; 49–51 Bedford Square; London WC1B 3DP; U.K.

2The UK Cochrane Centre; NHS R&D Programme; Summertown Pavilion; Middle Way;Oxford OX2 7LG; U.K.

3Department of Preventive Medicine & Biometrics; University of Colorado Health Sciences Center;Campus Box C245; 4200 East Ninth Avenue; Denver; CO 80262; U.S.A.

4Cochrane Injuries Group; Public Health Intervention Research Unit; Department of Epidemiology andPopulation Health; London School of Hygiene & Tropical Medicine; 49–51 Bedford Square;

London WC1B 3DP; U.K.

SUMMARY

A study was conducted to estimate the accuracy and reliability of reviewers when screening recordsfor relevant trials for a systematic review. A sensitive search of ten electronic bibliographic databasesyielded 22571 records of potentially relevant trials. Records were allocated to four reviewers such thattwo reviewers examined each record and so that identi�cation of trials by each reviewer could be com-pared with those identi�ed by each of the other reviewers. Agreement between reviewers was assessedusing Cohen’s kappa statistic. Ascertainment intersection methods were used to estimate the likely num-ber of trials missed by reviewers. Full copies of reports were obtained and assessed independently bytwo researchers for eligibility for the review. Eligible reports formed the ‘gold standard’ against whichan assessment was made about the accuracy of screening by reviewers. After screening, 301 of 22 571records were identi�ed by at least one reviewer as potentially relevant. Agreement was ‘almost perfect’(�¿0:8) within two pairs, ‘substantial’ (�¿0:6) within three pairs and ‘moderate’ (�¿0:4) within onepair. Of the 301 records selected, 273 complete reports were available. When pairs of reviewers agreedon the potential relevance of records, 81 per cent were eligible (range 69 to 91 per cent). If reviewersdisagreed, 22 per cent were eligible (range 12 to 45 per cent). Single reviewers missed on average8 per cent of eligible reports (range 0 to 24 per cent), whereas pairs of reviewers did not miss any(range 0 to 1 per cent). The use of two reviewers to screen records increased the number of randomizedtrials identi�ed by an average of 9 per cent (range 0 to 32 per cent). Reviewers can reliably identifypotentially relevant records when screening thousands of records for eligibility. Two reviewers shouldscreen records for eligibility, whenever possible, in order to maximize ascertainment of relevant trials.Copyright ? 2002 John Wiley & Sons, Ltd.

KEY WORDS: systematic reviews; screening; inter-observer reliability; ascertainment intersection

∗Correspondence to: Phil Edwards, Public Health Intervention Research Unit, Department of Epidemiology and Popu-lation Health, London School of Hygiene & Tropical Medicine, 49–51 Bedford Square, London WC1B 3DP, U.K.

†E-mail: [email protected]

Contract=grant sponsor: The BUPA Foundation

Copyright ? 2002 John Wiley & Sons, Ltd.

Page 2: Identification of randomized controlled trials in systematic reviews: accuracy and reliability of screening records

1636 P. EDWARDS ET AL.

1. INTRODUCTION

One of the most important steps in any systematic review of randomized controlled trialsis to identify and include all, or nearly all, of the relevant trials. Inclusion of all relevanttrials reduces random error in meta-analyses and, because ease of identi�cation of trials isassociated with treatment e�ects, complete ascertainment may also reduce bias [1]. To identifyas many of the relevant trials as possible, reviewers will typically conduct sensitive searches ofrelevant electronic bibliographic databases. These searches might yield thousands of records ofpotentially relevant reports that must then be screened to determine whether they are eligiblefor the review. Records that are excluded during screening are not usually considered again.Although screening records for eligibility may impact importantly on both precision and biasin systematic reviews, there is little information on the accuracy and reliability of the screeningprocess.Previous studies [2; 3] have examined agreement between reviewers but have not estimated

the number of relevant trials missed by reviewers, or the increase in the number of trialsidenti�ed using a second reviewer. We conducted a study to estimate the accuracy and relia-bility of screening records for relevant trials for a systematic review, using records retrievedby a search of ten databases.

2. METHODS

A study was designed to assess the accuracy and reliability of reviewers when screeningrecords. The records used were those identi�ed by the search strategy for a systematic reviewof methods to in�uence response to postal questionnaires [4]. This systematic review sought toidentify all reports of randomized controlled trials of strategies to in�uence response to postalquestionnaires, with no restriction by type of strategy, questionnaire topic, target populationor language. A search of ten electronic bibliographic databases yielded 26937 records ofpotentially relevant reports. Records including titles and abstracts were downloaded into aProCite reference database. After removing duplicate records, there were 22571 records ofpotentially relevant reports. Records were sorted and divided into six approximately equal sets(A to F).Four reviewers took part in the study, each being a co-reviewer on the systematic review [4].

Each reviewer had substantial experience of screening records for systematic reviews, apartfrom reviewer 2 who was relatively inexperienced. Each was allocated three of the sets toscreen (that is, each reviewer was to screen approximately 11286 records). The six sets wereallocated such that two reviewers examined each record, and so that identi�cation of trialsby each reviewer could be compared with each of the other reviewers (reviewer 1:A;B;C;reviewer 2:A;D;E; reviewer 3: B;D;F; reviewer 4: C;E;F). Prior to screening records, thereviewers met to discuss the criteria for trials to be eligible for inclusion in the systematicreview. Reviewers then worked independently, screening the title and abstract of each recordin their sets, selecting those they considered being potentially relevant.Agreement between reviewers on the relevance of records was assessed using Cohen’s

kappa statistic (�), which adjusts the proportion of records for which there was agreement bythe amount of agreement expected by chance alone [5]. We calculated 95 per cent con�denceintervals for kappa as ±1:96 times the standard error of kappa [6].

Copyright ? 2002 John Wiley & Sons, Ltd. Statist. Med. 2002; 21:1635–1640

Page 3: Identification of randomized controlled trials in systematic reviews: accuracy and reliability of screening records

IDENTIFICATION OF RANDOMIZED CONTROLLED TRIALS IN SYSTEMATIC REVIEWS 1637

Ascertainment intersection methods, analogous to capture–recapture methods used in ecol-ogy, were used to estimate the likely number of relevant records missed by pairs of reviewers[7; 8]. For example, the maximum likelihood estimate for the number of relevant recordsmissed by reviewer 1 and reviewer 2 during screening is

bca+ 1

where b is the number of relevant records identi�ed only by reviewer 1, c is the numberidenti�ed only by reviewer 2, and a is the number identi�ed by both reviewers. This methodproduces an unbiased estimate of relevant records missed by both reviewers [7]. Con�denceintervals for the estimate were calculated using the goodness-of-�t model [9].When screening was complete, full copies of the reports identi�ed by at least one reviewer

as potentially relevant were requested. Two researchers independently assessed each report foreligibility for inclusion in the systematic review. Disagreements about eligibility were referredto a third researcher. The eligible reports were used as the ‘gold standard’ against which anassessment was made about the accuracy of screening by reviewers.

3. RESULTS

After screening, 301 of 22571 records had been selected by at least one of the four review-ers as being potentially relevant. Reviewer 1 selected 102 records, reviewer 2 selected 74records, reviewer 3 selected 95 records and reviewer 4 selected 163 records. Of the six pos-sible comparisons between reviewers, kappa coe�cients of agreement ranged from 0.59 (95per cent CI 0.56 to 0.62) to 0.93 (95 per cent CI 0.90 to 0.96) (Table I). Agreement was‘almost perfect’ (�¿0:8) within two pairs, ‘substantial’ (�¿0:6) within three pairs, and ‘mod-erate’ (�¿0:4) within one pair [10]. Ascertainment intersection methods estimate that singlereviewers missed on average 22 per cent (range 3 to 55 per cent) of potentially relevantrecords. Pairs of reviewers missed 4 per cent (range 0 to 6 per cent) of potentially relevantrecords.Of the 301 records selected, we were able to obtain 273 complete reports for our assessment

of accuracy (the remainder being unobtainable despite extensive attempts through the BritishLibrary and other sources) (Table II). Of these, 156 (57 per cent) met the inclusion criteria

Table I. Kappa coe�cients of agreement (95 per cent con�dence interval) in selectionof potentially relevant records by four reviewers.

Reviewer 2 Reviewer 3 Reviewer 4

Reviewer 1 0.81 0.93 0.59(0:78; 0:84) (0:90; 0:96) (0:56; 0:62)

Reviewer 2 0.74 0.64(0:70; 0:77) (0:61; 0:67)

Reviewer 3 0.74(0:71; 0:77)

Copyright ? 2002 John Wiley & Sons, Ltd. Statist. Med. 2002; 21:1635–1640

Page 4: Identification of randomized controlled trials in systematic reviews: accuracy and reliability of screening records

1638 P. EDWARDS ET AL.

Table II. Numbers of records selected by four reviewers as potentially relevant (numberssubsequently found to be eligible). Ascertainment intersection estimates∗ of reports missed

by reviewers are shown in bold.

Reviewer 2 Reviewer 3 Reviewer 4

Reviewer 1 Yes No Yes No Yes NoYes 22 (20) 5 (2) 34 (28) 1 (1) 35 (26) 5 (2)No 3 (0) 1 (0) 4 (1) 0 (0) 40 (3) 6 (0)

Reviewer 2Yes 19 (17) 4 (1) 25 (22) 1 (0)No 8 (2) 2 (0) 24 (7) 1 (0)

Reviewer 3Yes 26 (18) 4 (1)No 13 (5) 2 (0)

Figures exclude 28 identi�ed reports that were not available.∗Ascertainment intersection estimates have been rounded to integers.

for the systematic review. Reviewer 1 identi�ed 79 eligible reports, reviewer 2 identi�ed 60eligible reports, reviewer 3 identi�ed 67 eligible reports and reviewer 4 identi�ed 81 eligiblereports. When pairs of reviewers agreed about the potential relevance of records, the majority(81 per cent) was found to be eligible (131 of 161 records). This proportion ranged acrossthe six pairings of reviewers from 69 per cent eligible (18 of 26 records) to 91 per centeligible (20 of 22 records). If two reviewers disagreed on the potential relevance of records,22 per cent were still found to be eligible (25 of 112 records). This proportion ranged from12 per cent eligible (1 of 8 records) to 45 per cent eligible (5 of 11 records). Ascertainmentintersection methods estimate that single reviewers missed on average 8 per cent (range 0to 24 per cent) of eligible reports and that pairs of reviewers did not miss any (range 0to 1 per cent). The average increase in the total number of randomized trials identi�ed byusing a second reviewer was 9 per cent, ranging across pairs from 0 per cent (both reviewersindividually identi�ed 29 trials) to 32 per cent (a second reviewer identi�ed an additionalseven trials increasing the total identi�ed to 29 trials).

4. DISCUSSION

Screening many thousands of records might be expected to be unreliable. This study hasshown, however, that pairs of reviewers can reliably identify potentially relevant records andthat together they can ensure that few eligible records are excluded. Our results on the inter-observer reliability of screening records were consistent with those reported in previous studies[2; 3]. There are, however, a number of methodological issues that may have a bearing onthe validity of these, and our other results.First of all, the eligibility of selected records was determined by two researchers. Although

this was done using the full reports, some studies may nevertheless have been misclassi�ed.Any misclassi�cation may have led us to underestimate the number of studies missed byreviewers. Secondly, for the ascertainment intersection estimates to be accurate, the sources

Copyright ? 2002 John Wiley & Sons, Ltd. Statist. Med. 2002; 21:1635–1640

Page 5: Identification of randomized controlled trials in systematic reviews: accuracy and reliability of screening records

IDENTIFICATION OF RANDOMIZED CONTROLLED TRIALS IN SYSTEMATIC REVIEWS 1639

used (reviewers screening records) must be independent. In this study, independence wasensured by having reviewers screen records in di�erent locations, remaining blind to theresults of other reviewers until all screening was complete.Another requirement for the ascertainment intersection estimates to be accurate is for the

probability with which relevant records are selected to be constant for any reviewer. Somevariation in the ease with which relevant records can be identi�ed, and in the skills of review-ers in identifying potentially relevant records, is likely. In this study, prior to screening, thecriteria for records to be eligible were discussed, in an attempt to achieve consistency betweenreviewers. However, it is possible that their competence in applying the criteria improved withpractice. In addition, decreased attentiveness during long periods of screening may also havein�uenced the probability of selection of relevant records. This was not supported by theresults, however, where reviewer 2 took over seven days to screen records and reviewer 1screened in less than two days, and yet agreement between them was ‘almost perfect’. Therewas considerable variation in the kappa coe�cients none the less. It might be thought thatgreater experience in screening would lead to greater agreement, but the kappa coe�cients ofagreement for the reviewer with the least experience (reviewer 2) were higher, on average,than those for reviewer 4. These results provide some evidence of variation in the probabilityof ascertainment of relevant records by reviewers. Although any resulting bias is likely to berelatively small [11], estimates of the numbers of reports missed by pairs of reviewers mustbe treated with a degree of caution.The limitations of the kappa coe�cient of agreement have been discussed elsewhere [12].

Of particular relevance is the fact that the kappa statistic can vary considerably depending onthe proportion of records that is eligible. In this study approximately 1 per cent of recordswas selected as potentially relevant. Similar levels of agreement between reviewers in otherstudies may not produce similar values of kappa if a greater proportion of records is eligible.For this study we used the records identi�ed by a search for randomized controlled trials

of methods to in�uence response to postal questionnaires. The records came from searches often bibliographic databases, which included the major general medical databases [4]. Althoughthe reports identi�ed by the four reviewers were from journals, books, conference proceedingsand other sources, covering a wide range of academic disciplines, the extent to which theseresults can be generalized (for example, to searches for randomized controlled trials of medicalinterventions only), is a matter for judgement.Screening records for eligibility may impact importantly on both precision and bias in

systematic reviews. Although a single reviewer is likely to identify the majority of relevanttrials, we found that a second reviewer can increase the number identi�ed by as much as one-third. Two reviewers should therefore screen records for eligibility whenever possible in orderto maximize ascertainment of relevant trials. Further research could examine how aspects ofelectronic records in�uence their ease of identi�cation for systematic reviews.

REFERENCES

1. Clarke MJ, Stewart LA. Obtaining data from randomised controlled trials: how much do we need for reliableand informative meta-analyses? British Medical Journal 1994; 309:1007–1010.

2. Castro AA, Clark OAC, Atallah AN. Improving retrieval of clinical trials in the LILACS database. Sixth AnnualCochrane Colloquium Abstracts, October 1998, Baltimore.

3. Strouse D, Hylton K, Sweat M, Elmquist L, Kay L, Sogolow E, Semaan S. Identifying HIV prevention trials forsystematic review: comparing manual and automated methods of retrieval. Seventh Annual Cochrane ColloquiumAbstracts, October 1999, Rome.

Copyright ? 2002 John Wiley & Sons, Ltd. Statist. Med. 2002; 21:1635–1640

Page 6: Identification of randomized controlled trials in systematic reviews: accuracy and reliability of screening records

1640 P. EDWARDS ET AL.

4. Edwards P, Clarke M, Diguiseppi C, Roberts I, Wentz R. Methods for increasing response to postalquestionnaires (Protocol for a Cochrane Review). In The Cochrane Library, Issue 1. Update Software: Oxford,2000.

5. Cohen J. A coe�cient of agreement for nominal scales. Educational and Psychological Measurement 1960;20:37–46.

6. Fleiss JL. Statistical Methods for Rates and Proportions. 2nd edn. Wiley: New York, 1981; 220–221.7. Hook EB, Regal RR. The value of capture–recapture methods even for apparent exhaustive surveys. The needfor adjustment for source of ascertainment intersection in attempted complete prevalence studies. AmericanJournal of Epidemiology 1992; 135:1060–1067.

8. Spoor P, Airey M, Bennett C, Greensill J, Williams R. Use of the capture–recapture technique to evaluate thecompleteness of systematic literature searches. British Medical Journal 1996; 313:342–343.

9. Regal RR, Hook EB. Goodness-of-�t based con�dence intervals for estimates of the size of a closed population.Statistics in Medicine 1984; 3:288–291.

10. Landis JR, Koch GG. The measurement of observer agreement for categorical data. Biometrics 1977; 33:159–174.

11. Hook EB, Regal RR. E�ect of variation in probability of ascertainment by sources (‘Variable Catchability’) upon‘Capture–Recapture’ estimates of prevalence. American Journal of Epidemiology 1993; 137(10):1148–1166.

12. Maclure M, Willett WC. Misinterpretation and misuse of the kappa statistic. American Journal of Epidemiology1987; 126(2):161–169.

Copyright ? 2002 John Wiley & Sons, Ltd. Statist. Med. 2002; 21:1635–1640