Published on 12.06.12 in Vol 14, No 3 (2012): May-Jun
Sensitivity and Predictive Value of 15 PubMed Search Strategies to Answer Clinical Questions Rated Against Full Systematic Reviews
Background: Clinicians perform searches in PubMed daily, but retrieving relevant studies is challenging due to the rapid expansion of medical knowledge. Little is known about the performance of search strategies when they are applied to answer specific clinical questions.
Objective: To compare the performance of 15 PubMed search strategies in retrieving relevant clinical trials on therapeutic interventions.
Methods: We used Cochrane systematic reviews to identify relevant trials for 30 clinical questions. Search terms were extracted from the abstract using a predefined procedure based on the population, interventions, comparison, outcomes (PICO) framework and combined into queries. We tested 15 search strategies that varied in their query (PIC or PICO), use of PubMed’s Clinical Queries therapeutic filters (broad or narrow), search limits, and PubMed links to related articles. We assessed sensitivity (recall) and positive predictive value (precision) of each strategy on the first 2 PubMed pages (40 articles) and on the complete search output.
Results: The performance of the search strategies varied widely according to the clinical question. Unfiltered searches and those using the broad filter of Clinical Queries produced large outputs and retrieved few relevant articles within the first 2 pages, resulting in a median sensitivity of only 10%–25%. In contrast, all searches using the narrow filter performed significantly better, with a median sensitivity of about 50% (all P < .001 compared with unfiltered queries) and positive predictive values of 20%–30% (P < .001 compared with unfiltered queries). This benefit was consistent for most clinical questions. Searches based on related articles retrieved about a third of the relevant studies.
Conclusions: The Clinical Queries narrow filter, along with well-formulated queries based on the PICO framework, provided the greatest aid in retrieving relevant clinical trials within the 2 first PubMed pages. These results can help clinicians apply effective strategies to answer their questions at the point of care.
J Med Internet Res 2012;14(3):e85)
- Evidence-based medicine;
- information retrieval;
- medical literature;
- search strategy;
- clinical queries;
- search filters;
- positive predictive value;
Searching the literature for evidence has become a central skill in clinical practice . Physicians’ information needs are considerable [ - ], and evidence-based decisions often require the identification and appraisal of current research findings [ ]. As Glasziou et al commented, “the use of search engines is now as essential as the stethoscope” [ ]. However, retrieving relevant information has also become increasingly challenging, given the rapid expansion of medical knowledge: PubMed now comprises more than 20 million citations [ ], and 2000–4000 more are added every day [ ], including 75 clinical trials and 11 systematic reviews [ ].
In the past decade, several solutions have been implemented to improve access to current research . Preappraised resources, such as evidence-based medicine journals, or Web-based summaries such as UpToDate [ ], have been adopted by many clinicians [ , ]. But these resources are limited by delayed processing [ , ], insufficient coverage [ , ], or cost [ , ]. Thus, PubMed remains the most popular search engine used to retrieve original studies [ - ], either alone or as a complement to preappraised resources [ , ].
However, searching PubMed is not as intuitive as searching other commonly used engines such as Google. While clinicians often perform short unstructured queries of 2–3 terms , these tend to produce large and diluted outputs. More efficient search strategies can be proposed ( ), using existing search tools and based on expert recommendations. First, the clinical question is translated into search terms that are combined into a query using Boolean operators (eg, OR, AND) [ , ]. The use of the population, interventions, comparison, outcomes (PICO) framework helps formulate more precise queries that combine search terms for these four factors [ , ]. The size of the output can be further reduced using limits [ ] or methodological filters, such as PubMed’s Clinical Queries [ ], designed to retrieve high-quality randomized controlled trials for therapeutic interventions [ ]. Finally, additional strategies, such as the use of the related articles link in PubMed, can identify studies based on a relevant article that was initially found ( [ ]).
Evidence regarding the performance of search strategies to answer clinical questions is scarce. A few small studies have assessed the impact of giving clinicians search tutorials on the retrieval of specific sets of articles identified by experts, but search strategies could not be compared, as clinicians were free to use their own informal strategies [, ]. Other studies have reported on the performances of search filters in retrieving high-quality clinical trials, but these filters were assessed independently of user queries [ ] or clinical questions [ ]. To our knowledge, there is little evidence to help clinicians organize their searches and combine existing search tools into effective strategies that are applicable at the point of care.
In this study, we compared the performance of 15 PubMed search strategies in retrieving relevant clinical trials, as identified by high-quality systematic reviews on specific clinical questions. We devised these 15 search strategies by choosing and combining the following search components that are easily applicable at the point of care: formulation of queries using the PICO framework, use of Clinical Queries therapeutic filters (broad or narrow), use of several search limits, and use of PubMed links to related articles. Our aim was to identify search components and tools that would most likely help clinicians answer questions on therapeutic interventions at the point of care.
Sample of Systematic Reviews for the Identification of Relevant Articles
For a specific clinical question, a well-conducted systematic review is considered the gold standard for the identification of all relevant articles [- ], since it implies a systematic and comprehensive search, as well as an appraisal of the articles’ scientific rigor. Moreover, recent systematic reviews address questions of clinical interest for which relative uncertainty exists and are thus a good proxy of the questions that may arise in real clinical practice.
We selected 30 clinical questions from a wide spectrum of topics of general interest, on the basis of 30 recent Cochrane systematic reviews (, ). We favored Cochrane reviews because of their well-established quality and methodological rigor [ - ], and focused on reviews that included randomized controlled trials, as these constitute the highest level of evidence for questions about therapeutic interventions. In February 2010, we searched the Cochrane Database of Systematic Reviews [ ] for all reviews on therapeutic interventions published since the beginning of 2010. To focus on the most recent reviews, or reviews that were recently updated, we restricted the search to citations with the following record status: new review, new search, or conclusion changed. Of the 287 potentially eligible reviews, we excluded those with fewer than 4 studies (n = 113), those that were less relevant for a general audience (n = 82), and those with composite questions that would result in complex queries for clinical practice (n = 62). All clinical trials included in the remaining 30 reviews that were also retrievable in PubMed defined the subset of relevant studies used in our assessment.
Extraction of Search Terms and Formulation of PICO Query
We selected search terms using a predefined procedure that we determined before applying the searches strategies. Following the PICO framework (), we categorized the exact wording of the objective and selection criteria of each review’s abstract into four sections for the population, interventions, comparison, and outcomes [ , ]. When we could find no information in the abstract for one of these categories, we retrieved the relevant keywords from the methods section (eg, type of intervention or type of outcome). Then for each PICO category, we extracted the smallest set of search terms that best expressed the clinical question. This predefined procedure was performed by 2 assessors (TA and AM), both trained in clinical epidemiology and evidence-based medicine. In a learning phase, together they extracted search terms from a pilot sample of 10 systematic reviews. Then they independently extracted PICO search terms from the 30 reviews of the study sample and agreed on the smallest set of terms that best expressed the clinical questions. They resolved discrepancies by consensus. Finally, all coauthors, including experienced clinicians, approved the retained search terms. For half of the reviews, PICO’s comparison component shared common terms with interventions (eg, composite vs single intervention), and in these cases no term was retained for comparison. Examples of search term extraction are shown in .
To obtain the final search query (), search terms were combined with the Boolean operator OR within each PICO category, so as to increase sensitivity. Then all categories were combined with AND to retrieve only the citations matching all PICO elements [ - ]. For example, for a review on the effects of oral mucolytic agents in adults with stable chronic bronchitis or chronic obstructive pulmonary disease (COPD) [ ], the final full PICO query was ((chronic bronchitis) OR COPD) AND mucolytics AND placebo AND exacerbations). Search terms consisting of groups of words (eg, chronic bronchitis) were put into parentheses, not in quotes. The detailed wording of all 30 queries can be found in .
Design of Search Strategies
The purpose and context of the search determine the choice of search strategy. Our objective was to identify search components and tools that could help clinicians build more effective strategies to answer questions at the point of care. Therefore, selecting sophisticated strategies used for performing systematic reviews , such as the Cochrane Highly Sensitive Search Strategy [ ], would have been inappropriate. Since there is very little evidence on strategies adapted to our purpose, we devised our 15 search strategies ( , ) based on the few expert recommendations available [ - , , , - ]. We favored search components and search tools that are applicable and easy to use in clinical practice.
We thus obtained strategies by varying the following parameters. First, the search query was either the full PICO query or a truncated PIC query that did not include terms of outcomes. Although the full PICO is usually recommended for clinicians to help them formulate more precise searches [- ], such strategies may also miss relevant articles [ ], as the outcomes are less often mentioned in the abstract or assigned Medical Subject Headings (MeSH) in PubMed. Moreover, formulating shorter PIC queries may be less burdensome in practice and may be more appropriate for clinicians who are interested in many potential outcomes. We entered search terms using PubMed’s automatic term mapping, which systematically searches each term in the MeSH thesaurus, as well as the “[all fields]” tag ( ). We did not test strategies using the restriction to MeSH terms, as few clinicians manage to perform the necessary steps for this restriction in practice [ ], and this proportion hardly increases after medical residents take a search tutorial [ ].
Second, we combined these queries with either the broad or narrow filter of the Clinical Queries for therapeutic interventions. We chose Clinical Queries over other filters meant to improve the retrieval of high-quality randomized controlled trials  because Clinical Queries were specifically designed to help clinicians answer their questions on therapeutic interventions and are implemented in PubMed [ ]. To use these filters, clinicians simply need to enter their search terms on PubMed’s Clinical Queries page [ ]. After taking a search tutorial, about 80% of clinicians adopted the use of Clinical Queries [ ]. Third, we repeated these strategies adding search limits [ ] that restricted the searches to citations in English and research on humans. PICO searches with human and English limits were repeated with a further restriction to a set of 119 core medical journals listed in the Abridged Index Medicus (AIM) [ ]. We did not further add the limit “clinical trials [pt],” as it would have been redundant with the Clinical Queries filters.
Finally, the last 3 searches tested the PubMed link to related articles, a search tool that identifies content similarity in the title, abstract, and index terms . To use this tool in practice, clinicians must identify a citation they consider to be potentially relevant in a first search output (ie, a citation that best corresponds to their PICO question) and click on its link to related articles. They can then scan a new search output with citations ranked by content similarity with the initial citation. To test this strategy, 2 assessors (TA and AM) screened the first page of search #11 (PICO, therapy narrow filter, limited to human studies in English), a search strategy that yielded short outputs ( ), and identified by consensus the 3 citations that were closest to the corresponding PICO question. They performed this selection being blinded to the citations included in the review, so as to correspond to the situation that clinicians encounter in practice. Search performance was assessed on each output of these 3 citations’ related articles. We performed this strategy based on 3 citations, instead of 1, to minimize the impact of the subjective component of the choice of initial citation.
Analysis of Search Performance
We applied all 15 strategies for each of the 30 clinical questions (450 searches in total). We restricted each search to the date when the corresponding review was assessed as up-to-date, so that the time frame of the search was the same as its corresponding review. For each search we collected the number of articles in the output, the number of gold standard articles retrieved, and their position in the output. Then we computed the sensitivity (also called recall: the proportion of relevant papers that were retrieved) of each search and its positive predictive value ([PPV], also called precision: the proportion of the output corresponding to relevant papers), as defined in[ , , ].
Because most physicians screen only 2 pages of the PubMed output, or at most 40 items [, ], our primary outcomes for search performance were sensitivity and PPV for this cut-off. However, we also analyzed overall sensitivity and PPV on the full search output and further explored how these properties varied according to the cumulative number of items screened in the output.
We summarized the performance of the search strategies over the 30 clinical questions using nonparametric statistics and box plots. Filtered versus unfiltered searches were compared using the Wilcoxon signed rank tests (paired). Finally, we explored whether retrieval performances differed according to the characteristics of the corresponding review. All analyses were performed using R 2.12.1 software .
Characteristics of Systematic Reviews and Corresponding PICO Queries
The 30 systematic reviews addressed a broad pattern of clinical topics and were produced by 15 different Cochrane groups, with a range of 1–4 reviews per group. Of these, 13 reviews (43%) had their record status labeled as new review, and 13 others (43%) as new search with no change to the conclusion and 4 (13%) with conclusion changed. They included a median of 15 studies (interquartile range [IQR] 7–23, range 5–49), of which 85% (IQR 66%–100%) were retrievable in PubMed (). This led to a median of 12 relevant clinical trials per review (IQR 7–18) that we considered to be the gold standard for each clinical question.
The predefined extraction procedure of search terms from the reviews’ abstracts led to a median of 2 terms (IQR 1–2) for population 1 term (IQR 1–2) for interventions, 1 term (IQR 0–1) for comparison, and 2.5 search terms (IQR 2–3) for outcomes. The required number of search terms was variable, as these terms were strictly tailored to the reviews’ questions. Indeed, these were sometimes more complex or consisted of grouping of words (eg, generalized onset tonic–clonic seizures). Overall, queries resulted in a median of 4 terms (IQR 3–6) for the PIC query and 7 terms (IQR 5–8) for the PICO query (for detailed wording, see).
Performance of Search Strategies
We observed important differences in the sizes of the output across search strategies (). Unfiltered PIC queries resulted in the largest outputs, with a median of 173 items, while the PICO queries halved the output size. The use of the Clinical Queries broad filter reduced the output by about 20%, and use of the narrow filter reduced it by about 80%. In contrast, searches based on related articles typically retrieved hundreds of articles.
The sensitivity and the positive predictive value (PPV) were also highly variable within each search strategy (). When the full outputs were screened for relevant studies, about 85% were detected by PIC queries and 69% by PICO queries ( A). Overall sensitivity remained comparable when we used Clinical Queries filters (with or without limits), although the Clinical Queries narrow filter was associated with slightly lower overall sensitivities ( A). In contrast, the use of the AIM limit systematically lowered sensitivity to about 15% ( A). The overall sensitivity of searches based on related articles was also extremely variable, with a median of about 60%.
When the screening of relevant articles was limited to the first 2 pages (ie, 40 articles), sensitivity dropped to 10% for unfiltered searches and PPV decreased to 2.5%–10% (B). In contrast, all search strategies that used the Clinical Queries narrow filter (S4, S5, S10, and S11 in ) showed significantly higher sensitivities, overall around 50% (P < .001 when compared with their corresponding unfiltered queries) and higher PPV values, between 20% and 30% (P < .001) within the first 2 pages of the output. When looking at the 30 questions individually, adding a Clinical Queries narrow filter to a query increased the sensitivity of the search for 21 questions (70%), kept it stable for 6 (20%), and decreased it for 3 (10%), whereas PPV increased in 27 questions (90%) ( B). Moreover, adding a Clinical Queries narrow filter to a PIC query reduced the risk of finding no relevant articles in the first 2 pages from 11 questions (37%) to 4 questions (13%). The broad Clinical Queries filter did not improve search performance; nor did searches using related articles and the additional use of limits. Overall, PICO queries had slightly higher performances than PIC queries, although the differences were not statistically significant.
The exploration of search performance according to the number of articles screened in the output showed that sensitivity rose gradually up to about 100 items in unfiltered searches (). In contrast, the sensitivity of searches using the Clinical Queries narrow filter rose much more steeply, peaking earlier at 50 to 60 articles screened. Finally, we found no significant association between search performance and characteristics of the reviews, namely the absolute number of relevant articles included, the number of PICO search terms required to summarize its question, or the presence of mortality as an outcome (data not shown).
The performance of the search strategies tested was highly variable according to the clinical question. None of the 15 strategies showed a consistently high sensitivity in retrieving relevant articles identified by systematic reviews. Even the best strategies had a sensitivity of about 50%, with a range from 0% to 100% across the 30 questions. Therefore, clinicians who perform a specific PubMed search cannot foresee to what extent their search will be successful.
Nevertheless, on average, some search strategies were more successful than others. Unfiltered searches based on PIC or PICO queries, as well as those using the broad Clinical Queries filter, produced large outputs and retrieved few relevant articles within the first 2 pages. In contrast, searches that used the Clinical Queries narrow filter had a significantly higher sensitivity (around 50%–55%) and a higher PPV (around 20%–30%). These improved performances were observed for most clinical questions. Additional use of limits had only a marginal effect, except for the limit to core medical journals, which actually reduced sensitivity. Finally, searches based on related articles retrieved about a third of relevant studies in the first 2 pages. However, their PPV remained low, despite the ranking of citations by content similarity, because they also retrieved hundreds of nonrelevant articles. This may be because related articles will also retrieve studies that are not randomized trials, in contrast with Clinical Queries therapeutic filters.
Clinicians who use PubMed at the point of care favor short queries of 2–3 terms, without any limits or filters, and generally screen only the first 2 pages of the output (ie, 20–40 items) [, ]. Our results clearly show that such strategies will miss many relevant articles. Although these strategies may reach high overall sensitivities, relevant citations will be scattered over many pages that will not be screened in real life. To increase the density of relevant articles in the output (ie, the PPV), we recommend the use of the PIC(O) framework. Such queries including 5–8 terms kept a satisfactory overall sensitivity of about 85% ( A). However, when only the first 2 pages of the output were scanned, which is what usually occurs in clinical practice [ ], this sensitivity dropped to about 10%, with an even lower PPV of 2.5% (corresponding to a median of 1 relevant article per screened output).
To improve search performance within readable outputs, an important finding is the usefulness of the Clinical Queries narrow therapy filter in identifying relevant studies in the first 2 pages of PubMed output. Clinical Queries filters were designed by Haynes et al  based on a hand search of 161 clinical journals that identified methodologically sound randomized controlled trials. When applied to the whole Medline database, the Clinical Queries broad filter retained 99% of hand-selected studies, while the narrow filter retained 93% [ ]. Hoogendam et al found similar results [ ]. Our study confirmed the high sensitivity of the Clinical Queries filters when applied to specific clinical questions. A surprising finding was the increased sensitivity of the narrow filter compared with the broad filter when we screened only a realistic portion of the search output; this was because the narrow filter concentrated relevant articles on the first pages of the output. In other words, sensitivity, assessed on 2 pages of output, was better as PPV improved.
Based on our results we recommend that clinicians at the point of care start their PubMed searches by formulating a PIC query combined with a narrow filter, and then adapt it according to the output size and relevancy of the first items screened. If this strategy retrieves too many irrelevant citations, search terms describing the outcome (full PICO query) and further search limits can be added. In the opposite case, filters can be deactivated and sensitivity can be gained by adding search terms or by using the related articles link from the first relevant studies found. Finally, a general recommendation would be to screen beyond the first page of the output, as even the most efficient strategies require at least 50 to 60 items to approach their maximum sensitivity ().
Study Limitations and Strengths
The main limitation of our study is that we assessed search strategies based solely on their retrieval performance. The searches were not performed by clinicians at the point of care, so we did not capture the iterative process of searching PubMed, based on trial and error [, ], and could not assess whether relevant articles would be identified in practice. Furthermore, we did not examine the impact of search strategies on medical decisions and patients’ outcomes. However, performing searches that retrieve relevant articles is a prerequisite for their further use in clinical practice. Moreover, by using predefined search procedures, we were able to isolate and compare the effect of several components of search strategies in relation to a specific clinical question. As we focused on retrieval, we considered each relevant article to be equally important, although this may not be the case. In particular, we did not examine selection bias attributable to suboptimal identification of relevant articles. Finally, our results apply only to therapeutic interventions, assessed by clinical trials, and cannot be generalized to questions on prevention or diagnosis [ ]. For the latter, similar studies could be conducted using alternative Clinical Queries filters designed to improve the retrieval of high-quality studies on diagnosis or prognosis [ ].
Among its strengths, our design allowed an unbiased comparison of strategies, by controlling important sources of variation such as clinicians’ searching skills or previous knowledge [, ]. This exploration was broader and more systematic than in previous studies. Moreover, the tested search strategies relied on search components that are easily applicable in practice, as they are directly implemented in PubMed. Finally, we used reputable systematic reviews to identify relevant studies on specific clinical topics.
Conclusion and Future Prospects
Countless PubMed searches are performed daily by clinicians, but the effectiveness of this activity is poorly understood . We have shown that a well-formulated PIC query used with the Clinical Queries narrow filter was most likely to retrieve relevant clinical trials within readable outputs. These results can help clinicians build more effective strategies to answer their questions at the point of care, and thus reduce the gap between evidence from clinical trials and its actual implementation in practice. Further research should focus on the performance and clinical usefulness of selected strategies when they are performed in real practice [ ]. Meanwhile, PubMed remains a perfectible tool. Areas of improvement include the development of content filters for specific clinical disciplines [ ], or the implementation of new search interfaces that help clinicians formulate effective queries [ ] and conduct parallel searches combining methodological and content filters at will.
All persons who contributed significantly to this project are coauthors.
There was no external funding
Study concept, design: Agoritsas, Merglen, Perneger. Acquisition of data: Agoritsas, Merglen. Statistical analysis: Agoritsas, Courvoisier, Combescure. Interpretation of data: Agoritsas, Merglen, Courvoisier, Combescure, Garin, Perrier, Perneger. Drafting of the article: Agoritsas. Critical revision of the article for important intellectual content: Merglen, Courvoisier, Combescure, Garin, Perrier, Perneger. Supervision: Perneger.
Agoritsas had full access to all the data in the study and takes responsibility for the integrity of the data and the accuracy of the data analysis.
Conflicts of Interest
Multimedia Appendix 1
Detailed description of the 30 Cochrane systematic reviews on therapeutic interventions, used to identify relevant studies on 30 clinical questions.PDF File (Adobe PDF File), 85KB
Multimedia Appendix 2
Two examples of search term extraction forms for one complex (review #1) and one simpler (review #7) clinical questions.PDF File (Adobe PDF File), 68KB
Multimedia Appendix 3
Detailed search terms for 30 clinical questions, extracted from the abstract of the corresponding Cochrane systematic review according to the PICO framework.PDF File (Adobe PDF File), 69KB
- Straus SE, Richardson WS, Glasziou P, Haynes RB. Evidence-Based Medicine: How to Practice and Teach EBM. 3rd edition. Edinburgh: Elsevier/Churchill Livingstone; 2005.
- González-González AI, Dawes M, Sánchez-Mateos J, Riesgo-Fuertes R, Escortell-Mayor E, Sanz-Cuesta T, et al. Information needs and information-seeking behavior of primary care physicians. Ann Fam Med 2007;5(4):345-352 [FREE Full text] [CrossRef] [Medline]
- Green ML, Ciampi MA, Ellis PJ. Residents' medical information needs in clinic: are they being met? Am J Med 2000 Aug 15;109(3):218-223. [Medline]
- Graber MA, Randles BD, Ely JW, Monnahan J. Answering clinical questions in the ED. Am J Emerg Med 2008 Feb;26(2):144-147. [CrossRef] [Medline]
- Straus SE, McAlister FA. Evidence-based medicine: a commentary on common criticisms. CMAJ 2000 Oct 3;163(7):837-841. [Medline]
- Glasziou P, Burls A, Gilbert R. Evidence based medicine and the medical curriculum. BMJ 2008;337:a1253. [Medline]
- US National Library of Medicine, National Institutes of Health. PubMed URL: http://www.ncbi.nlm.nih.gov/sites/entrez?otool=ichbfmglib [accessed 2011-11-30] [WebCite Cache]
- Bastian H, Glasziou P, Chalmers I. Seventy-five trials and eleven systematic reviews a day: how will we ever keep up? PLoS Med 2010 Sep;7(9):e1000326 [FREE Full text] [CrossRef] [Medline]
- Haynes RB. Of studies, syntheses, synopses, summaries, and systems: the “5S” evolution of information services for evidence-based healthcare decisions. Evid Based Med 2006 Dec;11(6):162-164. [CrossRef] [Medline]
- UpToDate, Inc. 2011. Helping Clinicians Provide the Best Patient Care URL: http://www.uptodate.com/index [accessed 2011-11-30] [WebCite Cache]
- Thiele RH, Poiro NC, Scalzo DC, Nemergut EC. Speed, accuracy, and confidence in Google, Ovid, PubMed, and UpToDate: results of a randomised trial. Postgrad Med J 2010 Aug;86(1018):459-465. [CrossRef] [Medline]
- Hoogendam A, Stalenhoef AF, Robbé PF, Overbeke AJ. Answers to questions posed during daily patient care are more likely to be answered by UpToDate than PubMed. J Med Internet Res 2008;10(4):e29 [FREE Full text] [CrossRef] [Medline]
- Banzi R, Cinquini M, Liberati A, Moschetti I, Pecoraro V, Tagliabue L, et al. Speed of updating online evidence based point of care summaries: prospective cohort analysis. BMJ 2011;343:d5856 [FREE Full text] [Medline]
- Haynes RB. ACP Journal Club: the best new evidence for patient care. ACP J Club 2008 May 20;148(3):2. [Medline]
- Patel MR, Schardt CM, Sanders LL, Keitz SA. Randomized trial for answers to clinical questions: evaluating a pre-appraised versus a MEDLINE search protocol. J Med Libr Assoc 2006 Oct;94(4):382-387. [Medline]
- Bonis PA, Pickens GT, Rind DM, Foster DA. Association of a clinical knowledge support system with improved patient safety, reduced complications and shorter length of stay among Medicare beneficiaries in acute care hospitals in the United States. Int J Med Inform 2008 Nov;77(11):745-753. [CrossRef] [Medline]
- McKibbon KA, Fridsma DB. Effectiveness of clinician-selected electronic information resources for answering primary care physicians' information needs. J Am Med Inform Assoc 2006;13(6):653-659 [FREE Full text] [CrossRef] [Medline]
- Schilling LM, Steiner JF, Lundahl K, Anderson RJ. Residents' patient-specific clinical questions: opportunities for evidence-based learning. Acad Med 2005 Jan;80(1):51-56. [Medline]
- Shariff SZ, Bejaimal SA, Sontrop JM, Iansavichus AV, Weir MA, Haynes RB, et al. Searching for medical information online: a survey of Canadian nephrologists. J Nephrol 2011;24(6):723-732. [CrossRef] [Medline]
- Herskovic JR, Tanaka LY, Hersh W, Bernstam EV. A day in the life of PubMed: analysis of a typical day's query log. J Am Med Inform Assoc 2007;14(2):212-220 [FREE Full text] [CrossRef] [Medline]
- Hoogendam A, Stalenhoef AF, Robbé PF, Overbeke AJ. Analysis of queries sent to PubMed at the point of care: observation of search behaviour in a medical teaching hospital. BMC Med Inform Decis Mak 2008;8:42 [FREE Full text] [CrossRef] [Medline]
- Webster AC, Cross NB, Mitchell R, Craig JC. How to get the most from the medical literature: searching the medical literature effectively. Nephrology (Carlton) 2010 Feb;15(1):12-19. [CrossRef] [Medline]
- Doig GS, Simpson F. Efficient literature searching: a core skill for the practice of evidence-based medicine. Intensive Care Med 2003 Dec;29(12):2119-2127. [CrossRef] [Medline]
- Schardt C, Adams MB, Owens T, Keitz S, Fontelo P. Utilization of the PICO framework to improve searching PubMed for clinical questions. BMC Med Inform Decis Mak 2007;7:16 [FREE Full text] [CrossRef] [Medline]
- US National Library of Medicine, National Institutes of Health. PubMed Limits URL: http://www.ncbi.nlm.nih.gov/pubmed/limits [accessed 2011-11-30] [WebCite Cache]
- US National Library of Medicine, National Institutes of Health. PubMed Clinical Queries URL: http://www.ncbi.nlm.nih.gov/pubmed/clinical [accessed 2011-11-30] [WebCite Cache]
- Haynes RB, McKibbon KA, Wilczynski NL, Walter SD, Werre SR, Hedges Team. Optimal search strategies for retrieving scientifically strong studies of treatment from Medline: analytical survey. BMJ 2005 May 21;330(7501):1179. [CrossRef] [Medline]
- US National Library of Medicine, National Institutes of Health. 2011 Nov 8. Abridged Index Medicus (AIM or “Core Clinical”) Journal Titles URL: http://www.nlm.nih.gov/bsd/aim.html [accessed 2011-11-30] [WebCite Cache]
- Stark R, Helenius IM, Schimming LM, Takahara N, Kronish I, Korenstein D. Real-time EBM: from bed board to keyboard and back. J Gen Intern Med 2007 Dec;22(12):1656-1660. [CrossRef] [Medline]
- Gruppen LD, Rana GK, Arndt TS. A controlled comparison study of the efficacy of training medical students in evidence-based medicine literature searching skills. Acad Med 2005 Oct;80(10):940-944. [Medline]
- Hoogendam A, de Vries Robbé PF, Stalenhoef AF, Overbeke AJ. Evaluation of PubMed filters used for evidence-based searching: validation using relative recall. J Med Libr Assoc 2009 Jul;97(3):186-193. [CrossRef] [Medline]
- Kastner M, Wilczynski NL, McKibbon AK, Garg AX, Haynes RB. Diagnostic test systematic reviews: bibliographic search filters (“Clinical Queries”) for diagnostic accuracy studies perform well. J Clin Epidemiol 2009 Sep;62(9):974-981. [CrossRef] [Medline]
- Sampson M, Zhang L, Morrison A, Barrowman NJ, Clifford TJ, Platt RW, et al. An alternative to the hand searching gold standard: validating methodological search filters using relative recall. BMC Med Res Methodol 2006;6:33 [FREE Full text] [CrossRef] [Medline]
- Delaney A, Bagshaw SM, Ferland A, Laupland K, Manns B, Doig C. The quality of reports of critical care meta-analyses in the Cochrane Database of Systematic Reviews: an independent appraisal. Crit Care Med 2007 Feb;35(2):589-594. [CrossRef] [Medline]
- Jadad AR, Cook DJ, Jones A, Klassen TP, Tugwell P, Moher M, et al. Methodology and reports of systematic reviews and meta-analyses: a comparison of Cochrane reviews with articles published in paper-based journals. JAMA 1998 Jul 15;280(3):278-280. [Medline]
- Moher D, Tetzlaff J, Tricco AC, Sampson M, Altman DG. Epidemiology and reporting characteristics of systematic reviews. PLoS Med 2007 Mar 27;4(3):e78 [FREE Full text] [CrossRef] [Medline]
- The Cochrane Collaboration. John Wiley & Sons, Inc. 2012. The Cochrane Library: Advanced Search URL: http://onlinelibrary.wiley.com/o/cochrane/mainSearch?action=displayForm [accessed 2012-05-21] [WebCite Cache]
- Poole P, Black PN. Mucolytic agents for chronic bronchitis or chronic obstructive pulmonary disease. Cochrane Database Syst Rev 2010(2):CD001287. [CrossRef] [Medline]
- McKibbon KA, Wilczynski NL, Haynes RB, Hedges Team. Retrieving randomized controlled trials from medline: a comparison of 38 published search filters. Health Info Libr J 2009 Sep;26(3):187-202. [CrossRef] [Medline]
- Higgins JPT, Green S,E. The Cochrane Collaboration. 2011 Mar. Cochrane Handbook for Systematic Reviews of Interventions, Version 5.1.0, Section 6.4.2, Structure of a Search Strategy URL: http://www.cochrane-handbook.org/ [accessed 2012-05-16] [WebCite Cache]
- Lin J, Wilbur WJ. PubMed related articles: a probabilistic topic-based model for content similarity. BMC Bioinformatics 2007;8:423 [FREE Full text] [CrossRef] [Medline]
- R Development Core Team. R Foundation for Statistical Computing. 2008. The R Project for Statistical Computing URL: http://www.r-project.org/ [accessed 2011-11-30] [WebCite Cache]
- Hersh WR, Crabtree MK, Hickam DH, Sacherek L, Friedman CP, Tidmarsh P, et al. Factors associated with success in searching MEDLINE and applying evidence to answer clinical questions. J Am Med Inform Assoc 2002;9(3):283-293 [FREE Full text] [Medline]
- Vakkari P, Huuskonen S. Search effort degrades search output but improves task outcome. J Am Soc Inf Sci Technol 2012;63(4):657-670. [CrossRef]
- Shariff SZ, Cuerden MS, Haynes RB, McKibbon KA, Wilczynski NL, Iansavichus AV, et al. Evaluating the impact of MEDLINE filters on evidence retrieval: study protocol. Implement Sci 2010;5:58 [FREE Full text] [CrossRef] [Medline]
- Garg AX, Iansavichus AV, Wilczynski NL, Kastner M, Baier LA, Shariff SZ, et al. Filtering Medline for a clinical discipline: diagnostic test assessment framework. BMJ 2009;339:b3435 [FREE Full text] [Medline]
|AIM: Abridged Index Medicus|
|COPD: chronic obstructive pulmonary disease|
|IQR: interquartile range|
|MeSH: Medical Subject Headings|
|PIC(O): population, interventions, comparison, (outcomes)|
|PPV: positive predictive value|
Edited by G Eysenbach; submitted 04.12.11; peer-reviewed by M Sampson, S Shariff, K Robinson; comments to author 05.02.12; revised version received 11.03.12; accepted 13.04.12; published 12.06.12
©Thomas Agoritsas, Arnaud Merglen, Delphine S Courvoisier, Christophe Combescure, Nicolas Garin, Arnaud Perrier, Thomas V Perneger. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 12.06.2012.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.