Skip to main content

Preferred Reporting Items for Overviews of Reviews (PRIOR): a protocol for development of a reporting guideline for overviews of reviews of healthcare interventions

Abstract

Background

Overviews of reviews (i.e., overviews) compile information from multiple systematic reviews to provide a single synthesis of relevant evidence for healthcare decision-making. Despite their increasing popularity, there are currently no systematically developed reporting guidelines for overviews. This is problematic because the reporting of published overviews varies considerably and is often substandard. Our objective is to use explicit, systematic, and transparent methods to develop an evidence-based and agreement-based reporting guideline for overviews of reviews of healthcare interventions (PRIOR, Preferred Reporting Items for Overviews of Reviews).

Methods

We will develop the PRIOR reporting guideline in four stages, using established methods for developing reporting guidelines in health research. First, we will establish an international and multidisciplinary expert advisory board that will oversee the conduct of the project and provide methodological support. Second, we will use the results of comprehensive literature reviews to develop a list of prospective checklist items for the reporting guideline. Third, we will use a modified Delphi exercise to achieve a high level of expert agreement on the list of items to be included in the PRIOR reporting guideline. We will identify and recruit a group of up to 100 international experts who will provide input into the guideline in three Delphi rounds: the first two rounds will occur via online survey, and the third round will occur during a smaller (8 to 10 participants) in-person meeting that will use a nominal group technique. Fourth, we will produce and publish the PRIOR reporting guideline.

Discussion

A systematically developed reporting guideline for overviews could help to improve the accuracy, completeness, and transparency of overviews. This, in turn, could help maximize the value and impact of overviews by allowing more efficient interpretation and use of their research findings.

Peer Review reports

Background

Overviews of reviews specific to healthcare interventions (sometimes referred to as “overviews of systematic reviews,” “reviews of reviews,” “reviews of systematic reviews,” or “umbrella reviews”; hereafter referred to as “overviews”) use explicit and systematic methods to search for, identify, extract data from, and analyze the results of multiple related systematic reviews (Table 1). Their aim is to provide a single synthesis of systematic review evidence to answer different types of questions related to the efficacy, effectiveness, and/or safety of healthcare interventions for preventing or treating various clinical conditions [1]. Because overviews have been gaining momentum as an increasingly popular knowledge synthesis product, methods for conducting overviews have evolved in recent years [2,3,4,5,6]. For example, in 2016, we published a scoping review summarizing existing guidance for conducting overviews of healthcare interventions [7], and we used the results to update the chapter on overview methods in the Cochrane Handbook for Systematic Reviews of Interventions [1]. Despite advances in methods for conducting overviews, the reporting of overviews varies considerably and is often substandard (Table 2) [2,3,4].

Table 1 Types of questions about healthcare interventions that overviews can examinea
Table 2 Percentage of overviews (published up to 2016 or 2017) reporting on key aspects of methods and results

A reporting guideline is defined as “a checklist, flow diagram, or explicit text to guide authors in reporting a specific type of research, developed using explicit methodology” [8]. To date, we are not aware of any published guidance that was developed based on the evidence-based and agreement-based process recommended by the Enhancing QUAlity and Transparency Of health Research (EQUATOR) Network [8]. We are, however, aware of several relevant documents that narratively describe (based on personal experience and principles of “good practice”) issues related to reporting overviews. Onishi and Furukawa [9] developed a checklist for overviews of reviews focused on clinical topics based on the recommendations of Cochrane [10], the Preferred Reporting Items for Systematic reviews and Meta-Analyses (PRISMA) statement [11], and the AMSTAR (A MeaSurement Tool to Assess systematic Reviews) quality assessment tool [12]. Bougioukas et al. [13] published a checklist (PRIO, Preferred Reporting Items for Overviews of Systematic Reviews) with specific emphasis on harms reporting that was designed by combining recommendations from PRISMA [11], PRISMA-harms [14], PRISMA-protocols (PRISMA-P) [15], and other relevant documents. The same research group published the first available guidance on the reporting of abstracts for overviews of reviews of healthcare interventions [16], generated by combining features of PRISMA for abstracts [17], the PRIO checklist [13], important features of published abstracts identified in their literature search, and expert input. Li et al. [18] produced a checklist for overviews of reviews of healthcare interventions based on the recommendations in the Cochrane handbook [10], PRISMA statement [11], and the Overview Quality Assessment Questionnaire (OQAQ) [19]. Finally, Singh [20] published a reporting checklist for metareviews (MARQ, Metareview Assessment of Reporting Quality checklist), a type of overview that descriptively compares the reporting characteristics of systematic reviews, based on AMSTAR [12] and Cochrane guidance [10].

An abstract for STROVI (Standards for Reporting of Overviews of Reviews and Umbrella Reviews statement) was presented at the 2017 Cochrane Colloquium [21], but our informal contact with the authors indicated that to date further work on the project has not been undertaken. We have included the primary author of the STROVI abstract, as well as authors of the other previously mentioned published checklists, as participants in our Delphi process, based on their interest and expertise in overview methodology. While the aforementioned guidance can be a good starting point for authors preparing overviews, there is a lack of clarity about which might be the best to use and under what circumstance. In addition, much of the evidence used to develop the guidelines has since been updated (e.g., a new Cochrane chapter on overviews was published [1], and AMSTAR 2 [22] was developed). There is a need for an up-to-date, rigorously developed [8] reporting guideline for overviews that can be endorsed by funders and publishers [8] to help authors report their methods and results in a consistent, complete, and transparent way [8, 23, 24]. This, in turn, could help end-users to better assess the reliability, validity, and applicability of overview results when making healthcare decisions [8, 25].

Our objective is to develop an evidence-based and agreement-based reporting guideline for overviews (PRIOR, Preferred Reporting Items for Overviews of Reviews) using explicit, systematic, and transparent methods, as recommended by the EQUATOR Network [8]. Because the unit of analysis differs from that of systematic reviews, we plan to develop the guideline de novo, rather than undertaking a PRISMA extension. Though we expect that some of the reporting guidance will be similar to PRISMA, the development of a stand-alone guideline will allow us to focus on the particular challenges in reporting overviews and to facilitate future guideline extensions (e.g., for diagnostic or prognostic accuracy overviews, etiology overviews). This guideline will consist of a “minimum essential set of items that should be reported” in overviews of healthcare interventions [8]. The PRIOR guideline will focus on overviews that examine the efficacy, effectiveness, and/or safety of healthcare interventions and that present narrative summaries and/or meta-analyses of quantitative outcome data. The target audience of the PRIOR reporting guideline will be overview authors, peer reviewers, journal editors, and healthcare decision-makers.

Methods

This study will follow the key steps recommended by the EQUATOR Network for developing reporting guidelines in health research [8], which have been used to successfully develop reporting guidelines for other similar knowledge syntheses such as systematic reviews [11] and network meta-analyses [26]. We will develop the reporting guideline in four stages: (1) project launch, (2) literature review, (3) modified Delphi exercise, and (4) development of the guidance statement. These stages are illustrated in Fig. 1 and described below. We will record protocol amendments and describe these in our final manuscript. We have registered our intent to develop the PRIOR reporting guideline for overviews on the EQUATOR Network website (“reporting guidelines under development” section) [27]. The EQUATOR Network encourages registration of all reporting guidelines and extensions under development to raise awareness and help to prevent duplication. Before launching the project, we obtained ethics approval from the University of Alberta Health Research Ethics Board (#Pro00086094).

Fig. 1
figure 1

Four stages of the reporting guideline

Project launch

At project launch (December 2018), we established a core project team who will be responsible for the day-to-day operations of the planned project, consisting of methodologists with expertise in overview methodology and the development of reporting guidelines (MP [Canada]; LH [Canada]; RMF [Portugal]; DP [Germany]; ACT [Canada]), and research staff (AG and MG [Canada]).

We then established an international and multidisciplinary expert advisory board, with expertise in overview methodology and the development of reporting guidelines, who will advise on the development of PRIOR: Dr. David Moher (Canada); Dr. Tianjing Li (USA); Dr. Sue Brennan (Australia). The advisory board will be consulted regularly throughout the guideline development process. They will be asked to recommend relevant documents for the literature review; nominate participants for the Delphi exercise, review the checklist items for inclusion in the first round of the Delphi exercise, provide feedback after each round of the Delphi exercise (e.g., interpret results of the previous round, approve content for the next round), help plan and co-facilitate the in-person meeting, contribute to the production of the final reporting guideline, and assist with dissemination and knowledge translation activities.

Literature review

To support the development of the PRIOR reporting guideline, we will conduct a methodological SR examining the quality of reporting of a sample of overviews of healthcare interventions published from 2012 to 2016 (in progress) [28]. To ensure that the most recently available data are used to inform the PRIOR guideline, we have supplemented our own data with those recently published by Lunny et al. in 2019 [4]. We will also search for and summarize methodological documents related to conducting and reporting overviews (e.g., documents that provide guidance for reporting overviews, documents that summarize methods used to conduct overviews, and empirical studies that evaluate methods for conducting overviews) (in progress). To identify relevant methodological documents, we will use and expand upon the most relevant search strategies contained within the scoping review by Pollock et al. [7], the methodological SR by Pieper et al. [28], and the evidence map by Lunny et al. [5, 6]. Our searches, which will continue until the final version of the PRIOR guideline is completed, will consist of: web searches (Google Scholar); reference tracking [29, 30]; monitoring article alerts; hand-searching websites, conference proceedings, and personal files; and asking experts (e.g., advisory board members; Delphi participants during the first survey round) for relevant articles. Two independent reviewers will assess titles and abstracts and all potentially relevant full-text articles for inclusion, with discrepancies resolved by consensus or third party adjudication. Results of the literature review will be used to develop and refine a list of checklist items that a panel of expert participants will assess for inclusion in the PRIOR reporting guideline.

Modified Delphi exercise

A panel of experts (i.e., “participants”) will participate in a modified Delphi exercise to achieve a high level of agreement (≥ 70%) on the list of items to be included in the PRIOR reporting guideline [8]. The participants will provide feedback on potential items during three Delphi rounds, using structured feedback between rounds to help reconcile individual opinions and achieve group agreement [31,32,33]. Three rounds are likely to result in convergence of opinions between participants [32, 33]. The first two Delphi rounds will include all participants and occur online via self-administered survey; the third Delphi round will include a smaller subset of the participants (i.e., “expert panelists”) and occur during an in-person meeting that will use a nominal group technique (i.e., a formal consensus technique where expert panelists systematically review, discuss, and re-rate outstanding items) to achieve agreement [33, 34]. Should a high level of agreement (≥ 70%) not be reached following the in-person meeting, we may implement a third online survey among the expert panelists to assist in achieving agreement. Before the study, we will pilot test the survey’s usability, clarity, and face validity by sending it to five individuals familiar with overview methods but uninvolved in the current project. Their feedback will be used to revise the survey format and checklist items as needed.

Participant recruitment

We will use a purposive sampling technique [31] to identify and recruit a panel of up to 100 participants with experience coordinating, conducting, reviewing, disseminating, and/or using overviews of healthcare interventions (e.g., editors, authors, peer reviewers, and end-users of published overviews such as guideline developers). We will aim to recruit international participants who have diverse roles (e.g., researchers, healthcare professionals, patients, journal editors, policy-makers, funding agency representatives) and are employed in a range of settings (e.g., universities, hospitals, government, non-profit organizations, for-profit organizations). Participants will form a single panel for the analysis of survey results, feedback between rounds, and criteria for agreement. A list of potential participants will be prioritized by the core project team, in order to provide representation from our populations of interest (i.e., editors, authors, peer-reviewers, end-users) and major evidence synthesis centers (i.e., Joanna Briggs Institute, Canadian Agency for Drugs and Technologies in Health, Campbell Collaboration, Cochrane, Agency for Healthcare Research and Quality Effective Health Care Program, Canadian Task Force on Preventive Health Care, Centre for Evidence-Based Medicine, Centre for Evidence-based Health Care). The core project team and expert advisory board will also provide input, based on their knowledge of the potential participants, on each participant’s level of expertise and potential for unique contribution to the development of the guideline (e.g., statistical expertise, information specialists). Participants will be invited via personalized email that will describe the PRIOR guideline development project and explain the objective, process, and timelines of the Delphi exercise. We will obtain informed consent from participants using an online consent form.

Round one: Online survey

The participants will be asked to use a self-administered online survey to rate, on a four-point Likert scale, the extent to which they agree with the inclusion of each checklist item in the PRIOR reporting guideline (1 = strongly disagree, 2 = somewhat disagree, 3 = somewhat agree, 4 = strongly agree) [34]. Participants may also choose to answer “I don’t know” and provide an explanation [35]. For each item, a free text box will be provided for general comments (e.g., justification for their decision, proposed wording changes). Items will be presented in an order that reflects the progression of reporting in overviews (e.g., title, abstract, background, methods, results, discussion, other). At the end of the survey, two free text boxes will be provided for participants to suggest additional checklist items, and relevant methods papers. The Dillman principles for constructing respondent-friendly web surveys will be used to design the survey and its component items [36, 37]. Round one of the survey will remain open for 1 month (February–March 2020), during which bi-weekly reminder emails will be sent. The survey will be completed quasi-anonymously (i.e., the core project team but not the other study participants will know the identities and responses of the participants [31]), using a versatile online Delphi platform (e.g., Welphi). We will collate and summarize survey results, with agreement defined a priori as ≥ 70% for inclusion in (i.e., score of 3–4) or exclusion from (i.e., score of 1–2) the reporting guideline [38], based on the total number of responses obtained per question.

Round two: Online survey

In the second online survey (June–July 2020), the same participants will view and/or re-rate the checklist items presented in the first survey [32, 33]. The content, structure, and process will be similar to that of the first survey, with two differences. First, checklist items may be re-worded and/or re-formatted (e.g., items may be split or combined) based on the free-text comments collected in the first survey. Second, we will provide participants with their individual responses and a summary of anonymized group ratings, including all free-text comments, from the first survey. Participants will be asked to consider the structured feedback to inform their responses [32, 33]. Checklist items will be presented in the same order as previously. The items that reached agreement (≥ 70%) after the first survey will be presented for information purposes only (i.e. no more voting will occur, though participants may respond to free-text comments). The items that did not reach a high level of agreement during the first survey will be re-rated. The participants will be asked to determine whether and how they wish to modify their original answers in light of the group responses and comments; they may also provide feedback on free-text comments if desired. At the end of the survey, we will ask the participants to rate and provide comments for each new checklist item generated by participants during the first survey. We will collate and summarize survey results, with agreement defined as previously described (≥ 70%).

Round three: In-person meeting

With input from the advisory board, we will select and invite a minimum of 8 to 10 participants (i.e., “expert panelists”), ideally representing different stakeholder groups, to convene at a 1-day, in-person meeting (October 2020), where a nominal group technique and real-time voting will be used to systematically discuss and resolve outstanding disagreements [8, 33, 34]. If needed, a subset of these expert panelists may participate remotely using interactive software that allows for real-time screen sharing, audio discussion, and user comments (e.g., Adobe Connect, GoToMeeting). Moderators with expertise in overview methods and previous experience conducting consensus meetings for reporting guidelines will facilitate the meeting. We will audio record the meeting and take meeting minutes. The objective of the meeting will be to obtain final agreement on the list of items to be included in the PRIOR reporting guideline.

Prior to the meeting, members of the expert panel will receive a copy of their second-round survey results. We will begin the meeting by briefly summarizing the items reaching agreement for inclusion in, and exclusion from, the reporting guideline. No further voting will occur for these items, but outstanding free-text comments will be presented and resolved as needed; this could include rewording or reformatting of existing items. The bulk of the meeting will use a nominal group technique to obtain a high level of agreement (i.e., ≥ 70% agreement, as previously described) on those checklist items still lacking agreement [33, 34]. Each item will be reviewed sequentially, using a three-step process. First, facilitators will present the group survey results, all free-text comments, and all relevant methodological literature related to each item. This can help structure the interaction, provide a common starting point for participants, and promote evidence-based discussions about guideline content [33]. Second, the expert panel will discuss, debate, and aim to resolve discrepancies in a structured large-group discussion [33, 34]. Third, expert panelists will vote on the inclusion or exclusion of remaining items from the guideline.

The expert panel will be asked to re-rate the extent to which they agree with the inclusion of each checklist item. The content, structure, and process of the survey will be similar to the online surveys, with two changes: (1) the “I don’t know” option will no longer be available and (2) no free-text comments will be solicited. Expert panelists will complete the survey anonymously using a secure, online, live voting platform (e.g., Sli.do) that they will access using their personal electronic devices, with agreement defined as previously described (≥ 70%). Aggregate survey results will be automatically compiled by the software and presented to the group at the end of the survey. If agreement has not been reached on all items following the survey, we will ask the expert panelists to divide themselves into one small group per checklist item. Each small group will engage in unmoderated discussion to achieve final agreement, with rationale, for each outstanding item. Post-meeting discussions may continue over email or teleconference (i.e., including the expert panelists and facilitated by the core team) to achieve agreement, or we may implement an additional online survey with expert panelists, if needed. The meeting will conclude by discussing the strategy for producing, publishing, and disseminating the final guideline. We will discuss the inclusion of a flow diagram, development of an accompanying explanation and elaboration document, who will be involved in which activities, and publication and knowledge translation strategies [8].

Development of the guidance statement

A small writing group will iteratively draft the final version of the PRIOR guidance document based on the final decisions of the expert panel. The writing group will consist of the core project team, with an open invitation issued to the advisory board members and expert panel members. We will aim to provide clear, concise, and unambiguous wording for each PRIOR checklist item. The reporting guideline will be circulated among all advisory board members and expert panel members to obtain final input and approval prior to publication.

Discussion

Despite the growing number of published overviews and the commonly observed deficiencies in reporting of overviews [2,3,4], there are currently no systematically developed reporting guidelines for overviews of healthcare interventions. This protocol will help address this gap in guidance by using a four-stage process to develop the PRIOR reporting guideline, an evidence-based and agreement-based reporting guideline for overviews of healthcare interventions. Once completed, we will submit the PRIOR reporting guideline for publication to an appropriate peer-reviewed journal (we may also seek co-publication in multiple journals, if appropriate). The guideline will also be published on the EQUATOR website and the Cochrane Comparing Multiple Interventions Methods Group (CMIMG) website. We will aim to present the guideline at conferences and workshops, disseminate the guideline via email lists, and solicit journal editors to actively endorse the guideline. Other potential methods that we will use to disseminate the guideline could include videos, infographics, and social media.

The PRIOR reporting guideline will be developed using a modified Delphi process, which is commonly used to develop reporting guidelines in the health sciences [11, 26]. There are established benefits to using a modified Delphi process. For example, the online surveys provide a time- and cost-effective way to obtain preliminary agreement, while the more intensive in-person meeting with nominal group technique allows for face-to-face, in-depth discussion of outstanding issues [33]. Using an explicit, controlled, and scientifically credible process to achieve a high level of agreement among a group of experts can help to leverage the benefits of individual expertise and group decision-making, while simultaneously minimizing the biases associated with informal decision-making [33]. This can enhance the credibility of the guideline development process and help to ensure widespread acceptance and uptake of the reporting guideline. We expect some participant attrition between Delphi rounds; we will minimize this by following Dillman’s principles [36, 37] and conducting pilot tests to enhance the usability of the survey, providing bi-weekly email reminders, and choosing a survey platform that supports participant convenience by allowing them to leave and return at any time. Additionally, the PRIOR reporting guideline will only capture current expertise based on the existing state of knowledge, and we expect that requirements for reporting overviews will evolve over time as overview methods evolve. Thus, the aim of the PRIOR reporting guideline will not be to provide a definitive or unchanging list of reporting requirements, but rather to capture current expertise and knowledge upon which future research can build.

Once completed, the planned work will result in an evidence-informed, consensus-based reporting guideline for overviews of reviews of healthcare interventions. The PRIOR reporting guideline will help overview authors improve the accuracy, completeness, and transparency of reporting. It will also provide a framework for peer reviewers, journal editors, and healthcare decision-makers to critically appraise submitted or published overviews. Strengthening the reporting of overviews can help healthcare decision-makers better evaluate the reliability, validity, and applicability of overview results. This, in turn, can maximize the impact of overviews by allowing more accurate interpretation and use of their research findings.

Availability of data and materials

Not applicable.

Abbreviations

PRIOR:

Preferred Reporting Items for Overviews of Reviews

SR:

Systematic review

References

  1. Pollock M, Fernandes RM, Becker LA, Pieper D, Hartling L. Chapter V: overviews of reviews [draft]. In: Higgins JPT, Thomas J, Chandler J, Cumpston M, Li T, Page M, et al., editors. Cochrane Handbook for Systematic Reviews of Interventions. London: Cochrane; 2018.

    Google Scholar 

  2. Hartling L, Chisholm A, Thomson D, Dryden DM. A descriptive analysis of overviews of reviews published between 2000 and 2011. PLoS One. 2012;7(11):e49667.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  3. Pieper D, Buechter R, Jerinic P, Eikermann M. Overviews of reviews often have limited rigor: a systematic review. J Clin Epidemiol. 2012;65(12):1267–73.

    Article  PubMed  Google Scholar 

  4. Lunny C, Brennan S, Reid J, McDonald S, McKenzie J. Overviews of reviews incompletely report methods for handling overlapping, discordant and problematic data. J Clin Epidemiol. 2019. https://doi.org/10.1016/j.jclinepi.2019.09.025.

    Article  PubMed  Google Scholar 

  5. Lunny C, Brennan SE, McDonald S, McKenzie JE. Toward a comprehensive evidence map of overview of systematic review methods: paper 1 – purpose, eligibility, search and data extraction. Syst Rev. 2017;6(1):231.

    Article  PubMed  PubMed Central  Google Scholar 

  6. Lunny C, Brennan SE, McDonald S, McKenzie JE. Toward a comprehensive evidence map of overview of systematic review methods: paper 2 — risk of bias assessment; synthesis, presentation and summary of the findings; and assessment of the certainty of the evidence. Syst Rev. 2018;7(1):159.

    Article  PubMed  PubMed Central  Google Scholar 

  7. Pollock M, Fernandes RM, Becker LA, Featherstone R, Hartling L. What guidance is available for researchers conducting overviews of reviews of healthcare interventions? A scoping review and qualitative metasummary. Syst Rev. 2016;5(1):190.

    Article  PubMed  PubMed Central  Google Scholar 

  8. Moher D, Schulz KF, Simera I, Altman DG. Guidance for developers of health research reporting guidelines. PLoS Med. 2010;7(2):e1000217.

    Article  PubMed  PubMed Central  Google Scholar 

  9. Onishi A, Furakawa TA. Chapter 13: State-of-the-art reporting. In: Biondi-Zoccai G, editor. Umbrella reviews. Cham: Springer International Publishing; 2016.

    Google Scholar 

  10. Becker LA, Oxman AD. Chapter 22: Overviews of reviews. In: Higgins JPT, Green S, editors. Cochrane handbook for systematic reviews of interventions (version 5.1.0). London: The Cochrane Collaboration; 2011.

    Google Scholar 

  11. Liberati A, Altman DG, Tetzlaff J, Mulrow C, Gøtzsche PC, Ioannidis JP, et al. The PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate health care interventions: explanation and elaboration. BMJ. 2009;339:b2700.

    Article  PubMed  PubMed Central  Google Scholar 

  12. Shea BJ, Grimshaw JM, Wells GA, Boers M, Andersson N, Hamel C, et al. Development of AMSTAR: a measurement tool to assess the methodological quality of systematic reviews. BMC Med Res Methodol. 2007;7:10.

    Article  PubMed  PubMed Central  Google Scholar 

  13. Bougioukas KI, Liakos A, Tsapas A, Ntzani E, Haidich AB. Preferred reporting items for overviews of systematic reviews including harms checklist: a pilot tool to be used for balanced reporting of benefits and harms. J Clin Epidemiol. 2018;93:9–24.

    Article  PubMed  Google Scholar 

  14. Zorzela L, Loke YK, Ioannidis JP, Golder S, Santaguida P, Altman DG, et al. PRISMA harms checklist: improving harms reporting in systematic reviews. BMJ. 2016;352:i157.

    Article  PubMed  Google Scholar 

  15. Shamseer L, Moher D, Clarke M, Ghersi D, Liberati A, Petticrew M, et al. Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015: elaboration and explanation. BMJ. 2015;349:g7647.

    Article  Google Scholar 

  16. Bougioukas KI, Bouras E, Apostolidou-Kiouti F, Kokkali S, Arvanitidou M, Haidich AB. Reporting guidelines on how to write a complete and transparent abstract for overviews of systematic reviews of health care interventions. J Clin Epidemiol. 2019;106:70–9.

    Article  PubMed  Google Scholar 

  17. Beller EM, Glasziou PP, Altman DG, Hopewell S, Bastian H, Chalmers I, et al. PRISMA for abstracts: reporting systematic reviews in journal and conference abstracts. PLoS Med. 2013;10(4):e1001419.

    Article  PubMed  PubMed Central  Google Scholar 

  18. Li L, Tian J, Tian H, Sun R, Liu Y, Yang K. Quality and transparency of overviews of systematic reviews. J Evid Based Med. 2012;5(3):166–73.

    Article  PubMed  Google Scholar 

  19. Oxman AD, Guyatt GH. Validation of an index of the quality of review articles. J Clin Epidemiol. 1991;44:1271–8.

    Article  CAS  PubMed  Google Scholar 

  20. Singh J. Development of the metareview assessment of reporting quality (MARQ) checklist. Rev Fac Med. 2012;60:285–92.

    Google Scholar 

  21. Posadzki P. Standards for reporting of overviews of reviews and umbrella reviews (STROVI) statement. Cape Town: Cochrane Colloquium; 2017.

    Google Scholar 

  22. Shea BJ, Reeves BC, Wells G, Thuku M, Hamel C, Moran J, et al. AMSTAR 2: a critical appraisal tool for systematic reviews that include randomised or non-randomised studies of healthcare interventions, or both. BMJ. 2017;358:j4008.

    Article  PubMed  PubMed Central  Google Scholar 

  23. Plint AC, Moher D, Morrison A, Schulz K, Altman DG, Hill C, et al. Does the CONSORT checklist improve the quality of reports of randomised controlled trials? A systematic review. Med J Aust. 2006;185(5):263.18.

    Article  Google Scholar 

  24. Smidt N, Rutjes A, Van der Windt D, Ostelo R, Bossuyt P, Reitsma J, et al. The quality of diagnostic accuracy studies since the STARD statement: has it improved? Neurol. 2006;67(5):792–7.19.

    Article  CAS  Google Scholar 

  25. Simera I, Moher D, Hoey J, Schulz KF, Altman DG. The EQUATOR Network and reporting guidelines: helping to achieve high standards in reporting health research studies. Maturitas. 2009;63(1):4–6.

    Article  PubMed  Google Scholar 

  26. Hutton B, Salanti G, Caldwell DM, Chaimani A, Schmid CH, Cameron C, et al. The PRISMA extension statement for reporting of systematic reviews incorporating network meta-analyses of health care interventions: checklist and explanations. Ann Intern Med. 2015;162(11):777–84.

    Article  PubMed  Google Scholar 

  27. EQUATOR Network. Reporting guidelines under development for systematic reviews. EQUATOR. 2018. http://www.equator-network.org/library/reporting-guidelines-under-development/reporting-guidelines-under-development-for-systematic-reviews/#72. Accessed 16 Jan 2019.

  28. Pieper D, Pollock M, Fernandes RM, Büchter RB, Hartling L. Epidemiology and reporting characteristics of overviews of reviews of healthcare interventions published 2012–2016: protocol for a systematic review. Syst Rev. 2017;6(1):73.

    Article  PubMed  PubMed Central  Google Scholar 

  29. Greenhalgh T, Peacock R. Effectiveness and efficiency of search methods in systematic reviews of complex evidence: audit of primary sources. BMJ. 2005;331(7524):1064–5.

    Article  PubMed  PubMed Central  Google Scholar 

  30. Horsley T, Dingwall O, Sampson M. Checking reference lists to find additional studies for systematic reviews. Cochrane Database Syst Rev. 2011;8.

  31. Hasson F, Keeney S, McKenna H. Research guidelines for the Delphi survey technique. J Adv Nurs. 2000;32(4):1008–15.

    CAS  PubMed  Google Scholar 

  32. Hsu C-C, Sandford BA. The Delphi technique: making sense of consensus. PARE. 2010:344–7.

  33. Murphy M. Consensus development methods, and their use in clinical guideline development. Health Technol Assess. 1998;2(3):1–88.

    Article  Google Scholar 

  34. Jones J, Hunter D. Consensus methods for medical and health services research. BMJ. 1995;311(7001):376.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  35. Streiner DL, Norman GR, Cairney J. Health measurement scales: a practical guide to their development and use. New York: Oxford University Press; 2015.

    Book  Google Scholar 

  36. Dillman DA, Tortora RD, Bowker D. Principles for constructing web surveys. In: Joint meetings of the American Statistical Association; 1998.

    Google Scholar 

  37. Dillman DA. Mail and internet surveys: the tailored design method–2007 update with new internet, visual, and mixed-mode guide. New York: Wiley; 2011.

    Google Scholar 

  38. Diamond IR, Grant RC, Feldman BM, Pencharz PB, Ling SC, Moore AM, et al. Defining consensus: a systematic review recommends methodologic criteria for reporting of Delphi studies. J Clin Epidemiol. 2014;67(4):401–9.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

We thank Drs. Amanda Newton and Shannon Scott for input into the development of the protocol.

Funding

There is no specific funding for this project. ACT is funded by a Tier 2 Canada Research Chair in knowledge synthesis. LH is funded by a Tier 1 Canada Research Chair in knowledge synthesis and translation.

Author information

Authors and Affiliations

Authors

Contributions

All authors made substantial contributions to the conception and design of the study. MP drafted the protocol with input from RMF, DP, ACT, and LH. All authors reviewed the manuscript critically for important intellectual content and approve of the final version.

Corresponding author

Correspondence to Lisa Hartling.

Ethics declarations

We have obtained ethics approval from the University of Alberta Health Research Ethics Board (#Pro00086094). Participants will provide informed consent prior to participation.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Pollock, M., Fernandes, R.M., Pieper, D. et al. Preferred Reporting Items for Overviews of Reviews (PRIOR): a protocol for development of a reporting guideline for overviews of reviews of healthcare interventions. Syst Rev 8, 335 (2019). https://doi.org/10.1186/s13643-019-1252-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13643-019-1252-9

Keywords