Skip to main content

Methodological quality and reporting standards in systematic reviews with meta-analysis of physical activity studies: a report from the Strengthening the Evidence in Exercise Sciences Initiative (SEES Initiative)



Several resources have been developed (e.g., reporting guidelines) to promote high-standard practices in health research. However, there was no continuous and systematic assessment of recommended practices in published systematic reviews with meta-analysis (SRMAs), which increases the usability of the available resources. Therefore, we aimed to assess the methodological and reporting standards in SRMAs of physical activity studies. This report presents the main results of the SEES Initiative in 2019.


Our approach is based on a prospective systematic review methodology to implement post-publication surveillance of research practices in exercise sciences. Briefly, during the year 2019, pre-specified searches were conducted monthly (PubMed/MEDLINE) in journals from the exercise sciences (n = 9) and medicine (n = 5). The assessments were independently conducted by two authors, based on 36 items/practices derived from established statements/tools (PRISMA, AMSTAR 2, ROBIS). To be eligible, SRMAs should summarize studies that had, at least, one arm consisting of physical activity interventions/exposures and one health or behavioral outcome.


Out of 1028 studies assessed for eligibility, 103 SRMAs were included. The minimum adherence was 13/36 items, whereas only one SRMA adhered to all items. Some highly contemplated items included identification of title as SRMA (97.1%) and descriptions of the main outcome in the abstract (95.1%) and risk of bias (RoB) assessment (95.1%). Some poorly contemplated items included publicly available protocol (4.9%), discussion of the results in light of RoB in studies included (32.0%), and data sharing statements (35.9%).


In summary, there is a suboptimal adherence to recommended practices on methodological quality and reporting standards in the SRMAs of physical activity intervention/exposure evaluated from the selected journals in 2019, which likely reduce the reproducibility and usefulness of these studies. This incipient evidence from our first 12 months of post-publication surveillance should serve as a call for attention and action for multiple stakeholders (e.g., authors, reviewers, editors, funders, academic institutions) in this important health research field.

Peer Review reports


Over the last decades, there has been a consistent increase in the publication of systematic reviews with or without meta-analysis in the health sciences [1,2,3]. This type of literature review is accepted as a valuable source of evidence [4,5,6], being highly cited [7]. Additionally, systematic reviews with meta-analysis (SRMAs) are used to inform clinical decision-making, development of clinical practice guidelines and health policies [2, 8]. In the health sciences, numerous SRMAs have been poorly conducted with low methodological quality and poor reporting [2, 9,10,11]. In the exercise sciences, SRMAs and clinical trials are essential to inform strategies of prevention and treatment in several clinical conditions [12, 13]; however, methodological and reproducibility issues have been scarcely addressed [14,15,16]. Such evidence of low-quality practices affects reproducibility, since it makes findings less verifiable, credible, and informative [17], with likely waste of financial or non-financial resources [18].

Several strategies have been developed to improve the quality of evidence syntheses, such as open handbooks for standardized methods [19], registry platforms (e.g., International Prospective Register of Systematic Review (PROSPERO)), data sharing policies, reporting guidelines, and methodological quality assessment tools [20]. In this context, the “Preferred Reporting Items for Systematic Reviews and Meta-Analyses Statement” (PRISMA Statement) [21], a reporting guideline, and the “A MeaSurement Tool to Assess systematic Reviews” (AMSTAR 2) [22], a methodological quality assessment tool, are carefully conceived documents to improve the quality of SRMAs. However, reproducibility requires methodological rigor, adequate reporting, and transparency in the design, conduction, analysis, and critical appraisal of scientific evidence [18, 23], denoting that adherence to existent resources is of foremost importance. Since adherence is still quite low [1, 2], advocacy, education/training, implementation, and surveillance are issues to be addressed in a more intense and in-depth way [1, 2, 24, 25].

Thus, we have developed and implemented a discipline-based initiative with a surveillance approach to monthly evaluate the practices from the publications in exercise sciences, named SEES Initiative. Specifically, our purpose was to prospective and systematically assess methodological and reporting standards of systematic reviews with meta-analyses of physical activity from selected journals in the field of exercise sciences and general medicine, as well as to disseminate our findings publicly and provide direct feedback to authors and journal editors.

The present report brings the initial evidence of our first 12 months of assessments in 2019 in order to attract the attention of different stakeholders to the adherence or not of relevant aspects of the quality standard (methodological and reporting) in a considerable sample of SRMAs in the exercise sciences published in top-tier journals and, finally, show the importance of conducting such initiative.


Institutional review board approval and informed consent were not required, since this study used publicly available data from the SRMAs included. The Preferred Reporting Items for Systematic Reviews and Meta-Analyses: The PRISMA Statement and the PRISMA checklist were used to ensure accurate reporting (see Additional file 1).

Design and organization

The Strengthening the Evidence in Exercise Sciences Initiative (SEES Initiative) is a collaborative, nonprofit, living project for assessment of published research (i.e., randomized controlled trials and SRMAs) and dissemination of recommended practices of transparency, reproducibility, and integrity. This project was launched in January 2019, and the results to be presented herein refer to the SRMA assessments from January to December 2019. Our protocol is available at the Open Science Framework repository (OSF) [26].

The SEES Initiative is conducted by trained collaborators organized in three committees. The pre-assessment committee is responsible for literature searches in selected journals and retrieval. The assessment committee carries out the assessment of eligibility criteria and conducts data extraction on randomized controlled trials and SRMAs. The post-assessment committee conducts the data management (i.e., analyses, storage, and sharing) and dissemination of evidence appraisal to authors, journal editors, and websites. To ensure internal consistency, planned tasks of each committee follow standardized operational procedures (

These collaborators do not have exclusive dedication or receive any payment. Therefore, to achieve the SEES Initiative ambitious aim of continuously conducting post-publication surveillance, disseminating the findings publicly, and providing direct feedback to authors and journal editors, we had to select and limit the number of journals to carry out literature searches. The criteria used for the journal’s selection are described in the section “Literature search”.

Our methodological design relates to a meta-research prospective approach, mostly regarding post-publication analyses and, therefore, submission to ethical committees is not applicable for this study design.

Eligibility criteria

We included SRMAs which synthesized primary studies having at least one intervention arm of physical activity or exercise programs with a research question related to health outcomes or health behaviors at a minimal extent. In addition, SRMAs of observational studies with well-defined physical activity exposures were eligible for inclusion, as well as reviews with exploratory approaches (e.g., meta-regression) using meta-analytic techniques.

Literature search

We conducted searches in PubMed/MEDLINE between the 3rd and 7th day of each month of 2019. To reduce the burden with eligibility analyses, after conducting pilot tests and examining specific literature, search strategies (see Additional file 2) were built using highly sensitive filters for SRMAs [27, 28]. Moreover, we applied date filters restricting the searches for the two past months. For example, February was included in the search conducted both in March as well as in April. As expected, this strategy resulted in duplicated references because each month was looked up twice. Since MEDLINE/PubMed does have some timing variability in indexing of references, we implemented such overlapping procedure because non-overlapped searches yielded loss of some references in our pilot tests.

The searches were conducted and retrieved on a monthly basis in 9 exercise science journals (American Journal of Sports Medicine, British Journal of Sports Medicine, European Journal of Preventive Cardiology, International Journal of Behavioral Nutrition and Physical Activity, Journal of Physiotherapy, Journal of Science and Medicine in Sport, Medicine and Science in Sports and Exercise, Scandinavian Journal of Medicine & Science in Sports and Sports Medicine) and on a quarterly basis from 5 general medicine journals (Annals of Internal Medicine, British Medical Journal, Journal of the American Medical Association, Lancet, and New England Journal of Medicine). The choice of these exercise science journals was guided by the audience reach, for which we considered whether a given journal was linked to a professional or scientific society, journal impact factor, and by the expected volume of SRMA publications. Regarding the general medicine journals, our choice was based on the public reach and media coverage that articles addressing physical activity exposures/interventions usually achieve whenever published by these journals. The protocol document presents further reasoning regarding the journals’ choice [26].


Two trained collaborators (N.L.O and C.E.B) composed the assessment committee for SRMAs and screened independently the title/abstracts for eligibility. The same collaborators read the full texts of potential eligible SRMAs for a final selection. Any disagreements were solved by a third collaborator (D.U.). Full data set with publications assessed for eligibility is publicly available on OSF (

Data extraction (SRMA assessment)

We set up a standardized form to assess the methodological and reporting quality of the SRMAs, which was completed independently by two collaborators (N.L.O and C.E.B). Disagreements were resolved with participation of a third collaborator (D.U.). The assessment form was constructed based on documents widely available and referred to in the literature. The questions were therefore operationalized based on recommendation items from the PRISMA Statement [21], the AMSTAR 2 [22], and the “Risk of Bias in Systematic Reviews” (ROBIS) [29]. Additional file 3 presents the items included in the assessment form and respective guideline/tools of reference.

In this report, methodological quality refers to how well SRMA was planned and conducted, taking into account the processes of analysis, synthesis, and critical appraisal of the selected evidence [22]. In addition, the quality of reporting refers to the details and completeness of the information throughout the research report [21].

Items and domains

The SRMAs (unit of analysis) were analyzed by (1) cumulative number of items (from 0 to 36) in accordance with recommended practices and (2) seven domains that aggregate these items based on contributions in methodological or reporting aspects. The seven domains were transparency (n = 4 items), completeness (n = 11), methodological rigor (n = 7), participants (n = 2), interventions/exposures (n = 2), outcome (n = 5), and critical appraisal (n = 5). Although the domains are intrinsically related, items were distributed in an exclusive approach to avoid overlapped counts (which is only present for one item associated with two domains in the SRMA assessment) (see Additional file 3). We underscore that the “domains” are labeled as “components” both on the website and in the SEES Initiative protocol [26]. In this report, registration and methodological protocol were considered separately. When we cite “publicly available methodological protocol,” we are referring to a detailed protocol that allows the reproducibility of the study and is publicly available, but not necessarily published in a journal. This document may be available on an OSF page, Google Drive link, institutional webpages, Google sites, among others.

Dissemination of individual results

After SRMA assessments were completed, we disseminated reports by three means: (1) summarized reports at our website (, (2) deposit of full reports were filed in OSF ( folder: assessment reports/SRMA), and (3) full reports sent by email to the corresponding authors. We encouraged authors to send us requests for clarification or correction.

Statistical analysis

A general summary is presented in median and interquartile range (25th–75th percentiles). Data for each item evaluated are presented as absolute frequencies (n) and percentages of SRMA publications that received each possible answer (for example, yes or no). Regarding the domains, the data are summarized as frequencies and percentages of SRMAs that scored as “Yes” or a “positive” evaluation in a specific number of items in each individual domain. In most items, the option “Yes” was indicative of a recommended practice (item) that was followed in the SRMA. Items that did not have a binary (yes/no) response option included those related to the time range of literature search, duplicated process for selection, extraction, and assessment of risk of biases (yes vs. partially yes), number of languages considered for study eligibility, description of risk of bias within studies, description for protocol deviations during the course of the review, and discussion limitations at the study/outcome and/or at the review level. Another type of exception regards the occurrence of a potential spin bias, in which “No” is the positive (recommended) evaluation. We describe these exceptions in detail and their possible operationalization in Additional file 4.

Although we did compute a cumulative frequency of adherence to recommended practices per study, we chose not to disseminate study results primarily based on scores, grades, or single index, because the use of such metrics can hide critical flaws in important items for reproducibility, credibility, and trustworthiness of the SRMA findings [22]. In this sense, we attempt to communicate such frequencies by also showing all the items included in the assessment.

An exploratory analysis was performed to evaluate whether the adherence to a recommended practice (item) is different between SRMAs registered and not registered a priori and between SRMAs with open access and without open access. The chi-square test was used for this analysis. However, when at least 80% of the expected frequencies (counts) are not equal to or greater than 5, the Fisher exact test was used.

The data was entered into a Google Sheet, which was generated from a standardized Google Form, filled in duplicate. All variable manipulation and related analyses were carried out using R software version 4.0.0 (R Development Core Team, Vienna, Austria). The R script and data are openly available in our study repository ( Statistical significance was set at p < 0.05.


A total of 1028 records were retrieved from the electronic literature searches performed monthly in 2019. Nine hundred and eighty-five were screened by title and abstract, and 882 were excluded. Thus, after confirmation by full-text screening, 103 SRMAs were eligible for assessment (Fig. 1). Only one SRMA included was published in general medicine journals. Most publications were from Sports Medicine (n = 36) and British Journal of Sports Medicine (n = 34) (see Additional file 5).

Fig. 1
figure 1

Flow diagram of the study selection process. SRMAs, systematic reviews with meta-analysis

Overall, one SRMA addressed appropriately the 36 core items (Fig. 2; see Additional file 6 for more details). The median of items that received “Yes” or other “positive” evaluation was 28 items (24–29 items). The item regarding the identification as a systematic review, meta-analysis, or both in the study title (completeness domain) was the one with the highest adherence (100/103, 97.1%). On the other hand, the item regarding the existence of a publicly available methodological protocol (transparency domain) was the one with the lowest adherence (5/103, 4.9%) (Table 1).

Fig. 2
figure 2

Overview of the adherence to recommended practices: items (n = 36) per SRMA (n = 103). Rows refer to items/recommended practices (n = 36) and columns refer to systematic reviews with meta-analysis (n = 103). The studies are organized in descending order of adherence to the items/recommended practices (from left to right). Four items of transparency: 1T: Registration; 2T: Protocol; 3T: Available searches; 4T: Data Statement. Eleven items of completeness: 1C: Title as SRMA; 2C: Data sources (abstract); 3C: Key eligibility criteria (abstract); 4C: Number of included studies (abstract); 5C: Research question; 6C: PICOS explanation; 7C: Number of references; 8C: Description of sample sizes; 9C: Duration of included studies; 10C: Sources of funding; 11C: Potential conflicts of interest. Two items of participants: 1P: Description of participants (abstract); 2P: Detailed studies' characteristics. Two items of interventions/exposures: 1I: Description of interventions/exposures (ab); 2I: Detailed studies' characteristics. Five items of outcome: 1O: Main outcome of interest (abstract); 2O: Statistical methods; 3O: Statistical heterogeneity; 4O: Meta-analytic summary estimates; 5O: Statistics per study. Seven items of methodological rigor 1M: Searches in grey literature; 2M: Searches from inception or with justification; 3M: Number of languages; 4M: Study selection in duplicate; 5M: Data extraction in duplicate; 6M: Description of Risk of Bias assessment; 7M: Risk of Bias assessment in duplicate. Five items of critical appraisal: 1Cr: Risk of Bias results within studies; 2Cr: Description of protocol deviations; 3Cr: Presence of spin bias; 4Cr: Discussion addressing Risk of Bias; 5Cr: Limitations thoroughly addressed. The list with the identification of the studies in this same order is available in the Additional file 7

Table 1 Frequency distribution of items/recommended practices by domains for the total number of systematic reviews with meta-analysis (SRMAs)


For the four items of this domain, only 2.9% (3/103) of SRMAs received “yes” on all (full yes) but 11.7% (12/103) received “no” on all (full no). The item with the most positive evaluation was related to the availability of at least one full-search query (83/103, 80.6% of SRMAs) (Table 1). Just over a third of the reports made a statement regarding the data sharing (37/103, 35.9%) and just over half (57/103, 55.3%) were registered in a public database.


Thirty SRMAs (29.1%) complied to all the 11 items of this domain. In addition, 7 items presented a high percentage of “yes” (93.2% to 97.1%) among the SRMAs. However, 31.1% (32/103) of SRMAs did not report a detailed explanation of eligibility criteria for PICOS elements (population, intervention, comparator, outcome, setting), and the abstract of 25.2% SRMAs (26/103) did not inform some of the key eligibility criteria for study selection (Table 1).

Participants and interventions/exposures

These two domains were composed of two items, one of which (i.e., full description of PICOS elements in the Results section) is repeated in both domains. Eighty-eight (85.4%) SRMAs scored “yes” in this item. Also, 69 (67.0%) SRMAs received full yes in the participants domain and 82 (79.6%) SRMAs in the interventions/exposures domain. However, 21.4% (22/103) of SRMAs did not provide a description regarding the population (participants) or main condition(s) addressed in the abstract section (Table 1).


Thirty-five SRMAs (34.0%) complied to all the 5 items of this domain. The item with the most positive evaluation was related to the result description for the main outcome in the abstract (98/103, 95.1%). However, almost half of SRMAs (46/103, 44.7%) failed to report a full description of individual results for studies composing the meta-analysis (i.e., effects size, imprecision measure, and percentage weight) in the Results section. Moreover, 30.1% (31/103) of SRMAs did not describe meta-analytic summary estimates as recommended (i.e., binary outcomes as frequencies with and without the event (or as proportions such as 12/45); continuous outcomes as the mean, standard deviation, and sample size for each group) (Table 1).

Methodological rigor

A full yes was reached in 9.7% (10/103) of SRMAs in this domain of 7 items. The item with the most positive evaluation was related to the description of the risk of bias assessment in the Methods section (98/103, 95.1%), but 34.0% (35/103) of SRMA did not perform this assessment in duplicate. In addition, most of SRMAs considered only one language to study eligibility (62/103, 60.2%) and almost half did not extract data in duplicate (47/103, 45.6%) (Table 1).

Critical appraisal

Eighteen SRMAs (17.5%) presented a full yes in this domain. A high percentage of SRMAs described the risk of bias within included studies in the Results section (89/103, 86.4%). However, only 32.0% (33/103) of SRMAs discussed the results in light of the risk of biases in individual studies. A description for non-planned modifications during the course of the review was not carried out in 17.5% (18/103) after comparing the published report to registration. Also, this comparison was not possible in 40.8% (42/103) of SRMAs due to the lack of registration (Table 1).

Answers to SEES emails

Nine of the 103 corresponding authors questioned the assessment of some items. The median of claims was 4 items, ranging from 1 to 10. These SRMAs are identified in Additional file 7.

Exploratory analysis

The registered SRMAs, when compared to those without registration, addressed more adequately the following items/recommended practices: availability of at least one full-search query (transparency domain) (p = 0.002), key eligibility criteria for study selection (completeness domain) (p = 0.045), detailed explanation of eligibility criteria for PICOS (completeness domain) (p = 0.014), statement of funding sources (completeness domain) (p = 0.043), potential conflicts of interest (completeness domain) (p = 0.043), description of the risk of bias assessment (methodological rigor domain) (p = 0.016), risk of bias results within studies (critical appraisal domain) (p = 0.000), and description for non-planned modifications (critical appraisal domain) (p = 0.000) (see Additional file 8). Regarding open access, we observed that 27 SRMAs had this condition and only two items, key eligibility criteria for study selection (completeness domain) (p = 0.049) and search in gray literature (methodological rigor) (p = 0.030), showed statistical significance, with greater adherence in the open access SRMAs (see Additional file 9).


This study presents the main report of methodological and reporting surveillance on SRMAs of physical activity intervention/exposure assessed by the SEES Initiative in the year of 2019. Across the seven domains operationalized to aggregate the items/recommended practices by the PRISMA Statement, ROBIS, and AMSTAR 2 tools, some of the most adopted practices were in the completeness domain: identification as a systematic review, meta-analysis, or both in the study title (97.1%) and more 2 items with 95.1%; outcome domain: description of the main outcome of interest in the abstract (95.1%); and methodological rigor domain: description of RoB assessment (95.1%). In opposite, the existence of a publicly available methodological protocol (4.9%) (transparency domain), discussion of the results in light of RoB in individual studies (32.0%) (critical appraisal domain), and statement on data sharing (35.9%) (transparency domain) were among the practices least adhered to, therefore indicating possible emphases that can be addressed through educational resources and editorial policies.

Practices most and least present in assessed articles

In the transparency domain, modest existence of reviews registered in public databases and only a minimal fraction (5/103) presenting a detailed methodological protocol in addition to the registration may reduce the potential for methodological reproducibility [30]. Although most SRMAs retrieve disseminated and aggregated data, data sharing statements were still less than modest. There are similar results indicating low adherence to practices of registration [1, 24] or data sharing [31] in other disciplines, denoting that transparent practices need more implementation in varied research fields.

We observed that most studies adhered to items under the completeness, participants, and interventions/exposures domains. Taken together, these three domains included 14 items, which allow stakeholders to understand the scope of the study, the research question, its applicability, and, finally, the replication of the study if desired [21]. However, our results indicate that the information on the eligibility criteria for the selection of individual studies was not adequately reported in a relevant number of SRMAs, either in abstracts (25.2%) or in methods (31.1%). Furthermore, 22.4% of SRMAs did not provide a clear description of participants or main conditions under study in the abstract section. We underscore the importance of such information since it is key for stakeholders to decide whether to access (possibly by payment) the article.

In domains of methodological rigor, outcome reporting, and critical appraisal, we combined items (practices) that could more directly affect the interpretation of evidence generated by SRMAs. Surprisingly, several reviews did lack duplicate processes in data extraction and risk of bias assessments. Although similar results were observed in meta-research studies in health sciences [1, 24, 32], it is widely recommended that these procedures be performed by two independent researchers or primarily conducted by one researcher and fully verified by another one [22, 29]. Additionally, we found that nearly two-thirds of SRMAs (60.3%) considered only the English language to study eligibility. Such restriction is commonly applied in literature searches [33], but can introduce a systematic error, a language bias, possibly modifying meta-analytic estimates [33, 34]. In the outcome domain, items such as individual results for each analyzed study and meta-analytic summary estimates still need substantial reporting improvement to facilitate the understanding and replicability of meta-analyses. These items were considered as properly reported when all recommended elements were addressed; therefore, our assessment may have been very strict. However, to note, previous studies have also reported similar findings in the biomedical sciences [31, 35].

In the critical appraisal domain, more than 80% of our sample presented the risk of bias assessment; however, only a third (33/103) discussed their review findings in the context of potential biases identified through the assessment. Part of our results are in line with data by Buttner et al. [36], who focused only on the risk of bias assessment, and also described a high percentage of these assessments in systematic reviews and SRMAs of exercise interventions/exposures published in a single journal. Moreover, studies in biomedical literature have also reported a reduced use of this type of assessment in the discussion of SRMAs [9, 32], therefore, in agreement with our findings. In contrast, Buttner et al. [36] found more favorable results, indicating that risk of bias assessments were incorporated in the interpretation/discussion sections of 86% of their sample. Although both studies, Buttner et al. [36] and ours, have distinct samples and criteria to assess this practice, we reason that, especially when the risk of bias is high in individual studies, such results should be thoroughly addressed in Discussion sections. Finally, we assessed whether SRMAs reported methodological changes during the course of the review, which were not planned as described in the registry or methodological protocol. These changes in design, conduct, or analysis can introduce bias in the review findings and therefore should be reported to the readers’ knowledge [22, 29]. Unfortunately, our results show that 17.5% (18/103) of SRMAs made unplanned changes and did not describe them in the final publication. Additionally, in 41.7% (43/103) of SRMAs, it was not possible to carry out this verification due to the lack of a public registration record, which again reinforces the need for transparent practices at the pre-study stage.

The results observed show the importance of this type of initiative, with a characteristic of “post-publication criticism,” for the realization of a call for attention and action for the different stakeholders. It is noteworthy that the SEES Initiative could have a significant impact if authors and editors had actively engaged in the cordial dialogue about the results of their manuscript assessments, proposed by this initiative in the emails sent to these key actors. However, the reluctance to respond to criticism does not seem to be unique to the field of exercise science [37].

Prospective registration and open access publication

The results of our exploratory analysis indicate that prospective registration is associated with better adherence to some relevant items/recommended practices in the domains of transparency, completeness, methodological rigor, and critical appraisal. To our knowledge, there are no studies on the influence of the prospective registration on the methodological and reporting standards of SRMAs in exercise sciences, but a recent study investigated this topic in systematic reviews of healthcare interventions [38]. In a similar way to our study, they evaluated a sample (n = 150) of systematic reviews published in 1 year (i.e., 2015) through PRISMA and AMSTAR tool. However, unlike us, they used total scores for these analyses. As main results, they observed that the total scores on PRISMA and AMSTAR were higher for those systematic reviews registered a priori, being significant only for the total scores on AMSTAR. Therefore, they concluded that prospective registration may at least indirectly improve the overall methodological quality of systematic reviews. Regarding open access, Pastorino et al. [39], evaluating only 47 systematic reviews with and without meta-analysis in the field of oncology (open access, n = 15 and no open access, n = 32), published in 2013, observed that the overall methodological and reporting standards were comparable between studies with and without open access. Similar to our study, they found a significant difference in only one item (description of the methods used to assess the risk of bias in individual studies). However, one must be aware of the large amount of statistical comparisons, which increases the probability that a difference is significant by chance. In addition, these stratified analyses were not planned. Therefore, studies specifically designed to investigate these issues (registration and open access) must be conducted to confirm these findings.

Strengths and limitations

Several strategies exist to improve the methodological and reporting quality of SRMAs in the biomedical literature [18, 20, 24, 32, 40,41,42]. Our study highlights that solely relying on the availability of these documents/tools could be insufficient for adoption. In exercise sciences, there is a need for most journals to provide endorsement of the guidelines, which may trigger quality improvements [37, 38]. The use of guidelines in the peer review process has also been proposed [24, 32]. We point out that the main strength of this study is the “translational approach,” since we used established guidelines in a lively project, giving feedback to editors and authors (despite the low number of replies). In methodological grounds, we did not homogenize our assessments through a total score, allowing readers a more detailed view of adherence to items and domains. Indeed, generating such scores is not recommended, since it assumes that all items have the same weight [22]. Finally, we included some ROBIS items in our assessment form, therefore addressing more directly the methodological quality assessment of the SRMAs.

Some limitations should be considered. First, our sample of SRMAs represents only the year 2019 and only a few pre-selected scientific journals. Therefore, the results observed in our assessments are not promptly generalizable to other SRMA publications in exercise sciences. However, it seems unlikely that the quality of reporting of previous SRMAs is superior to the current ones, given that important documents providing guidance and recommendations for planning, conducting, and reporting SRMAs were published a few years ago (i.e., PRISMA in 2009, AMSTAR 2 in 2017, ROBIS in 2016); our results are in line with several studies of this nature in other health disciplines. In addition, the selected journals had features such as connection to a professional or scientific societies, considerable public reach, and impact factor. Since these features increase visibility, we believe the quality of studies is a common priority among the selected journals. Finally, unlike other studies of this kind, we proposed to assess SRMAs prospectively and systematically (i.e., monthly basis), as well as to provide direct feedback to authors and journal editors, encouraging them to question the assessments when they do not agree with these. Thus, focusing on some journals became essential to increase our chances of being able to contemplate this proposal. Second, a possible limitation is the fact that our assessment form has not been formally validated. Although the form was elaborated on the basis of a reporting guideline for systematic reviews (i.e., PRISMA) and well-planned methodological quality assessment tools (i.e., AMSTAR and ROBIS), testing the properties of our form would be important.


There was a heterogeneous pattern for adherence to reporting standards proposed by the PRISMA Statement in SRMAs in exercise sciences, with structure (completeness) at a better status than practices related to transparency and full information for adequate critical appraisal. This suboptimal (sometimes low) adherence to recommended practices on quality of reporting and methodological rigor limits the reproducibility and credibility of the SRMA findings, and consequently, the evidence uptake for clinical decision-making. In other words, this can lead to an avoidable waste of resources and, ultimately, impair the application/use of the physical activity interventions investigated, the results achieved by those undergoing these interventions (patients/individuals in general), and the overall confidence in the exercise sciences. Therefore, the implementation of strategies such as SEES, which proposes to monitor published research continuously and discuss its findings openly and cordially with the authors and editors, can be a way to improve the synthesis of evidence in the field of exercise sciences. However, it seems necessary to discuss aims, outcomes, and diversify strategies to increase engagement and responsibility of all stakeholders. Finally, as future directions, we can point out that after observing the overall picture with this study, we will move on to identifying critical areas to focus on, with the few resources we have, and explore it in further publications and educational actions in social media (e.g., exploration of integrity between outcomes defined in the registry and/or protocol and the outcomes reported in the final manuscript).

Availability of data and materials

All data generated by the SEES Initiative are publicly available via Open Science Framework (; DOI: under a CC-By Attribution 4.0 International license. Data sheets: Eligibility (; SRMA assessments (





A MeaSurement Tool to Assess systematic Reviews


Population, intervention, comparator, outcome, setting


Preferred Reporting Items for Systematic Reviews and Meta-Analyses Statement


Risk of bias


Risk of Bias in Systematic Reviews

SEES Initiative:

Strengthening the Evidence in Exercise Sciences Initiative


Systematic reviews with meta-analysis


  1. Pussegoda K, Turner L, Garritty C, Mayhew A, Skidmore B, Stevens A, et al. Systematic review adherence to methodological or reporting quality. Syst Rev. 2017;6:131.

    Article  Google Scholar 

  2. Page MJ, Shamseer L, Altman DG, Tetzlaff J, Sampson M, Tricco AC, et al. Epidemiology and reporting characteristics of systematic reviews of biomedical research: a cross-sectional study. PLoS Med. 2016;13:e1002028.

    Article  Google Scholar 

  3. Bastian H, Glasziou P, Chalmers I. Seventy-five trials and eleven systematic reviews a day: how will we ever keep up? PLoS Med. 2010;7:e1000326.

    Article  Google Scholar 

  4. Howick J, Chalmers I, Glasziou P, Greenhalgh T, Heneghan C, Liberati A, Moschetti I, Phillips B, Thornton H. Explanation of the 2011 Oxford Centre for Evidence-Based Medicine (OCEBM) levels of evidence (background document). Oxford Centre for Evidence-Based Medicine. Accessed 12 July 2020.

  5. Harbour R, Miller J. A new system for grading recommendations in evidence based guidelines. BMJ. 2001;323:334–6.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  6. Krist AH, Wolff TA, Jonas DE, Harris RP, LeFevre ML, Kemper AR, et al. Update on the methods of the U.S. Preventive Services Task Force: methods for understanding certainty and net benefit when making recommendations. Am J Prev Med. 2018;54:S11–8.

    Article  Google Scholar 

  7. Patsopoulos NA, Analatos AA, Ioannidis JPA. Relative citation impact of various study designs in the health sciences. JAMA. 2005;293:2362–6.

    Article  CAS  Google Scholar 

  8. Moher D, Tetzlaff J, Tricco AC, Sampson M, Altman DG. Epidemiology and reporting characteristics of systematic reviews. PLoS Med. 2007;4:e78.

    Article  PubMed  PubMed Central  Google Scholar 

  9. Xu C, Liu Y, Jia P-L, Li L, Liu T-Z, Cheng L-L, et al. The methodological quality of dose-response meta-analyses needed substantial improvement: a cross-sectional survey and proposed recommendations. J Clin Epidemiol. 2019;107:1–11.

    Article  Google Scholar 

  10. Brugha TS, Matthews R, Morgan Z, Hill T, Alonso J, Jones DR. Methodology and reporting of systematic reviews and meta-analyses of observational studies in psychiatric epidemiology: systematic review. Br J Psychiatry. 2012;200:446–53.

    Article  PubMed  Google Scholar 

  11. Adie S, Ma D, Harris IA, Naylor JM, Craig JC. Quality of conduct and reporting of meta-analyses of surgical interventions. Ann Surg. 2015;261:685–94.

    Article  PubMed  Google Scholar 

  12. Pedersen BK, Saltin B. Exercise as medicine - evidence for prescribing exercise as therapy in 26 different chronic diseases. Scand J Med Sci Sports. 2015;25:1–72.

    Article  PubMed  Google Scholar 

  13. Bull FC, Al-Ansari SS, Biddle S, Borodulin K, Buman MP, Cardon G, et al. World Health Organization 2020 guidelines on physical activity and sedentary behaviour. Br J Sports Med. 2020;54:1451–62.

    Article  Google Scholar 

  14. Adherence to Consolidated Standards of Reporting Trials (CONSORT) Guideline Items in Randomized Trials of Physical Activity Published in 5 Sports Medicine Journals | Peer Review Congress. Accessed 12 July 2020.

  15. Silverman SR, Schertz LA, Yuen HK, Lowman JD, Bickel CS. Systematic review of the methodological quality and outcome measures utilized in exercise interventions for adults with spinal cord injury. Spinal Cord. 2012;50:718–27.

    Article  CAS  Google Scholar 

  16. Abell B, Glasziou P, Hoffmann T. Reporting and replicating trials of exercise-based cardiac rehabilitation: do we know what the researchers actually did? Circ Cardiovasc Qual Outcomes. 2015;8:187–94.

    Article  Google Scholar 

  17. Goodman SN, Fanelli D, Ioannidis JPA. What does research reproducibility mean? Sci Transl Med. 2016;8:341ps12.

    Article  PubMed  Google Scholar 

  18. Glasziou P, Altman DG, Bossuyt P, Boutron I, Clarke M, Julious S, et al. Reducing waste from incomplete or unusable reports of biomedical research. Lancet. 2014;383:267–76.

    Article  Google Scholar 

  19. Website. Website. Higgins JPT, Thomas J, Chandler J, Cumpston M, Li T, Page MJ, Welch VA (editors). Cochrane handbook for systematic reviews of interventions version 6.1 (updated September 2020). Cochrane; 2020. Available from Accessed 20 Dec 2020.

  20. Munafò MR, Nosek BA, Bishop DVM, Button KS, Chambers CD, Percie du Sert N, et al. A manifesto for reproducible science. Nat Hum Behav. 2017;1:e124.

    Article  Google Scholar 

  21. Liberati A. The PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate health care interventions: explanation and elaboration. Ann Intern Med. 2009;151:W.

    Article  Google Scholar 

  22. Shea BJ, Reeves BC, Wells G, Thuku M, Hamel C, Moran J, et al. AMSTAR 2: a critical appraisal tool for systematic reviews that include randomised or non-randomised studies of healthcare interventions, or both. BMJ. 2017;358:j4008.

    Article  PubMed  PubMed Central  Google Scholar 

  23. Ioannidis JPA, Greenland S, Hlatky MA, Khoury MJ, Macleod MR, Moher D, et al. Increasing value and reducing waste in research design, conduct, and analysis. Lancet. 2014;383:166–75.

    Article  Google Scholar 

  24. Sharma S, Oremus M. PRISMA and AMSTAR show systematic reviews on health literacy and cancer screening are of good quality. J Clin Epidemiol. 2018;99:123–31.

    Article  Google Scholar 

  25. Blanco D, Altman D, Moher D, Boutron I, Kirkham JJ, Cobo E. Scoping review on interventions to improve adherence to reporting guidelines in health research. BMJ Open. 2019;9:e026589.

    Article  Google Scholar 

  26. Umpierre D, De Nardi AT, Botton CE, Helal L, Santos LP, Oliveira NL. SEES-Initiative protocol: Open Science Framework; 2019.

  27. Boynton J, Glanville J, McDaid D, Lefebvre C. Identifying systematic reviews in MEDLINE: developing an objective approach to search strategy design. J Inf Sci. 1998;24:137–54.

    Article  Google Scholar 

  28. Lee E, Dobbins M, Decorby K, McRae L, Tirilis D, Husson H. An optimal search filter for retrieving systematic reviews and meta-analyses. BMC Med Res Methodol. 2012;12:51.

    Article  Google Scholar 

  29. Whiting P, Savović J, Higgins JPT, Caldwell DM, Reeves BC, Shea B, et al. ROBIS: a new tool to assess risk of bias in systematic reviews was developed. J Clin Epidemiol. 2016;69:225–34.

    Article  Google Scholar 

  30. Stewart L, Moher D, Shekelle P. Why prospective registration of systematic reviews makes sense. Syst Rev. 2012;1.

  31. Page MJ, Altman DG, Shamseer L, McKenzie JE, Ahmadzai N, Wolfe D, et al. Reproducible research practices are underused in systematic reviews of biomedical interventions. J Clin Epidemiol. 2018;94:8–18.

    Article  Google Scholar 

  32. Nissen T, Wayant C, Wahlstrom A, Sinnett P, Fugate C, Herrington J, et al. Methodological quality, completeness of reporting and use of systematic reviews as evidence in clinical practice guidelines for paediatric overweight and obesity. Clin Obes. 2017;7:34–45.

    Article  CAS  Google Scholar 

  33. Grégoire G, Derderian F, Le Lorier J. Selecting the language of the publications included in a meta-analysis: is there a tower of babel bias? J Clin Epidemiol. 1995;48:159–63.

    Article  PubMed  Google Scholar 

  34. Pham B, Klassen TP, Lawson ML, Moher D. Language of publication restrictions in systematic reviews gave different results depending on whether the intervention was conventional or complementary. J Clin Epidemiol. 2005;58:769–76.e2.

    Article  PubMed  Google Scholar 

  35. Wasiak J, Tyack Z, Ware R, Goodwin N, Faggion CM Jr. Poor methodological quality and reporting standards of systematic reviews in burn care management. Int Wound J. 2017;14:754–63.

    Article  Google Scholar 

  36. Büttner F, Winters M, Delahunt E, Elbers R, Lura CB, Khan KM, et al. Identifying the “incredible”! Part 2: spot the difference - a rigorous risk of bias assessment can alter the main findings of a systematic review. Br J Sports Med. 2020;54:801–8.

    Article  PubMed  Google Scholar 

  37. Schriger DL, Altman DG. Inadequate post-publication review of medical research. BMJ. 2010;341:c3803.

    Article  Google Scholar 

  38. Ge L, Tian J-H, Li Y-N, Pan J-X, Li G, Wei D, et al. Association between prospective registration and overall reporting and methodological quality of systematic reviews: a meta-epidemiological study. J Clin Epidemiol. 2018;93:45–55.

    Article  Google Scholar 

  39. Pastorino R, Milovanovic S, Stojanovic J, Efremov L, Amore R, Boccia S. Quality assessment of studies published in open access and subscription journals: results of a systematic evaluation. PLoS One. 2016;11:e0154217.

    Article  Google Scholar 

  40. Panic N, Leoncini E, de Belvis G, Ricciardi W, Boccia S. Evaluation of the endorsement of the preferred reporting items for systematic reviews and meta-analysis (PRISMA) statement on the quality of published systematic review and meta-analyses. PLoS One. 2013;8:e83138.

    Article  Google Scholar 

  41. Cullis PS, Gudlaugsdottir K, Andrews J. A systematic review of the quality of conduct and reporting of systematic reviews and meta-analyses in paediatric surgery. PLoS One. 2017;12:e0175213.

    Article  Google Scholar 

  42. Blanco D, Schroter S, Aldcroft A, Moher D, Boutron I, Kirkham JJ, et al. Effect of an editorial intervention to improve the completeness of reporting of randomised trials: a randomised controlled trial. BMJ Open. 2020;10:e036799.

    Article  Google Scholar 

Download references


Not applicable.


No funding has been specifically awarded to the SEES Initiative. The project is housed at the Hospital de Clínicas de Porto Alegre (Porto Alegre, Brazil). NLO receives post-doctoral fellowship funding support from the National Institute of Science and Technology for Health Technology Assessment (IATS, Brazil). ATDN receives doctoral funding support from the Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES, Brazil – Finance Code 001). DU receives research productivity grant and support from the Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq foundation).

Author information

Authors and Affiliations



DU created the SEES Initiative. NLO and DU conceived the initial idea of this manuscript. All authors contributed to the development of the methodological protocol. DU and ATD developed the search strategy and search the database. NLO and CEB performed the eligibility and assessment of SRMAs. NLO and DU analyzed the data. NLO wrote the manuscript with strong feedback from all authors. All authors read and approved the final version of the manuscript.

Corresponding author

Correspondence to Nórton Luís Oliveira.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1.

PRISMA Checklist.

Additional file 2.

Search strategies for journals of exercise sciences and general medicine.

Additional file 3.

Full assessment of SRMA with individual items, associated component labels (i.e., domains) and, whether applicable, the guideline or tool of reference.

Additional file 4.

Exceptions and operationalization of the answers of some items.

Additional file 5: Table.

Number of Systematic Reviews with Meta-Analysis (SRMAs) assessed by journal.

Additional file 6.

Spreadsheet with complete assessments.

Additional file 7.

List of the studies in descending order of adherence to the items/recommended practices.

Additional file 8: Table.

Frequency distribution of items/recommended practices by domains for the Systematic Reviews with Meta-Analysis (SRMAs) with registration and without registration.

Additional file 9: Table.

Frequency distribution of items/recommended practices by domains for the Systematic Reviews with Meta-Analysis (SRMAs) with open access and without open access.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Oliveira, N.L., Botton, C.E., De Nardi, A.T. et al. Methodological quality and reporting standards in systematic reviews with meta-analysis of physical activity studies: a report from the Strengthening the Evidence in Exercise Sciences Initiative (SEES Initiative). Syst Rev 10, 304 (2021).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: