AMSTAR-2 Checklist
AMSTAR-2 Checklist
AMSTAR-2 Checklist
See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
ORIGINAL ARTICLE
(wileyonlinelibrary.com) doi: 10.1002/leap.1463 Received: 8 February 2022 | Accepted: 28 April 2022 | Published online in Wiley Online Library: 7 June 2022
1
Leeds Institute of Medical Research at St. James’s,
Abstract: AMSTAR-2 is a critical appraisal instrument for systematic re-
University of Leeds, Leeds, UK views and may have a role in editorial processes. This study explored
2
Division of General Surgery, Department of Surgery, whether associations exist between AMSTAR-2 assessments and editorial
University of Toronto, Toronto, Ontario, Canada
decisions. A retrospective, cross-sectional study of manuscripts submitted
3
Department of Surgery, Amsterdam UMC, University to a single journal between 2015 and 2017 was undertaken. All submis-
of Amsterdam (location AMC), Amsterdam,
The Netherlands
sions that reported an eligible systematic review were assessed using
4
AMSTAR-2 by two assessors. Inter-rater agreement (IRR) was calculated
Department of Surgery, Faculty of Medical and Health
Sciences, University of Auckland, Auckland, for all AMSTAR-2 items. Associations between AMSTAR-2 assessments
New Zealand and the editorial decision, final publication status in any journal, and mea-
5
Department of Gastrointestinal Surgery, Ghent sures of impact were explored. One hundred and twenty-two manuscripts
University Hospital, Ghent, Belgium
were included. Across all AMSTAR-2 items, the IRR varied from 0.03
6
Royal Devon & Exeter Hospital, Royal Devon & Exeter
(slight agreement) to 0.82 (substantial agreement). All submissions con-
NHS Foundation Trust, Exeter, UK
tained at least two critical methodological weaknesses. There was no dif-
ORCID: ference in the number of weaknesses (median: 4; IQR: 3–5 vs. median: 4;
S. J. Chapman: 0000-0003-2413-5690
IQR: 3.5–4.5; p = 0.482) between accepted and rejected submissions.
F. Dossa: 0000-0002-4670-7445
E. J. de Groof: 0000-0002-7191-4964 Neither was there a difference between rejected submissions published
C. Keane: 0000-0002-9611-2335 elsewhere and those which remained unpublished (median: 4; IQR:
G. H. van Ramshorst: 0000-0002-5368-582X
N. J. Smart: 0000-0002-3043-8324
3.5–4.5 vs. median: 4; IQR: 4.5–5; p = 0.103). The number of weaknesses
was not associated with academic impact. There was no association with
*Corresponding author: Stephen J. Chapman, NIHR
AMSTAR-2 assessments and editorial outcomes. Further work is required
Doctoral Research Fellow Room 7.16, Clinical Sciences
Building, Leeds Institute of Medical Research at to explore whether the instrument can be prospectively operationalized
St. James’s, University of Leeds, LS9 7TF, Leeds, UK. for use during editorial processes.
E-mail: [email protected]
Learned Publishing 2022; 35: 529–538 www.learned-publishing.org © 2022 The Authors. 529
Learned Publishing published by John Wiley & Sons Ltd on behalf of ALPSP.
This is an open access article under the terms of the Creative Commons Attribution License,
which permits use, distribution and reproduction in any medium, provided the original work is properly cited.
17414857, 2022, 4, Downloaded from https://onlinelibrary.wiley.com/doi/10.1002/leap.1463 by Nat Prov Indonesia, Wiley Online Library on [11/03/2023]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
530 S.J. Chapman et al.
1 Non-critical Did the research questions and inclusion criteria for the review include the components Yes
of PICO? No
2 Critical Did the report of the review contain an explicit statement that the review methods Yes
were established prior to the conduct of the review and did the report justify any Partial yes
significant deviations from the protocol? No
3 Non-critical Did the review authors explain their selection of the study designs for inclusion in the Yes
review? No
4 Critical Did the review authors use a comprehensive literature search strategy? Yes
Partial yes
No
5 Non-critical Did the review authors perform study selection in duplicate? Yes
No
6 Non-critical Did the review authors perform data extraction in duplicate? Yes
No
7 Critical Did the review authors provide a list of excluded studies and justify the exclusions? Yes
Partial yes
No
8 Non-critical Did the review authors describe the included studies in adequate detail? Yes
Partial yes
No
9 Critical Did the review authors use a satisfactory technique for assessing the risk of bias (RoB) Yes
in individual studies that were included in the review? Partial yes
No
10 Non-critical Did the review authors report on the sources of funding for the studies included in the Yes
review? No
11 Critical If meta-analysis was performed, did the review authors use appropriate methods for Yes
statistical combination of results? No
No MA conducted
12 Non-critical If meta-analysis was performed, did the review authors assess the potential impact of Yes
RoB in individual studies on the results of the meta-analysis or other evidence No
synthesis? No MA conducted
13 Critical Did the review authors account for RoB in individual studies when interpreting/ Yes
discussing the results of the review? No
14 Non-critical Did the review authors provide a satisfactory explanation for, and discussion of, any Yes
heterogeneity observed in the results of the review? No
15 Critical If they performed quantitative synthesis, did the review authors carry out an adequate Yes
investigation of publication bias (small study bias) and discuss its likely impact on the No
results of the review? No MA conducted
16 Non-critical Did the review authors report any potential sources of conflict of interest, including Yes
any funding they received for conducting the review? No
(high, moderate, low, or critically low) based on weaknesses identi- These were identified from an internal search of editorial records
fied across defined ‘critical’ and ‘non-critical’ domains. An outline of using the ScholarOne Manuscript editorial workflow management
the domains and criteria for decision-making is shown in Table 1. system. Simple literature/scoping reviews, diagnostic accuracy
reviews, reviews of reporting methods and studies reviewing epi-
demiological trends were excluded. No language exclusions were
Eligibility criteria
applicable since all submissions were received in English. All study
All original manuscripts submitted to Colorectal Disease between populations were eligible, including adults, pregnant adults, and
January 2015 and December 2017 which reported a systematic children and no exclusions were applied based on the clinical
review of a healthcare intervention were eligible for inclusion. subject area.
FIGURE 2 Scatter plots of critical weaknesses versus measures of research impact. (A–C) Scatter plots for submissions accepted in Colo-
rectal Disease after review, showing the correlation between number of critical weaknesses and (A) average annual citations according to
iCite; (B) relative citation ratio (RCR); (C) average annual citations according to Google Scholar. (D–F) scatter plots for submissions rejected
by Colorectal Disease and accepted by other journals, showing the correlation between number of critical weaknesses and (D) average
annual citations according to iCite; (E) relative citation ratio (RCR); (F) average annual citations according to Google Scholar.
approaches, including the number of forward citations according correlation coefficient. Inter-rater agreement (IRA) was explored
to the iCite portal (representing impact from MEDLINE-indexed using Cohen’s kappa, with the following parameters indicating the
sources), number of citations according to Google Scholar (rep- level of agreement: ≤0 no agreement; 0.01–0.20 slight agreement;
resenting impact from any published and online source) and the 0.21–0.40 fair agreement; 0.41–0.60 moderate agreement;
relative citation ratio (RCR; a field normalized measure of impact). 0.61–0.80 substantial agreement; and 0.81–1.00 almost perfect
agreement. Across all statistical tests, p < 0.05 was considered to
indicate statistical significance.
Data collection
Characteristic data including year of publication, colorectal
subject area (malignancy, inflammatory bowel disease, abdominal RESULTS
wall, other benign, peri-operative care), study design (systematic
review; systematic review with meta-analysis), and type of
Study characteristics
included evidence (randomized, non-randomized, both) were
collected by a single reviewer. Variables which were considered A total of 215 submissions were reviewed for eligibility and
to be wholly or partly subjective (such as colorectal subject area) 122 met all criteria for inclusion. Common reasons for exclusion
were checked by a second assessor, with disagreement addressed were submissions reporting a diagnostic accuracy review
through discussion and consensus. (n = 38), literature/scoping reviews (n = 27), and reviews of
epidemiological trends (n = 20) (Fig. 1). Submissions were most
often submitted by authors from China (n = 46; n = 37.7%) and
Statistical analysis
the UK (n = 28; 23.0%). A total of 95 (77.9%) included a pooled
A simple descriptive analysis was performed, including a summary meta-analysis. The most common subject areas were malignancy
of rates, averages (medians with interquartile ranges, IQR), and (n = 46; 37.7%), inflammatory bowel disease (n = 15; 12.3%) and
proportions. Associations between AMSTAR-2 summary outcomes other benign conditions (n = 30; 24.6%). A full outline of charac-
and editorial decisions were explored using Spearman’s Rho teristics is provided in Table 2.
Critical Items
Non-critical items
Note: ‘Partial Yes’ and ‘Yes’ responses for each item were considered to represent a satisfactory response (i.e. absence of the respective
weakness), as per the Method.
a
Cohen’s k with 95% confidence intervals.
b
A total of 27 manuscripts were not considered in these analyses since no formal meta-analysis was performed.
Summary of AMSTAR-2 assessments prospective methods) and item 7 (description of excluded studies),
while common sources of non-critical weaknesses were item
All 122 submissions demonstrated at least two critical weak-
3 ( justification of study design) and item 10 (description of study
nesses, indicating a ‘critically low’ level of confidence according
funding) (Table 3).
to the AMSTAR-2 instrument (Table 1). Overall, the median num-
ber of critical weaknesses was 4 (IQR: 3.5–4.875). When reviews
with and without meta-analysis were sub-analysed, the median
AMSTAR-2 inter-rater agreement
number of critical weaknesses was similar (median: 3.5, IQR 4–5;
and median: 4, IQR: 3–4.5, respectively) (p = 0.1556). The median The degree of IRA was variable across AMSTAR-2 items. Three
number of non-critical weaknesses was 4.5 (IQR: 3.5–5), which out of seven critical domains showed substantial agreement
was also similar when sub-analysed (median: 4.5; IQR: 3.5–5; and between assessors (items 2, 9 and 15), with one showing moder-
median: 4; IQR: 3.25–5, respectively) (p = 0.912). The most com- ate agreement (item 4) and the rest showing fair agreement.
mon sources of critical weakness were item 2 (statement of Across non-critical domains, one item showed near perfect
FIGURE 3 Scatter plots of non-critical weaknesses versus measures of research impact. (A–C) Scatter plots for submissions accepted in
Colorectal Disease after review, showing the correlation between number of non-critical weaknesses and (A) average annual citations
according to iCite; (B) relative citation ratio (RCR); (C) average annual citations according to Google Scholar. (D–F) scatter plots for submis-
sions rejected by Colorectal Disease and later accepted by other journals, showing the correlation between number of non-critical weak-
nesses and (D) average annual citations according to iCite; (E) relative citation ratio (RCR); (F) average annual citations according to
Google Scholar.
Average citations per year (Google Scholar) R = 0.112 p = 0.595 R= 0.104 p = 0.395
Average citations per year (Google Scholar) R = 0.225 p = 0.280 R = 0.003 p = 0.980
Note: All statistics are calculated using Spearman’s Rho correlation coefficient.
Abbreviation: RCR, relative citation ratio.
agreement (item 5), two showed substantial agreement (items: between the number of critical weaknesses and the average num-
8 and 10), three showed moderate agreement (items: 6, 12 and ber of citations per year via iCite (Colorectal Disease: R = 0.044;
14), and the rest showed only fair or slight agreement. A full out- Other journals: R = 0.056), the number of citations per year via
line of IRA calculations is provided in Table 3. Despite achieving Google Scholar (Colorectal Disease: R = 0.101; Other journals:
substantial agreement, the assessment of item 8 (description of R = 0.063), and the RCR (Colorectal Disease: R = 0.112; Other
included studies) was considered to be highly challenging by asses- journals: R = 0.104) (Fig. 2; Table 4). Similarly, the number of
sors. A post-hoc exercise was undertaken by assessors to review, non-critical weaknesses was not associated with academic
qualitatively identify and reflect on the perceived challenges impact, with all measures demonstrating weak or absent correla-
(Method reported fully in Data S1). Key reflections were: uncer- tion (Fig. 3; Table 4).
tainty about the level of detail required to satisfy the AMSTAR-2
criteria; disagreement on the definition of terms used by the
guidance documents; and difficulty applying the assessment
criteria to manuscripts due to incomplete or inconsistent DISCUSSION
reporting.
This study explored whether associations exist between
AMSTAR-2 assessments and editorial decision-making for manu-
AMSTAR-2 and editorial decisions scripts that report a systematic review of healthcare interven-
Twenty-five out of 122 (20.5%) submissions were accepted for tions. In this retrospective sample of manuscripts (performed
publication in Colorectal Disease. There was no difference in the after the completion of all editorial processes), all were found to
number of critical weaknesses (median: 4; IQR: 3–5 vs. median: contain multiple critical weaknesses, indicating a ‘critically low’
4; IQR: 3.5–4.5; p = 0.482) identified between accepted and level of confidence according to AMSTAR-2. Of note, there was
rejected submissions, respectively. Similarly, there was no differ- no difference in the number of critical weaknesses between man-
ence in the number of non-critical weaknesses (median: 5; IQR: uscripts accepted during the index submission and those which
3–5 vs. median: 4; IQR: 3.5–5; p = 0.787). Of the remaining were rejected, as well as no difference between manuscripts
97 rejected submissions, 69 (71.1%) were identified as being which were subsequently published in other journals and those
published in other MEDLINE-indexed journals. There were no which remained unpublished within the limits of this review.
differences in the number of critical weaknesses (median: 4; IQR: The number of critical weaknesses was not associated with the
3.5–4.5 vs. median: 4.5; IQR: 4–5; p = 0.103) or non-critical degree of subsequent academic impact, with some manuscripts
weaknesses (median: 4; IQR: 3–5 vs. median: 4.5; IQR: 3.875–5.5; of ‘critically low’ confidence achieving large volumes of citations.
p = 0.165) between submissions published in other journals and The mechanisms and determinants of editorial decision-
those where a publication was not identified, respectively. making continue to be the subject of scrutiny and debate. Some
studies have interrogated this by exploring predictors of success-
ful publication. A cross-sectional study of 1107 submitted manu-
AMSTAR-2 and research impact
scripts to three major medical journals (BMJ, the Lancet, Annals of
A total of 72 out of 94 (76.6%) submissions were published Internal Medicine) in 2003 found that acceptance was more
(including 25 in Colorectal Disease and 69 in other journals follow- likely if studies demonstrated high methodological quality, if they
ing the initial index rejection). The number of critical weaknesses had a randomized design, or if the corresponding author lived
identified in published submissions was not associated with sub- in the same country as the publishing journal, amongst other
sequent academic impact. Weak correlations were identified factors (Lee et al., 2006). In another study of 112 consecutive
meta-analyses submitted to JAMA between 1996 and 1997, a on study assessments. This was essential to eliminate avoidable
single factor—replicability of results—was associated with an between-assessor variation. Another strength was the sampling
increased rate of acceptance, whereas the overall methodological of both accepted and rejected manuscripts submitted to a single
quality was not (Stroup et al., 2001). Some studies have explored journal, which enabled a cross-sectional assessment of submis-
the impact of peer review comments on the final decision to sions which closely reflected a typical editorial pathway. Limita-
publish, with one report identifying comments on study design, tions of the study are also recognized. First, the study setting
originality, and the relationship between design, results, and con- involved submissions from a single sub-specialty journal, which is
clusions to be the most impactful factor on final decision-making a challenge for wider generalisability. Reproduction of the study
(Turcotte et al., 2004). The approach to understanding and facili- is encouraged to explore similarities and differences across differ-
tating peer review continues to be an evolving and complex ent journals and editorial processes. Secondly, it is notable that
domain of scholarly research with little clear empirical evidence many of the sampled manuscripts (submitted between 2015 and
to guide its use (Tennant & Ross-Hellauer, 2020). 2017) preceded the widespread publication of AMSTAR-2 in
The AMSTAR-2 critical appraisal tool published in 2017 pro- 2017. This was necessary to enable an assessment of forward
vides a standardized approach to assess the methodological qual- publication and academic impact in the years which followed
ity of systematic reviews that include a healthcare intervention. publication. In principle, this is not considered to be a major
Several reports have explored its usability and validity when weakness, since the checklist reflects general constructs of meth-
applied to broad samples of systematic reviews. In one report odological rigour. Furthermore, it may be considered a strength
involving 60 systematic reviews of treatments for depression, since the selection of this timeframe also pre-dates other major
AMSTAR-2 was found to have ‘moderate’ agreement between confounding changes in the publishing landscape, such as updates
four assessors (overall kappa: 0.42; 95% CI: 0.25–0.59) (Lorenz to the PRISMA checklist and revision of the Cochrane Handbook.
et al., 2019). In another report of 30 reviews relevant to anaes- Future research may consider whether the instrument can be
thesiology, the agreement between four assessors was ‘fair’ prospectively operationalized within the editorial process. This
(overall kappa: 0.3; 95% CI: 0.17 to 0.43), with the poorest agree- may involve a prospective study, with dedicated training provided
ment demonstrated for item 8 (description of included studies) to administrators, reviewers or editors. Alternatively, the feasibil-
(kappa: 0.09; 95% CI: 0.06 to 0.23 (Pieper et al., 2019). Of note, ity of other approaches to guide editorial decision-making may be
a recent report exploring the interpretation of AMSTAR-2 found of interest. Such approaches may involve alternative instruments,
that significant differences existed in the final assessment of such as the ROBIS tool which assesses the risk of bias in system-
confidence according to how the criteria were applied, leading atic reviews or a different approach altogether such as mandatory
to calls for greater transparency and clearer reporting (Pieper pre-registration and appraisal of protocols (Whiting et al., 2016).
et al., 2021). While most studies have focused on the use of Either way, due to the role of systematic review research in guid-
AMSTAR-2 from an author’s perspective, to our knowledge, there ing health policy, it is clear that more attention is required to
is little or no evidence describing its use as a tool within the refine peer review and editorial processes for systematic reviews.
editorial pathway. This is important to ensure that only the best evidence is selected
In this study, while the IRA for some items was substantial for publication and subsequent adoption in practice.
(critical items: 2, 9 and 15; non-critical items: 8 and 10), others
showed only moderate or fair agreement. The methodological
quality of all submissions was considered to be ‘critically low’, AUTHOR CONTRIBUTIONS
which is in line with previous reports demonstrating a large
Neil J. Smart and Gabrielle H. van Ramshorst conceptualized the
majority of systematic review reports with critical methodological
study. Stephen J. Chapman, Fahima Dossa, E. Joline de Groof and
weaknesses (Siemens et al., 2021). Of note, the number of critical
Celia Keane performed searches, study assessments and data
weaknesses identified in manuscripts that were accepted and
extraction. Stephen J. Chapman prepared the manuscript which
rejected during their index submission was not significantly differ-
was subsequently edited by all authors. Neil J. Smart is the study
ent. This may suggest that factors as well as methodological qual-
guarantor.
ity (such as topicality, community expectations and perceived
likelihood of citations) may be co-determinants of the final edito-
rial decision. It was not possible to explore these factors in the
CONFLICT OF INTEREST
present study, but they are important targets for future investiga-
tion. The completeness of reporting according to accepted frame- Neil J. Smart and Gabrielle H. van Ramshorst are Editor-in-Chief
works (such as PRISMA) is also important and closely related and Editor of Colorectal Disease, respectively. Stephen
since critical appraisal relies on the availability of transparent J. Chapman, Fahima Dossa, E. Joline de Groof and Celia Keane
information. The completeness of reporting was not explored in are early career members of the Colorectal Disease Editorial
the present study, but this may be an important consideration for Advisory Board. The views expressed in this publication are those
future work. of the authors and not necessarily those of the NHS, the National
A major strength of the present study was the process of Institute for Health and Care Research, Health Education England
familiarization and piloting amongst assessors prior to embarking or the Department of Health.
TRANSPARENCY Pieper, D., Koensgen, N., Breuing, J., Ge, L., & Wegewitz, U. (2018).
How is AMSTAR applied by authors - a call for better reporting.
The corresponding author (Stephen J. Chapman) attests that all BMC Medical Research Methodology, 18, 56. https://doi.org/10.
authors had access to study data, takes responsibility for the 1186/s12874-018-0520-z
accuracy of the analysis and had authority over manuscript prep- Pieper, D., Lorenz, R. C., Rombey, T., Jacobs, A., Rissling, O.,
aration and the decision to submit the manuscript for publication. Freitag, S., & Matthias, K. (2021). Authors should clearly report
how they derived the overall rating when applying AMSTAR 2-a
cross-sectional study. Journal of Clinical Epidemiology, 129,
DATA AVAILABILITY STATEMENT 97–103. https://doi.org/10.1016/j.jclinepi.2020.09.046
Coded data will be available from the authors on reasonable Pieper, D., Puljak, L., González-Lorenzo, M., & Minozzi, S. (2019).
request. Minor differences were found between AMSTAR 2 and ROBIS
in the assessment of systematic reviews including both
randomized and nonrandomized studies. Journal of Clinical
SUPPORTING INFORMATION Epidemiology, 108, 26–33. https://doi.org/10.1016/j.jclinepi.
2018.12.004
Additional supporting information may be found online in the
Royle, P., Kandala, N.-B., Barnard, K., & Waugh, N. (2013).
Supporting Information section at the end of the article:
Bibliometrics of systematic reviews: Analysis of citation rates and
Appendix S1. Supporting information. journal impact factors. Systematic Reviews, 12, 74.
Shea, B. J., Grimshaw, J. M., Wells, G. A., Boers, M., Andersson, N.,
Hamel, C., Porter, A. C., Tugwell, P., Moher, D., & Bouter, L. M.
REFERENCES (2007). Development of AMSTAR: A measurement tool to assess
Goldkuhle, M., Narayan, V. M., Weigl, A., Dahm, P., & Skoetz, N. the methodological quality of systematic reviews. BMC Medical
(2018). A systematic assessment of Cochrane reviews and sys- Research Methodology, 15, 10.
tematic reviews published in high-impact medical journals related Shea, B. J., Reeves, B. C., Wells, G., Thuku, M., Hamel, C., Moran, J.,
to cancer. BMJ Open, 8, e020869. https://doi.org/10.1136/ Moher, D., Tugwell, P., Welch, V., Kristjansson, E., & Henry, D. A.
bmjopen-2017-020869 (2017). AMSTAR 2: A critical appraisal tool for systematic reviews
Higgins, J. P. T., Thomas, J., Chandler, J., Cumpston, M., Li, T., that include randomised or non-randomised studies of healthcare
Page, M. J., & Welch, V. A. (Eds.). (2021). Cochrane handbook for interventions, or both. BMJ, 358, j4008.
systematic reviews of interventions version 6.2 (p. 2021. Available Siemens, W., Schwarzer, G., Rohe, M. S., Buroh, S., Meerpohl, J. J., &
from). Cochrane www.training.cochrane.org/handbook Becker, G. (2021). Methodological quality was critically low in
Katsura, M., Kuriyama, A., Tada, M., Tsujimoto, Y., Luo, Y., 9/10 systematic reviews in advanced cancer patients-a methodo-
Yamamoto, K., So, R., Aga, M., Matsushima, K., Fukuma, S., & logical study. Journal of Clinical Epidemiology, 136, 84–95. https://
Furukawa, T. A. (2021). High variability in results and methodologi- doi.org/10.1016/j.jclinepi.2021.03.010
cal quality among overlapping systematic reviews on the same Stroup, D. F., Thacker, S. B., Olsen, C. M., et al. (2001). Characteristics
topics in surgery: A meta-epidemiological study. The British Journal of meta-analyses related to acceptance for publication in a medi-
of Surgery, 108, 1521–1529. https://doi.org/10.1093/bjs/znab328 cal journal. Journal of Clinical Epidemiology, 54, 655–660. https://
Lee, K. P., Boyd, E. A., Holroyd-Leduc, J. M., Bacchetti, P., & doi.org/10.1016/S0895-4356(00)00362-0
Bero, L. A. (2006). Predictors of publication: Characteristics of Tennant, J. P., & Ross-Hellauer, T. (2020). The limitations to our
submitted manuscripts associated with acceptance at major bio- understanding of peer review. Research Integrity and Peer Review,
medical journals. The Medical Journal of Australia, 184, 621–626. 5, 6. https://doi.org/10.1186/s41073-020-00092-1
https://doi.org/10.5694/j.1326-5377.2006.tb00418.x Turcotte, C., Drolet, P., & Girard, M. (2004). Study design, originality
Lorenz, R. C., Matthias, K., Pieper, D., Wegewitz, U., Morche, J., and overall consistency influence acceptance or rejection of man-
Nocon, M., Rissling, O., Schirm, J., & Jacobs, A. (2019). A psycho- uscripts submitted to the journal. Canadian Journal of Anaesthesia,
metric study found AMSTAR 2 to be a valid and moderately reli- 51, 549–556. https://doi.org/10.1007/BF03018396
able appraisal tool. Journal of Clinical Epidemiology, 114, 133–140. Whiting, P., Savovic, J., Higgins, J. P. T., Caldwell, D. M., Reeves, B. C.,
https://doi.org/10.1016/j.jclinepi.2019.05.028 Shea, B., Davies, P., Kleijnen, J., Churchill, R., & ROBIS group.
Murade, M. H., & Wang, Z. (2017). Guidelines for reporting meta- (2016). ROBIS: A new tool to assess risk of bias in systematic
epidemiological methodology research. Evidence-Based Medicine, reviews was developed. Journal of Clinical Epidemiology, 69,
22, 139–142. https://doi.org/10.1136/ebmed-2017-110713 225–234. https://doi.org/10.1016/j.jclinepi.2015.06.005