Skip to main content

A guide to evaluating systematic reviews for the busy clinicians or reluctant readers


Systematic reviews (SRs) provide a solution to handle information overload for busy clinicians by summarising and synthesizing studies on a specific issue. However, because SRs are complicated and often boring to read, the busy or reluctant reader may make do with the abstract. When, as it has been shown, many authors overstate efficacy or understate harm in their abstracts, not consulting the underlying article could be misleading. This means that the prudent reader must have the ability to identify the ‘tender points’ of SRs to avoid falling for ‘spin’. To this end we briefly review the method of SRs and ways to relatively quickly determine trustworthiness.


There has been an almost exponential growth in medical science-related journal publications [1], but scientific reports, in general, are not reader-friendly. For many clinicians the systematic review (SR) is, therefore, very helpful, as it summarizes the available evidence from the relevant literature.

Further, SRs are considered the most valuable type of research, placed at the top of the evidence pyramid [2], making them trusted and used to create guidelines and other policy documents. For the clinician, they represent a gold mine of information, all collected and reported in one place.

Unfortunately, though, anybody reading SRs will soon realise that all are not equally well performed. Therefore, results might have to be interpreted with caution [3]. This premise applies for all areas of health, including musculoskeletal conditions [4]. Also, consideration should be given to the possibility that an SR could be well performed but not well reported. Another problem is that SRs are often compact, technical, and quite boring to read, for which reason it is tempting for busy clinicians to read only the abstract, and they often do [5].

Unfortunately, abstracts of SRs have repeatedly been found to be deceptive, in that they may misrepresent their results by either exaggerating the good or diminishing the bad outcomes. This is called “spin” and has been shown to be particularly common for studies on low back pain with approximately 80% of abstracts being guilty of ‘spin’ [6]. Admittedly, this may not be life-threatening, but a study of SRs in emergency medicine showed a ‘spin’ percentage of approximately 30% [7], which is much more serious for patients.

This implies that the reader must be able to discern the trustable from the not so trustable SRs and be aware of the danger of reading the abstract only. To this end we will outline the key characteristics of a good quality SR and point out the ‘tender’ points. This will provide a relatively short and easy manual on SRs for the non-researcher, making it possible to relatively quickly assess, from a technical viewpoint, whether a SR is worth reading (and trusting) or not.

Key characteristics of a good quality systematic review

The major types of reviews

Reviews can be divided into three main groups: narrative, systematic, and other transparent types.

  1. a.

    Narrative reviews simply tell a story and may or may not have some transparent elements, but most commonly they consist of a summary of a topic, which can be both relevant and interesting. In fact, a topic may not be possible to deal with satisfactorily unless in a narrative form, as shown in this example which aimed to describe the outcomes, barriers, and facilitators relating to interprofessional practice involving chiropractors [8]. Unless the review is completely transparent, though, the reader cannot tell if what is written is also ‘true’ (including all relevant evidence, objective, and well-balanced).

  2. b.

    Systematic reviews, however, are inclusive and transparent. They can be purely descriptive or analytical and be with or without a critical element. However, increasingly the term ‘systematic’ seems to indicate that the review has also a critical view on the articles that were scrutinized. It is common that they deal with only one type of study design, but mixed design reviews are also seen, as in this example which reviews the impact of outcome measurements in clinical practice [9].

  3. c.

    Many other types of transparent reviews. There are many other types of semi-systematic or, at least, transparent reviews. These differ on how the literature is searched, the presence of quality appraisal, the approach to the analysis/synthesis, along with their presentation of results [10]. An example is the scoping review, which is increasingly often seen. It has a less rigorous approach to the search, may deal with several research designs, and lacks a critical element, because its purpose is to provide an overview over a topic without a stringent analysis, often because the topic is poorly understood and or confusing. A scoping review on Chiropractic Functional Neurology [11] is an example of how to approach a poorly defined clinical method that has only few scientific publications.

Systematic reviews are stringent research projects

A SR has many elements in common with an ‘ordinary’ research project, such as clear pre-hoc research questions, systematic collection of data, transparent analysis of data, and objective interpretation of the results [12]. Further, the data must be trustworthy, and this depends on two things: the quality of the data, i.e., the skills of the researchers of the original studies, and the quality of the review process, i.e., the skills of the reviewers. Many journals refuse to publish SRs if a detailed design or plan (protocol) has not been registered in a relevant research register, such as PROSPERO or INPLASY. Before registration, personnel at these registers will check the protocols, which constitutes a form of quality assurance.

The three steps of a systematic review

The method of data collection and analysis of SRs consist of three distinct steps:

  1. i)

    Finding the articles.

  2. ii)

    Extracting the data.

  3. iii)

    Analyzing/synthesizing the data.

Finding the articles

It is important to find and include all relevant literature [13]. This is best done by searching in the right places, using the right search terms, and aiming for a relevant time period. It may also be necessary to limit the search to certain foreign languages, depending on the language skills of the reviewers and assuming they do not trust Google translate. The research report must explain all this in such detail that somebody else can do the same search and find the same articles. Although anybody can look for articles on the Internet, to do it properly requires expertise to choose relevant databases and search terms. Therefore, the assistance of a research librarian is very helpful, perhaps even necessary.

The original search will result in a list of titles and abstracts of possibly suitable articles. Usually, this list is large, perhaps containing many thousand articles. The reviewers must therefore spend quite some time, meticulously selecting the relevant ones, according to pre-determined inclusion criteria, such as the relevant disease, the type of treatment and control treatment, depending on the topic and type of study design. Exclusion criteria are also defined in advance, such as not wanting to include animal studies. No articles are to be excluded at this stage because they are of poor quality, and certainly not because they provide the ‘wrong’ or ‘unexpected’ answers / outcomes. These extracted articles will correspond to the study subjects in a clinical study, and they will provide the data used to answer the research questions.

Extracting the data

Three types of data are usually extracted from these articles, listed in tables in an abbreviated form, to make the analysis perfectly transparent. These are: (i) descriptive information, which gives an overview of the reviewed articles on factors that may influence our understanding of the outcomes, (ii) information on various items that can be suspected of inducing bias or other quality information that may influence the credibility of data, and (iii) the variables (results) used to find answers to the research question(s).

All ‘evidence’ should be included in the report, i.e., the tables that (i) describe the articles, (ii) summarize the risk of bias for each article, and (iii) the raw data that provide the background for the results. Some articles scrutinize the technical quality instead of or in addition to the risk of bias. Sometimes these tables are presented in additional files or must be downloaded from the Internet.

Analyzing /synthesizing of data

The collected information can be analyzed (synthesized is another term) in several ways. The best-known analytic method is probably the meta-analysis, in which estimates from the articles are combined and re-analyzed with a specific statistical method to obtain a new summary estimate surrounded by a range of values that describes the uncertainty surrounding it (confidence interval). This confidence interval will typically be much smaller than the ones in the individual studies, as there will now be many more participants in the ‘meta-study’. The meta-analysis is often used to compare outcomes in two types of treatments, when there are several studies that approached the topic in similar ways. It can also be used to establish the prevalence of something in studies that define the study population and the outcome variable similarly.

Other analytical methods could be to count the number of studies that obtained one specific response vs. those finding something else. Such as, “eight of the ten included studies found no association between the habitual use of chewing gum and jaw pain”.

Results can also be presented as numbers, percentages, odds ratios, and suchlike to be summarized into a big picture.

Other analytical approaches could be to identify common concepts or approaches and to report this narratively. A narrative report of the results is not to be confused with a narrative review, which was explained under “Three major types of reviews”.

The critical aspect: checking if the extracted articles can be trusted

There is a strict procedure for how SRs should be performed. Part of this procedure deals with the trustworthiness of the results. This is pivotal and must be part of the analytical approach [14]. The credibility of the findings in an article depends on whether they are likely to be correctly obtained during the (original) research procedure (internal validity) and whether they are likely to be typical of the ‘real world’ (external validity). The risk of bias and other quality aspects of the methods used and the results reported are therefore scrutinized in properly conducted SRs.

Risk of bias

The skills of the original researchers in the extracted studies are thus assessed in a systematic manner, checking for points of (possible) bias that can arise e.g., when selecting study participants, treating these (if this is done), and summarizing the findings in the statistical analysis. ‘Bias’ is defined as an error that tends to push or distort results in a specific direction, i.e., resulting in a ‘systematic error’. Since it is not always possible to see if such an error occurs, one looks for the risk of it happening, i.e., “risk of bias”, referred to as ‘RoB’ in research articles. This is typically reported in tables (checklists, grids) with or without colors (green for low, yellow for moderate, and red for high RoB).

Different study designs require different research methods, which can result in different possibilities for bias. Therefore, there exist many different checklists for RoB assessment, each relevant for the various research designs (e.g., clinical trials, outcome, qualitative and animal experimentation studies) [15]. When no good checklist exists, or if there is one that needs some adaptations to meet the purposes of the study, the reviewer must design or amend an existing one in a transparent manner [16].

Quality issues

Errors that occur in a study are not necessarily ‘systematic’ (i.e., risking the influence of bias) but can also result in haphazard findings. One example is an inexperienced nurse responsible for taking the blood pressure, which results in nonsense (in any direction) blood pressure values. Another example could be a faulty questionnaire that lacks an appropriate answering category, with responders reacting in different, erratic ways to make it possible to provide an answer anyway. Checklists relating to quality items would probably have to be ‘invented’ by the research team, as they do not exist as prototypes. The quality of the study will also influence its credibility and its results, but this aspect is less often dealt with in SRs. Most authors concentrate on RoB.

Drawing conclusions on the RoB /quality

It is important, though, that these checklists (on RoB and possibly on quality) are carefully filled out and presented in the article for the readers to be able to see themselves if the included articles can be trusted or not. Thus, each article is defined as having a low, moderate, or high RoB, with the latter obviously being less trustworthy than the first. When the quality, rather than RoB, is under scrutiny, as would be the case in very technical studies, such as in laboratory studies, the same approach should be taken, defining studies as having low, moderate, or high quality. The cut-points, for when an item is ‘good’ or ‘bad’ must be pre-defined and explained in the Methods section. Their respective importance may differ with different study designs.

The whole reviewed research area can also be summarized in the same way, indicating if the evidence, in general, is trustable or not, using the concepts of RoB and/or quality. This is important, as poorly conducted research is more prone to bias and is therefore more likely to distort the true outcome of a study often resulting in ‘good’ outcomes, whereas well-conducted studies often have ‘poor’ outcomes, as was so dramatically demonstrated in an SR on spinal manipulation in the treatment of non-musculoskeletal disorders, where all the studies of low RoB revealed there to be no ‘effect’ of the treatment, whereas all the others reported good outcomes [17].

Taking into account the risk of bias and/or quality in the analysis

Whatever the method used to analyze/synthesize the data, it is important not to trust all results equally. The studies with a high RoB and/or of poor quality should not be considered equal to those done properly [18]. This means that one should discern whether or not the authors have either (i) separated the ‘good’ from the ‘bad’ articles and also reported the findings from these separately whilst indicating if the results are believable or not, (ii) or summarized/calculated the results whilst including all the articles – good and bad - and then labelled the research area (i.e., all the articles) as trustable, less trustable, or not trustable at all (the GRADE method). An example of the first approach is the SR on spinal manipulation and non-musculoskeletal conditions [17], whereas a SR on the potential effect of spinal manipulation on the autonomic nervous system used the GRADE approach [19].

Incredibly, many reviewers go through the tedious work of checking their collected articles for RoB (or possibly quality) and may even report the findings in a table to then forget all about it when drawing their conclusions on the results! This is particularly common in meta-analyses, where instead of checking what happens with the final estimate when the ‘bad’ studies have been removed or discussing the credibility of the whole research area, all studies are often pooled into one happy family, thus not taking into account at all the issue of trustworthiness. One could say that this omission is what separates a ‘research technician’ from a researcher, as ‘real’ researchers, as well as enlightened readers, are willing to accept and trust studies using ‘good’ methods and not only looking for the studies with the ‘good’ results. The inclusion of less robust/trustable study results risks leading to unreliable, often inflated estimates and can result in inaccurate recommendations influencing clinical practice.

Assessing the technical quality of a systematic review- checking for weaknesses at the ‘tender points’

A SR should have the same approach as in ‘ordinary’ research reports when interpreting results in the light of any methodological issues that could have influenced these. However, authors of SRs must deal with two layers of methodological issues: that of the others (i.e., the authors of the reviewed articles) and that of their own work. RoB and quality assessments take care of the first aspect, but it is important for the authors also to scrutinize the quality of their own review. Obviously, it is easier to find faults with others than with oneself, for which reason authors often seem to forget to take a critical look at their own activities. Therefore, it is often up to the reader to be critical because also SRs can have RoB and quality issues. Before reading the whole article (or perhaps ‘cheating’ by reading only the conclusion), it is, therefore, important to know if the SR was done ‘properly’, and some of this can be discerned by looking for some specific ‘tender points’. As a minimum, these should be acceptable. They are easy to detect also for the non-expert, and all readers should look for these.

Technical issues, which can result in both bias and quality issues in the review process itself, typically occur in four places, namely: (i) during the search for relevant articles, (ii) when extracting data, (iii) when analyzing the data, and (iv) in the interpretation of the results, such as producing a conclusion that can be described as “spin”.

  1. I.

    Searching for relevant articles: Thus, when searching for articles, a librarian does this best. The screening of articles should be performed by two persons, independently of each other, who then adjust any divergent findings afterwards in discussions. A third person, a referee, might be needed if the two reviewers cannot obtain consensus. If one person does this, as is often the case, this should either be done twice or with somebody checking the findings afterwards. The search terms should be listed somewhere in the text (or in an additional file) and the number of articles at each stage of the screening process should be presented in a flow-chart.

  2. II.

    Extracting data: Also this should be done by two people, independently of each other, as research texts can be complicated and may be misunderstood. Further, all extracted data should be visible in the results tables to make it possible for readers to check the information, theoretically making it possible to ensure there are no errors that could influence the results of the analyses. If the reader is unable to re-analyze the data based on the information in the various tables, the article is not fully transparent, and full transparency is the hallmark of a SR.

  3. III.

    Analysis of data: A major problem, seen quite often in SRs, is the disrespect for the issues of bias and/or quality during the analysis/synthesis of data. This is noticeable in three major ways: (i) there is no RoB assessment, although there quite clearly should be one, (ii) there could be a mention of a RoB assessment but no table that shows the results, neither in the main text nor in an additional file, or (iii) there is a RoB table but no explanation of how (if at all!) this had any consequences on the data synthesis and/or interpretation. It is difficult to understand why the reviewers went to all the trouble of extracting data on RoB/quality to then forget about it. The danger here is that readers might be so confused by the multitude of details that they lose the big picture, not realizing that RoB/quality played no role in the analysis or interpretation of data.

  4. IV.

    Interpretation of the results and risk of ‘spin’: Finally, the summary of findings at the beginning of the Discussion and in the Conclusions should correctly reflect the findings, which – in turn - should ultimately relate to the research questions posed at the end of the Background section. If these do not align, you know that you are dealing with an unskilled reviewer, and you have reasons to be cautious, as this mismatch may result in ‘spin’. A look at the affiliations of the authors of the review, including a search for any reported or unreported conflicts of interest, can also be helpful when watching for ‘spin’, as various types of obvious or less obvious conflicts of interest can influence biomedical research in important ways [20], also in SRs.

Additional help to the non-expert reader

As mentioned, there are checklists for most research designs both on how to conduct a study (mainly RoB tools) but also on how to write a report (writing guidelines are found on the Equator network). The same is true for SRs. Thus, for those who wish to penetrate deeper into this topic, there is a RoB checklist for SRs, called AMSTAR [21] and a writing guideline, the PRISMA guidelines [22]. It is usually a sign of reassurance when the SR contains a reference to this guideline.

Nevertheless, as previously mentioned, the technical quality of SRs is quite often not up to standard, which could mean that the results are inaccurate. For this reason, it is important for the reader to be able to quickly scrutinize such texts before investing reading time. For this reason, we include a list of ‘tender points’ to look for in SRs (Table 1). These can be detected by searching for these items in the text without spending time on actually reading it from the beginning to the end. Most of the relevant information to look for is found in the Methods section of the SR.

Table 1 Some common crucial points that readers should check to see if the systematic review is credible

A final consideration is one of the importance of the results [23]. That is, can the results be used in clinical practice?


It is important to note that even an impressive-looking systematic review can have limitations and that it’s important to resist the temptation of reading only the Conclusions. Instead, we recommend that readers follow our brief checklist, looking for the “tender points” that will indicate whether the results can be trusted or not. These tender points are particularly noticeable in the (i) article selection, (ii) data extraction, (iii) data analysis (which all may impact the trustworthiness of the reviewed literature), and in the (iv) authors’ interpretation of the results, which should not show any signs of ‘spin’.

Data Availability

Not applicable.



Risk of Bias


Systematic review


  1. Bornmann L, Mutz R. Growth rates of modern science: a bibliometric analysis based on the number of publications and cited references. J Association Inform Sci Technol. 2015;66(11):2215–22.

    Article  CAS  Google Scholar 

  2. Murad MH, Asi N, Alsawas M, Alahdab F. New evidence pyramid. BMJ Evidence-Based Medicine. 2016;21(4):125–7.

    Article  Google Scholar 

  3. Tian J, Zhang J, Ge L, Yang K, Song F. The methodological and reporting quality of systematic reviews from China and the USA are similar. J Clin Epidemiol. 2017;85:50–8.

    Article  PubMed  Google Scholar 

  4. Martinez-Calderon J, Flores-Cortes M, Morales-Asencio JM, Luque-Suarez A. Which psychological factors are involved in the onset and/or persistence of musculoskeletal pain? An umbrella review of systematic reviews and meta-analyses of prospective cohort studies. Clin J Pain. 2020;36(8):626–37.

    Article  PubMed  Google Scholar 

  5. King DW, Tenopir C, Clarke M. Measuring total readings of Journal Articles. D-Lib; 2006.

  6. Nascimento DP, Costa LO, Gonzalez GZ, Maher CG, Moseley AM. Abstracts of low back pain trials are poorly reported, contain spin of information, and are inconsistent with the full text: an overview study. Arch Phys Med Rehabil. 2019;100(10):1976–85. e1918.

    Article  Google Scholar 

  7. Ferrell MC, Schell J, Ottwell R, Arthur W, Bickford T, Gardner G, Goodrich W, Platts-Mills TF, Hartwell M, Sealey M. Evaluation of spin in the abstracts of emergency medicine systematic reviews and meta-analyses. Eur J Emerg Med. 2022;29(2):118–25.

    Article  Google Scholar 

  8. Myburgh C, Teglhus S, Engquist K, Vlachos E. Chiropractors in interprofessional practice settings: a narrative review exploring context, outcomes, barriers and facilitators. Chiropr Man Ther. 2022;30(1):1–11.

    Article  Google Scholar 

  9. Holmes MM, Lewith G, Newell D, Field J, Bishop FL. The impact of patient-reported outcome measures in clinical practice for pain: a systematic review. Qual Life Res. 2017;26:245–57.

    Article  PubMed  Google Scholar 

  10. Grant MJ, Booth A. A typology of reviews: an analysis of 14 review types and associated methodologies. Health Info Libr J. 2009;26(2):91–108.

    Article  PubMed  Google Scholar 

  11. Meyer A-L, Meyer A, Etherington S, Leboeuf-Yde C. Unravelling functional neurology: a scoping review of theories and clinical applications in a context of chiropractic manual therapy. Chiropr Man Ther. 2017;25(1):1–23.

    Article  Google Scholar 

  12. Hunt H, Pollock A, Campbell P, Estcourt L, Brunton G. An introduction to overviews of reviews: planning a relevant research question and objective for an overview. Syst reviews. 2018;7(1):1–9.

    Article  Google Scholar 

  13. Montori VM, Wilczynski NL, Morgan D, Haynes RB. Optimal search strategies for retrieving systematic reviews from Medline: analytical survey. BMJ. 2005;330(7482):68.

    Article  PubMed  PubMed Central  Google Scholar 

  14. Seehra J, Pandis N, Koletsi D, Fleming PS. Use of quality assessment tools in systematic reviews was varied and inconsistent. J Clin Epidemiol. 2016;69:179–184e175.

    Article  PubMed  Google Scholar 

  15. Whiting P, Wolff R, Mallett S, Simera I, Savovic J. A proposed framework for developing quality assessment tools. Syst Rev. 2017;6(1):204.

    Article  PubMed  PubMed Central  Google Scholar 

  16. Viswanathan M, Patnode CD, Berkman ND, Bass EB, Chang S, Hartling L, Murad MH, Treadwell JR, Kane RL. Recommendations for assessing the risk of bias in systematic reviews of health-care interventions. J Clin Epidemiol. 2018;97:26–34.

    Article  PubMed  Google Scholar 

  17. Côté P, Hartvigsen J, Axén I, Leboeuf-Yde C, Corso M, Shearer H, Wong J, Marchand A-A, Cassidy JD, French S. The global summit on the efficacy and effectiveness of spinal manipulative therapy for the prevention and treatment of non-musculoskeletal disorders: a systematic review of the literature. Chiropr Man Ther. 2021;29(1):1–23.

    Google Scholar 

  18. Campbell M, McKenzie JE, Sowden A, Katikireddi SV, Brennan SE, Ellis S, Hartmann-Boyce J, Ryan R, Shepperd S, Thomas J. Synthesis without meta-analysis (SWiM) in systematic reviews: reporting guideline. bmj 2020, 368.

  19. Picchiottino M, Leboeuf-Yde C, Gagey O, Hallman DM. The acute effects of joint manipulative techniques on markers of autonomic nervous system activity: a systematic review and meta-analysis of randomized sham-controlled trials. Chiropr Man Ther. 2019;27(1):1–21.

    Article  Google Scholar 

  20. Nejstgaard CH, Bero L, Hróbjartsson A, Jørgensen AW, Jørgensen KJ, Le M, Lundh A. Association between conflicts of interest and favourable recommendations in clinical guidelines, advisory committee reports, opinion pieces, and narrative reviews: systematic review. bmj 2020, 371.

  21. Shea BJ, Reeves BC, Wells G, Thuku M, Hamel C, Moran J, Moher D, Tugwell P, Welch V, Kristjansson E. AMSTAR 2: a critical appraisal tool for systematic reviews that include randomised or non-randomised studies of healthcare interventions, or both. bmj 2017, 358.

  22. Sarkis-Onofre R, Catalá-López F, Aromataris E, Lockwood C. How to properly use the PRISMA Statement. Syst Reviews. 2021;10(1):1–3.

    Article  Google Scholar 

  23. Evaluating systematic reviews and meta-analyses. In: Seminars in reproductive medicine: 2003: Copyright© 2003 by Thieme Medical Publishers, Inc., 333 Seventh Avenue, New … 5–106.

Download references


We would like to acknowledge the valuable feedback from Dr Jean Theroux PhD.


Not applicable.

Author information

Authors and Affiliations



SI was responsible for the study conceptualisation. SI and CL developed the initial draft. Both authors contributed to iterative drafts. Both authors read and approved the final manuscript.

Corresponding author

Correspondence to Stanley Innes.

Ethics declarations

Ethics approval and consent to participate

Not Applicable.

Consent for publication

Not applicable.

Competing interests


Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Innes, S., Leboeuf-Yde, C. A guide to evaluating systematic reviews for the busy clinicians or reluctant readers. Chiropr Man Therap 31, 38 (2023).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: