Keywords
presentation of findings, evidence summaries, summary of findings table, communication, mixed-methods systematic review
presentation of findings, evidence summaries, summary of findings table, communication, mixed-methods systematic review
Clinical guidelines support decision making to improve patient outcomes and quality of care in a cost-effective manner1. The development of a clinical guideline involves a rigorous synthesis of the best available evidence on a specific clinical topic. It may involve formal consensus methods with a range of multidisciplinary stakeholders2–5. Guideline development groups comprise a range of decision makers, often including healthcare professionals, methodologists, health policymakers, clinicians, and patient representatives – all of whom have varying levels of expertise in evidence synthesis methods. This complicates the consensus process as stakeholders may prioritise and understand the findings of evidence syntheses, such as systematic reviews, differently6.
While the methods and recognition of the importance of systematic reviews have advanced in recent decades7,8, there are still barriers to their creation and use9,10. A meta-analysis of nearly 200 systematic reviews registered on the International Prospective Register of Systematic Reviews (PROSPERO) registry found that the average systematic review, from registration to publication date, takes 67.3 weeks, involves an average of five authors, and requires the full-text screening of 63 papers (range: 0–4385)9. In order for a review to be conducted well, authors must have the time to review the evidence, knowledge to critically appraise study methods, and skills to synthesise the results. Time and resource constraints are becoming an increasing concern as the number of academic papers and systematic reviews being published in recent decades has rapidly increased11–13. This growth accelerated during the recent COVID-19 (coronavirus disease) pandemic, causing an unprecedented ‘infodemic’14. The expansion of evidence, and in decision makers openness to accepting trade-offs in validity15, has resulted in the growing popularity of other evidence synthesis methods, such as rapid reviews16. This increase in different types of evidence synthesis methods further complicates matters for guideline development groups, who may interpret different types of systematic reviews in different ways based on how familiar they might be with particular approaches.
For those using different types of evidence synthesis to inform clinical guideline development and health policy, the amount of included studies, length, and technical nature of evidence syntheses can make it difficult to find answers about the effectiveness of healthcare interventions10,17. Previous work has highlighted that decision makers more easily understand evidence summaries than complete systematic reviews18,19. These summaries can come in a variety of different formats such as policy briefs, one-page reports, abstracts, summary of findings tables, plain language summaries, visual abstracts or infographics, podcasts, and more. While formatting may vary, decision-makers have expressed several key preferences, such as succinct summaries highlighting contextual factors like local applicability and costs17,20.
Succinctness should be inherent in an evidence summary, but how this distilled information is formatted and presented affects the interpretation and use of systematic reviews21. It is currently unclear which evidence summary format is most helpful for decision making for different guideline development group stakeholders. For example, Cochrane recommends a ‘summary of findings’ table7 but testing with users familiar with the Cochrane library and evidence-based practices raised concerns around comprehension and presentation of results and the balance between precision and simplicity22. Others have tested the presentation of information using different formats such as an abstract, plain-language summary, podcast or podcast transcription with no clear answer regarding which format was most suited to which stakeholder and resulted in the best understanding23. Similarly, infographics, plain-language summaries, and traditional abstracts were found to be equally effective in transmitting knowledge to healthcare providers; however, there were differences in measures of acceptability (i.e., user-friendliness and reading experience)24.
To better support clinical guideline development groups and decision-makers, it is important to identify which format works best for which stakeholder. Previous reviews have focused on identifying barriers and facilitators to use, or have been solely based on summary of findings tables10,25. As impacts on decision-making and preferences for formats may be evaluated through different study designs, a comprehensive synthesis of the evidence is needed beyond a typical single method systematic review. Mixed methods systematic reviews (MMSR) can more easily identify discrepancies within available evidence, pinpoint how quantitative or qualitative research has focused on particular interest areas, and offer a deeper understanding of findings26. A MMSR is especially useful for this project as it brings together findings of effectiveness and experience so findings are more useful for decision makers27. Guideline developers need to consider diverse considerations in their work such as feasibility, priority, cost effectiveness, equity, acceptability, and patient values and preferences28,29. Similarly, a MMSR allows us to consider and integrate data from a variety of different questions and synthesize information in a single project.
The aim of this mixed methods systematic review is to evaluate the effectiveness of, preferences for, and attitudes towards, different communication formats of evidence summary findings amongst guideline development group members, including healthcare providers, policy makers and patient representatives. To achieve this, the proposed MMSR will answer the following questions:
The proposed systematic review will be conducted in accordance with the Joanna Briggs Institute (JBI) Manual for Evidence Synthesis which details the methodology for mixed methods systematic reviews (MMSR)26.
As this is a MMSR, we will include quantitative (i.e., randomised controlled trials), qualitative, and mixed methods studies evaluating the effectiveness and/or preferences for and attitudes towards evidence summary formats. We will exclude conference abstracts, case reports, case series, editorials, and letters. Further details regarding eligibility criteria are given within the review-relevant sections below.
We are interested in studies involving stakeholders such as policy makers, healthcare providers, and health systems managers, as well as other GDG members such as clinicians, patient representatives, and methodologists such as systematic review authors. We will exclude studies where the sole participants are students, the general population (not involved in the decision-making process), and journalists as communication to these populations is more complex given a wide variety of confounding factors. We will also exclude studies related to clinical decision-making for individual patients.
We have followed the Population, Intervention, Comparison, Outcome (PICO) format for the quantitative review (Table 1) and the Sample, Phenomenon of Interest, Design, Evaluation, Research type (SPiDER) format for the qualitative review (Table 2) and will present unique aspects of each methodological approach within the relevant sections below.
Quantitative systematic review. Due to the complexity of stakeholders, evidence synthesis types, and summary formats, there is a high potential that confounding factors will be extensive. Relatedly, randomised controlled trials (RCTs) are the most appropriate design to evaluate the effectiveness of the interventions in question. Thus, we chose to restrict to RCTs only in order to focus on the performance and impact of summary formats in optimal settings. We will include studies where the intervention is any summary mode (e.g., visual, audio, text-based) which communicates the findings from an evidence synthesis study (e.g., systematic review, qualitative evidence synthesis, rapid review) to policy-makers and decision makers, including guideline development groups (GDGs). We anticipate that included summary formats will encompass visual abstracts, Summary of Findings tables, one-page summaries, podcasts, Graphical Overview of Evidence Reviews (GofER) diagrams, and others. Studies in which the summaries are one component of a multi-component intervention will be excluded, as will decision aids for direct patient care.
For studies examining the effectiveness of evidence summary formats, we will include any comparison to an alternative active comparator. Studies where the comparison is no intervention (e.g., the plain full-text of a manuscript) will be excluded, though we do not anticipate finding evidence syntheses with no form of summary or abstract.
Our primary outcomes of interest are:
1. Effectiveness
a. User understanding and knowledge, and/or beliefs in key findings of evidence synthesis (e.g., changes in knowledge scores about the topic included in the summary)
b. Self-reported impact on decision‐making
c. Intervention metrics (e.g., the time needed to read the summary, expressed language accessibility issues or scale scores)
We will not be including outcomes related to health literacy, numeracy, nor risk communication in patient-centred care.
Qualitative evidence synthesis. Primary studies investigating the understanding and acceptability of evidence summary formats will include qualitative studies (e.g, interviews or focus groups). Mixed-methods studies with primary qualitative data collection will be included if they meet the inclusion criteria for a randomised controlled trial and where it is possible to extract the findings derived from the qualitative research.
Our primary outcomes of interest relate to participant’s views and experiences with summary formats. This includes their perceptions of the impact of summary formats on their understanding, knowledge, and decision making, and participant’s beliefs, attitudes, and feelings towards usability and readability.
The following databases will be searched from inception to May 2021: Ovid MEDLINE, EMBASE, APA PsycINFO, CINAHL (Cumulative Index to Nursing and Allied Health Literature), Web of Science, and Cochrane Library. The search strategy for Ovid MEDLINE includes a combination of keywords and medical subject headings (MeSH) terms for GDG members, evidence syntheses, and formats for the communication of findings (see Table 3). As we are looking for primary research on the impacts or effects of interventions and attitudes towards them, we do not anticipate that this literature will be found in grey literature sources such as government or agency websites. Additionally, it is anticipated that controlled trials will have short time points of assessment (and follow-up) thus we do not believe that searching registries will benefit our study. This search strategy has been informed by the strategies of similar reviews in the same topic area10,25. Aligned with the Peer Review of Electronic Search Straggles (PRESS) Statement30, we engaged a medical librarian after the MEDLINE search was drafted but before it was translated to the other databases. As we are including a range of study designs, we did not apply study design specific filters. Although we have used a PICO and SPiDER approach for the quantitative and qualitative reviews we used the PICO format to inform the search strategy as previous researchers found that the SPiDER approach for search strategies may be too restrictive and specific31. Language and date restrictions will not be applied.
Backwards citation identification on all eligible studies will be performed using the citationchaser Shiny application built in R version 1.432. This application performs backwards citation screening (reviewing reference lists) and internally de-duplicates results. Each step of the search is summarised for transparency and references are given as a downloadable RIS file.
All citations will be downloaded and stored in Zotero reference manager version 5.0. For ease, rather than using Zotero for screening, title and abstract screening will be managed using Covidence. Two reviewers will independently screen titles and abstracts for inclusion criteria. Disagreements for inclusion will be resolved through discussions. If it is still unclear if the paper should be included, both authors will review the full version of the paper and discuss it again. If there is still disagreement, a third review author will be consulted. The screening process will be documented in the final manuscript using the Preferred Reporting Items for Systematic Reviews (PRISMA) flow diagram33 and a supplemental file detailing the reason for exclusion for each individual study will be made publicly available.
Two review authors will independently extract data from each of the included studies using a standardised data-extraction form. If there are disagreements or discrepancies, the two authors will discuss and consult with a third review author if needed. Where possible, qualitative outcomes such as themes and categories will be extracted into the standardized form, however, included articles containing qualitative methods will be also imported in NVivo12 for line-by-line coding for information related to outcomes. This separate but parallel data extraction is important for our analytical approach of the qualitative data which is discussed in greater detail in the Qualitative Analysis section. The following data will be extracted:
Bibliometric data (first author, title, journal, year of publication, language)
Study characteristics (setting, participants demographics, country, study design, intervention, comparators, theoretical framework, analytical approach)
Intervention characteristics will be collected following the structure of the Template for Intervention Description and Replication (TiDieR) checklist34 to provide detailed information on the why, what, who, how, where, and when of the intervention described.
Outcomes (quantitative estimates of effectiveness and acceptability; qualitative expressions of views, attitudes, opinions, experiences, perceptions, beliefs, feelings, and understanding)
Data to critically appraise included studies (JBI critical appraisal tools)
Funding sources
If information is missing from the study report, we will contact authors to inquire about these gaps. We will provide narrative syntheses in lieu of imputing missing data.
The JBI critical appraisal checklists will be used for randomized controlled trials and qualitative research. As data from separate quantitative and qualitative evidence is integrated in a MMSR, the JBI does not recommend the GRADE nor ConQual approaches for assessing overall confidence in synthesized findings26. Two review authors will independently complete the critical appraisal checklist for each included study. Differences will be resolved through discussion and consultation with a third review author if necessary.
If quantitative data allows for a meta-analysis, a forest plot will be generated using R. If we find a low number of studies, large treatment effects, few events per trial, or all trials are of similar sizes, we will use the Harboard test for publication bias35 as it reduces the false positive rate. Egger’s test36 for funnel plot asymmetry will be used to investigate small study effects and publication bias.
A narrative synthesis will be performed, however, if appropriate, quantitative data from randomised control trials will be synthesised using meta-analysis. Heterogeneity will first be explored by assessing the study characteristics that may vary across the included studies (for example, participant group, study design, risk of bias, interventions or outcomes). If sufficient data is available, subgroup analyses (e.g. participant groups such as medical professionals versus policy makers or intervention type such as visual abstracts versus plain abstracts) will be conducted. Furthermore, statistical heterogeneity will be explored according to statistical guidance on heterogeneity7, an estimated I2 of 50-90% represents substantial heterogeneity. We will weigh this against an χ2 test for heterogeneity (<.10). If our results indicate 50% or greater and a low χ2 statistic, this indicates that the heterogeneity may not be due to chance, thus we will not pool results into a meta-analysis. If data can be pooled, effect sizes and accompanying 95% confidence intervals will be reported as either relative risks (for dichotomous and dichotomised ordinal data) or standardized mean differences (for continuous data).
Where possible, qualitative findings will be pooled together using the meta-aggregation approach, which allows a reviewer to present findings of included studies as originally intended by the original authors37. This approach organises and categorises findings based on similarity in meaning and avoids re-interpretation. Therefore, it does not violate paradigms and approaches used by the original study authors. This approach also enables meaningful generalizable recommendations for practitioners and policy makers38. If textual pooling is not available, a narrative summary will be presented.
Following JBI guidance for MMSR, we will use a convergent segregated approach that conducts separate quantitative and qualitative syntheses and then integrates the findings of each26,39. This integration process is called ‘configuration’ and allows one to arrange complementary evidence into a single line of reasoning. After separate analyses are conducted, they will be organized into a coherent whole, as they cannot be directly combined nor can one refute the other40. Data will be triangulated during the interpretation stage to identify areas where there is convergence, inconsistency, or contradiction in the data. If we have a sufficient number of included studies, subgroup analyses will be performed to investigate differences based on participant groups (e.g., clinicians versus policy makers) and outcomes (e.g., understanding, acceptability, etc.). Within this approach, we will use thematic synthesis. This method codes text ‘line-by-line’, develops descriptive themes, and finally consolidates and generates analytical themes41,42. Initial coding will be performed independently by two authors who will meet and discuss similarities and differences in coding to start grouping them into a hierarchical tree structure of descriptive themes. A drafted summary of findings will be created by one author, reviewed by both, and discussions will be held until a final version is agreed upon. Two authors will discuss the descriptive themes and, as a group, will draft the final analytical themes with accompanying detailed descriptions.
As the focus of this review is not evaluating health-related interventions nor outcomes, we will not register the protocol on PROSPERO. However, we will preregister the study on Open Science Framework. If an amendment to this protocol is necessary, the date of each amendment will be given alongside the rationale and description of the change(s). This information will be detailed in an appendix accompanying the final systematic review publication. Changes will not be incorporated into the protocol.
Findings will be disseminated as peer-reviewed publications. Data generated from the work proposed within this protocol will be made available on the aforementioned OSF project page.
This review will summarise the evidence on the effectiveness and acceptability of different evidence synthesis summary formats. By including a variety of evidence summary types and stakeholder participants, results can help tease apart the real-world complexity of guideline development groups and provide an overview of what summary formats work for which stakeholders in what circumstances. It is expected that review findings can support decision-making by policy-makers and GDGs, by establishing the best summary formats for presenting evidence synthesis findings.
No data are associated with this article.
OSF: PRISMA-P checklist for ‘Evidence synthesis summary formats for clinical guideline development group members: a mixed-methods systematic review protocol’. https://doi.org/10.17605/OSF.IO/SK4NX43
Data are available under the terms of the Creative Commons Attribution 4.0 International license (CC-BY 4.0).
We would like to thank Paul Murphy for his help with developing the search strategy and translating it to other databases.
Is the rationale for, and objectives of, the study clearly described?
Yes
Is the study design appropriate for the research question?
Yes
Are sufficient details of the methods provided to allow replication by others?
Partly
Are the datasets clearly presented in a useable and accessible format?
Not applicable
Competing Interests: No competing interests were disclosed.
Reviewer Expertise: Public health; Health communication
Is the rationale for, and objectives of, the study clearly described?
Yes
Is the study design appropriate for the research question?
Yes
Are sufficient details of the methods provided to allow replication by others?
No
Are the datasets clearly presented in a useable and accessible format?
Not applicable
Competing Interests: No competing interests were disclosed.
Reviewer Expertise: Methods
Alongside their report, reviewers assign a status to the article:
Invited Reviewers | ||
---|---|---|
1 | 2 | |
Version 2 (revision) 10 May 22 |
read | |
Version 1 15 Jul 21 |
read | read |
Provide sufficient details of any financial or non-financial competing interests to enable users to assess whether your comments might lead a reasonable person to question your impartiality. Consider the following examples, but note that this is not an exhaustive list:
Sign up for content alerts and receive a weekly or monthly email with all newly published articles
Register with HRB Open Research
Already registered? Sign in
Submission to HRB Open Research is open to all HRB grantholders or people working on a HRB-funded/co-funded grant on or since 1 January 2017. Sign up for information about developments, publishing and publications from HRB Open Research.
We'll keep you updated on any major new updates to HRB Open Research
The email address should be the one you originally registered with F1000.
You registered with F1000 via Google, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Google account password, please click here.
You registered with F1000 via Facebook, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Facebook account password, please click here.
If your email address is registered with us, we will email you instructions to reset your password.
If you think you should have received this email but it has not arrived, please check your spam filters and/or contact for further assistance.
1) the introduction section to more clearly relate to the main objectives of the mixed-methods systematic review, and
2) added clarifying information to the methods section. Specifically we
a) clarified information regarding the inclusion criteria (PICO);
b) added additional information about the quantitative systematic review including what study designs and summary formats were eligible and how we defined ‘health literacy’ as an exclusion outcome;
c) clarified information about the data extraction form and data collection;
d) edited the bias and quality assessments section for clarity;
e) clarified the exploration of heterogeneity/subgroup analyses;
f) elaborated and simplified the mixed methods synthesis section; and
g) added a few additional references to support our edits.
1) the introduction section to more clearly relate to the main objectives of the mixed-methods systematic review, and
2) added clarifying information to the methods section. Specifically we
a) clarified information regarding the inclusion criteria (PICO);
b) added additional information about the quantitative systematic review including what study designs and summary formats were eligible and how we defined ‘health literacy’ as an exclusion outcome;
c) clarified information about the data extraction form and data collection;
d) edited the bias and quality assessments section for clarity;
e) clarified the exploration of heterogeneity/subgroup analyses;
f) elaborated and simplified the mixed methods synthesis section; and
g) added a few additional references to support our edits.