Chapter 4 - Critical Appraisal of Qualitative Research: Key Points
Chapter 4 - Critical Appraisal of Qualitative Research: Key Points
This chapter should be cited as: Hannes K. Chapter 4: Critical appraisal of qualitative
research. In: Noyes J, Booth A, Hannes K, Harden A, Harris J, Lewin S, Lockwood C
(editors), Supplementary Guidance for Inclusion of Qualitative Research in Cochrane
Systematic Reviews of Interventions. Version 1 (updated August 2011). Cochrane
Collaboration Qualitative Methods Group, 2011. Available from URL
http://cqrmg.cochrane.org/supplemental-handbook-guidance
Key points
Introduction
Considerable debate exists on whether or not concepts such as validity and reliability
apply to qualitative research and if so how they could be assessed. Some researchers
have stated that qualitative research should establish validity, reliability and objectivity.
Others plead for an adjustment of these concepts to better fit the qualitative research
design. As a consequence, critical appraisal instruments might differ in the criteria they
1
list to complete a critical appraisal exercise. Some researchers consider appraisal
instruments a tool that can be utilized as part of the exploration and interpretation process
in qualitative research (Popay et al, 1998; Spencer, 2003). Edwards et al (2002) describes
the use of a “signal to noise” approach, where a balance is sought between the
methodological flaws of a study and the relevance of insights and findings it adds to the
overall synthesis. Other researchers do not acknowledge the value of critical appraisal of
qualitative research, stating that it stifles creativity (Dixon-Woods, 2004). While
recognising that all these views have some basis for consideration certain approaches
succeed in positioning the qualitative research enterprise as one that can produce a valid,
reliable and objective contribution to evidence synthesis. It is these that may therefore
have more potential to be generally accepted within the context of producing Cochrane
Intervention Reviews. The Cochrane Collaboration recommends a specific tool for
assessing the risk of bias in each included study in an intervention review, a process that
is facilitated through the use of appraisal instruments addressing the specific features of
the study design and focusing on the extent to which results of included studies should be
believed. This suggest that in assessing the methodological quality of qualitative studies
the core criterion to be evaluated is researcher bias. Believability in this context refers to
the ability and efforts of the researcher to make his or her influence and assumptions clear
and to provide accurate information on the extent to which the findings of a research
report hold true. However, it is the actual audit trail provided by researchers that allows
for an in-depth evaluation of a study. Most existing appraisal instruments use broader
criteria that account for reporting issues as well. We suggest that these issues should be
part of the appraisal exercise. Currently, there are four possibilities to make use of
qualitative research in the context of Cochrane Intervention reviews:
1. The use of qualitative research to define and refine review questions a Cochrane
Review (informing reviews).
2. The use of qualitative research identified whilst looking for evidence of
effectiveness (enhancing reviews).
3. The use of findings derived from a specific search for qualitative evidence that
addresses questions related to an effectiveness review (extending reviews).
4. Conducting a qualitative evidence synthesis to address questions other than
effectiveness (supplementing reviews).
The latter use (Supplementing) is beyond the scope of current Cochrane Collaboration
policy (Noyes et al, 2008). Stand alone qualitative reviews that supplement Cochrane
Intervention reviews need to be conducted and published outside of the Cochrane
context.
Critical appraisal applies to all of the above possibilities.
Reviewers should bear in mind that narratives used in reports of quantitative research
cannot be considered qualitative findings if they do not use a qualitative method of
datacollection and –analysis.Therefore, critical appraisal based on instruments developed
to assess qualitative studies is not applicable toreports that do not meet the criteria of
being a ‘qualitative study’..
2
This chapter breaks down in four sections. Section 1 addresses translated versions of
core criteria such as validity, reliability, generalisibility and objectivity of qualitative
studies. Section 2 presents an overview of different stages involved in quality
assessment. Section 3 guides the researcher through some of the instruments and
frameworks developed to facilitate critical appraisal and section 4 formulates suggestions
on how the outcome of an appraisal of qualitative studies can be used or reported in a
systematic review.
The concepts used in table 1 are based on Lincoln and Guba’s (1985) translation of
criteria to evaluate the trustworthiness of findings. Acknowledging the difference in
terminology does not obviate the rationale or process for critical appraisal. There might
be good congruence between the intent of meanings relevant to key aspects of
establishing study criteria, as demonstrated in table 1.
This scheme outlines some of the core elements to be considered in an assessment of the
quality of qualitative research. However, the concept of confirmability might not be
3
applicable to approaches inspired by phenomenology or critical paradigms in which the
researcher’s experience becomes part of the data (Morse, 2002). The choice of critical
appraisal instruments should preferably be inspired by those offering a multi-dimensional
concept of quality in research. Apart from methodological rigour, that would also include
quality of reporting and conceptual depth and bread.
There are a variety of evaluation techniques that authors might have included in their
original reports, that facilitate assessment by a reviewer and that are applicable to a broad
range of different approaches in qualitative research. However, it should be stated that
some of the techniques listed only apply for a specified set of qualitative research
designs.
The criteria listed might generate an understanding of what the basic methodological
standard is a qualitative study should be able to reach. However, a study may still be
4
judged to have followed the appropriate procedures for a particular approach, yet may
suffer from poor interpretation and offer little insight into the phenomenon at hand.
Consequently, another study may be flawed in terms of transparency of methodological
procedures and yet offer a compelling, vivid and insightful narrative, grounded in the data
(Dixon-Woods et al, 2004). Defining fatal flaws and balancing assessment against the
weight of a message remains a difficult exercise in the assessment of qualitative studies.
As in quantitative research, fatal flaws may depend on the specific design or method
chosen (Booth, 2001). This issue needs further research.
Debates in the field of quality assessment of qualitative research designs are centred
around a more theoretical approach to evaluating the quality of studies versus an
evaluation of the technical adequacy of a research design. How far criteria-based,
technical approaches offer significant advantages over expert intuitive judgement in
assessing the quality of qualitative research is being challenged by recent evidence
indicating that checklist-style approaches may be no better at promoting agreement
between reviewers (Dixon-Woods, 2007). However, these appraisal instruments might
succeed better in giving a clear explanation as to why certain papers have been excluded.
Given the fact that few studies are completely free from methodological flaws, both
approaches can probably complement each other.
Stage 1: Filtering:
Within the specific context of enhancing or extending Cochrane Reviews, and
viewing critical appraisal as a technical and paradigmatic exercise, it is worth
considering limiting the type of qualitative studies to be included in a systematic
review. We suggest restricting included qualitative research reports to empirical
studies with a description of the sampling strategy, data collection procedures and
the type of data-analysis considered. This should include the methodology chosen
and the methods or research techniques opted for, which facilitates the systematic
use of critical appraisal as well as a more paradigmatic appraisal process.
Descriptive papers, editorials or opinion papers would generally be excluded.
5
Critical appraisal instruments should be considered a technical tool to assist in the
appraisal of qualitative studies, looking for indications in the methods or
discussion section that add to the level of methodological soundness of the study.
This judgement determines the extent to which the reviewers may have
confidence in the researcher’s competence in being able to conduct research that
follows established norms (Morse, 2002) and is a minimum requirement for
critical assessment of qualitative studies. Criteria include but are not limited to
the appropriateness of the research design to meet the aims of the research, rigour
of data-collection and analysis, well-conducted and accurate sampling strategy,
clear statements of findings, accurate representation of participants’ voices,
outline of the researchers’ potential influences, background, assumptions,
justifications of the conclusion or whether or not it flows from the data, value and
transferability of the research project etc. For this type of appraisal one needs to
have a general understanding of qualitative criteria. Involving a researcher with a
qualitative background is generally recommended.
The Cochrane Qualitative Research Methods group recommends stage 3 whenever the
instrument chosen for stage 2 does not cover for a paradigmatic approach to judgment.
Other considerations include involving people with content expertise for the evaluation
exercise. They are believed to give more consistent assessments, which is in line with
what the Cochrane Collaboration suggests for the assessment of risk of bias in trials
(Oxman et al, 1993).
6
Section 3: A selection of instruments for quality assessment
A range of appraisal instruments and frameworks is available for use in the assessment of
the quality of qualitative research. Some are generic, being applicable to almost all
qualitative research designs; others have specifically been developed for use with certain
methods or techniques. The instruments also vary with regard to the criteria that they use
to guide the critical appraisal process. Some address paradigmatic aspects related to
qualitative research, others tend to focus on the quality of reporting more than theoretical
underpinnings. Nearly all of them address credibility to some extent. The list with
examples presented below is not exclusive with many instruments still in development or
yet to be validated and others not yet commonly used in practice. It draws on the findings
of a review of published qualitative evidence syntheses (Dixon-Woods et al, 2007) and
the ongoing update of it. Reviewers need to decide for themselves which instrument
appears to be most appropriate in the context of their review and use this judgement to
determine their choice. Researchers with a quantitative background also need to consider
an input from a researcher familiar with qualitative research, even when an appraisal
instrument suitable for novices in the field is opted for.
Examples:
QARI software developed by the Joanna Briggs Institute, Australia
URL: http://www.joannabriggs.edu.au/services/sumari.php
Used by: Pearson A, Porritt KA, Doran D, Vincent L, Craig D, Tucker D,
Long L, Henstridge V. A comprehensive systematic review of evidence on
the structure, process, characteristics and composition of a nursing team
that fosters a healthy environment. International Journal of Evidence-
Based Healthcare 2006; 4(2): 118-59.
1
URL: http://www.joannabriggs.edu.au/cqrmg/tools_3.html. A detailed guide on how to conduct a QARI
supported Systematic Review, including a detailed explanation of the10 critical appraisal criteria, can be
found on the JBI-website: http://www.joannabriggs.edu.au/pdf/sumari_user_guide.pdf
URL for instrument on process evaluation: http://eppi.ioe.ac.uk/cms/default.aspx?
tabid=2370&language=en-US. Tools exist which help to assess quality along three dimensions: quality of
reporting, sufficiency of strategies for increasing methodological rigour, and the extent to which study
methods and findings are appropriate to answer the review question (For an example, see Harden et al 2009
study).
7
Rhodes LG et al.Patient subjective experience and satisfaction during the
perioperative period in the day surgery setting: a systematic review. Int J
Nurs Pract 2006; 12(4): 178-92.
Examples:
Critical Appraisal Skills Programme (CASP):
http://www.phru.nhs.uk/Doc_Links/Qualitative%20Appraisal%20Tool.pdf
Used by: Kane GA et al. Parenting programmes: a systematic review and
synthesis of qualitative research. Child Care Health and Development
2007; 33(6): 784-793.
Modified versions of CASP, used by:
Campbell R, Pound P, Pope C, Britten N, Pill R, Morgan M, Donovan J.
Evaluating meta-ethnography: a synthesis of qualitative research on lay
experiences of diabetes and diabetes care. Social Science and Medicine
2003; 56: 671-84.
Malpass A, Shaw A, Sharp D, Walter F, Feder G, Ridd M, Kessler D.
‘Medication career" or "Moral career"? The two sides of managing
antidepressants: A meta-ethnography of patients' experience of
antidepressants. Soc Sci Med. 2009; 68(1):154-68.
8
Checklists developed by academics and commonly used in published
qualitative evidence syntheses: Such checklists have been selected and utilised
by other researchers in the specific context of an evidence synthesis.
Examples:
The Blaxter (1996) criteria for the evaluation of qualitative research papers,
used by:
Gately C et al. Integration of devices into long-term condition
management: a synthesis of qualitative studies. Chronic Illn 2008; 4(2):
135-48.
Khan N et al. Guided self-help in primary care mental health - Meta-
synthesis of qualitative studies of patient experience. British Journal of
Psychiatry 2007; 191: 206-211.
9
Section 4: Integrating outcomes of critical appraisal in a systematic
review.
In a ‘best case’ scenario a qualitative synthesis or primary study will achieve a positive
assessment or score for each of the criteria against which it has been assessed according
to the critical appraisal instrument used. However, this will most likely not be the case
for the majority of studies and researchers need to be aware of the fact that the
assessment or score might depend on the instrument that has been used, which increases
the value of involving a researcher with a qualitative background in the appraisal process.
For studies that fail to report sufficient information or it is clear that the study is weak
when matched against a certain criterion e.g. because of a methodological flaw a
decision needs to be made whether to include the study or not.
To include or exclude a study: In this particular case, only high quality studies are
included. The potential risk is that valuable insights are excluded from the
synthesis. Studies rated as “low quality” because of methodological flaws or lack
of reporting may nevertheless generate new insights, grounded in the data, while
methodological sound studies may suffer from poor interpretation of data, leading
to an insufficient insight into the phenomenon under study (Dixon-Woods, 2007).2
This approach was used by Carlsen et al (2007), who excluded 5 out of 17 studies
following quality appraisal.
*Authors may choose to give more weight to certain criteria and use this in their final judgment.
** H/L= High/Low
2
The loss of potential valuable studies is less likely if one would opt for a critical appraisal instrument
which assesses conceptual depth and breadth of findings as well as methodological rigour. Currently, there
is no guidance on how these two aspects might be balanced out. Sensitivity analyses could be considered.
10
*** For studies that are clearly on the verge between in- and exclusion researchers a judgement on
whether to include or exclude should be made and discussed with potential co-reviewers.
**** Authors should include a motivation for in- or exclusion, particularly for those cases where
judgments are being made.
To give more weight to studies that scored high on quality: In this particular case,
all valuable insights remain included. However, it might be complex to report on
the findings of the synthesis given the ‘subgroups’ of studies. One strategy to
cope with the issue of weighing studies is to report the findings of differently
rated studies in separate sections. However, this has an impact on the potential
richness of the presented synthesis, especially in those approaches generating new
theory based on all of the relevant and illuminating findings. No fixed parameters
currently exist to determine the weight of qualitative studies. Reviewers choosing
this approach need to evaluate which methodological flaws have a substantial
impact on the findings presented. The key issue to consider is the extent to which
the quality of reporting and choices made by the authors is acceptable in terms of
inclusion of the evidence in the Cochrane review which it aims to enhance.
Both approaches could benefit from a sensitivity analysis evaluating what happens to the
findings of a study when low or high quality studies are removed. Thomas et al (2004)
conducted such a sensitivity analysis and found that the findings of three studies rated as
low quality did not contradict those from studies of a higher quality. This was confirmed
by Noyes’ and Popays’ (2007) study on directly observed therapy and tuberculosis. The
synthesis would have come to the same conclusions with or without their inclusion. It
indicates that there might be little to gain from including lower quality studies in a
synthesis (Harden, 2008).
To describe what has been observed without excluding any studies: In this
particular case, all potential valuable insights remain included, because the worth
of individual studies might only become recognisable at the point of synthesis
rather than in the phase of appraisal. In this approach, the responsibility for
evaluating the quality of the studies is devolved to the reader from the researcher.
11
The Cochrane Qualitative Research Methods Group sees value in all of these approaches.
However, in line with current Cochrane policy, when conducting a Cochrane Intervention
review and integrating qualitative evidence we recommend the two first approaches
emphasizing the methodological soundness of studies rather than their contribution to
science in general. The decision lies with the review team. Regardless of the approach
eventually chosen for the quality assessment stage of the review there is a need to
preserve the transparency of the method through careful documentation of decisions
made. The convention of using at least two researchers for the quality assessment process
is a useful legacy from quantitative-based review processes; not so much for inter-rater
consistency purposes but, at the very least, to open up the data to a broader range of
possible interpretations.
Conclusion
References
Blaxter M. Criteria for evaluation of qualitative research. Medical Sociology News 1996;
22: 68-71.
12
Burns N. Standards for Qualitative Research. Nurs Sci Q 1989; 2(44): 44-52.
Carlsen B, Glenton C, Pope C. Thou shalt versus thou shalt not: a meta-synthesis of GPs'
attitudes to clinical practice guidelines. Br J Gen.Pract 2007;57:971-8.
Cohen DJ & Crabtree BF. Evaluative criteria for qualitative research in health care;
controversies and recommendations. Annals of Fam. Med 2008; 6 (4): 331-39.
Dixon-Woods M, Shaw RL, Agarwal S, Smith JA. The problem of appraising qualitative
research. QSHC 2004;13: 223-225.
Dixon-Woods M, Sutton AF, Shaw RL, Miller T, Smith J, Young B, Bonas S, Booth A,
Jones DR. Appraising qualitative research for inclusion in systematic reviews: a
quantitative and qualitative comparison of three methods. Journal of Health Service
Research and Policy 2007; 12: 42-7.
Lincoln YS, Guba EG. Naturalistic Inquiry, Sage, Newbury Park, CA, 1985 .
Guba EG. & Lincoln YS. Fourth Generation evaluation. Newbury Park. CA: Sage. 1989.
Morse JM, Barett M, Mayan M, Olson K, Spiers J. Verification strategies for establishing
reliability and validity in qualitative research. Int J of Qual Meth 2002; 1 (2):spring.
13
Mays N, Pope C. Qualitative research in health care Assessing quality in qualitative
research. BMJ 2000;320:50-52.
Noyes J, Popay J, Pearson A, Hannes K, Booth A. Chapter 20: Qualitative research and
Cochrane reviews. In: Higgins JPT, Green S (editors). Cochrane Handbook for
Systematic Reviews of Interventions. Version 5.0.1 [updated September 2008]. The
Cochrane Collaboration, 2008. Available from www.cochrane-handbook.org.
Oxman AD, Guyatt GH. The science of reviewing research. Annals of the New York
Academy of Sciences 1993; 703: 125-133.
Popay, J.,& Williams, G. (1998). Qualitative research and evidence based healthcare.
Journal of Research in Social Medicine,91(Suppl 35), 32–37.
Popay J, Rogers A, Williams G. Rationale and Standards for the Systematic Review of
Qualitative Literature in Health Care. Qual Health Research 1998; 8: 341-51.
14