thoughts on critical appraisal of qualitative research Esquire course 2015 University of Sheffield Karin Hannes Faculty of P sychology and E ducational S ciences KU Leuven ID: 314940
Download Presentation The PPT/PDF document "Advanced" is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.
Slide1
Advanced thoughts on critical appraisal of qualitative researchEsquire course 2015, University of Sheffield
Karin Hannes
Faculty
of
P
sychology
and
E
ducational
S
ciences, KU LeuvenSlide2
The aspirin caseSlide3
The aspirin case transferred to critical appraisalThe more you appraise, the lesser the chance to end up with flawed results. The main criterion is quality!
The more you appraise, the more it stifles creativity. The main criterion is
relevance
!
Qualitative
Scientists
Qualitative InquiristsSlide4
The aspirin case transferred to critical appraisalBasic questions
First issue
to
consider
Take a
good
, hard look at yourself and then answer the questions!Slide5
The worldview of a review author as a determining factor
Qualitative Science
Qualitative Inquiry
Meta-ethnography
Critical Interpretive synthesis
Meta-aggregation
Thematic synthesis
Framework synthesisMeta- Grounded TheoryMeta-narrativeEcological triangu-lationSlide6
The worldview of a review author as a determining factor
Meta-narrative
Critical
interpretive synthesis
Meta-ethno-
graphy
Grounded theory
Thematic synthesisThe JBI meta-aggregative approach
Framework synthesisEcological triangulationSubjective idealism
Subjective idealismObjective idealismObjective idealism
Critical realismCritical realism
Critical realismScientific realism
Based on Barnett-Page and Thomas, 2009Idealist Realist
There is no shared reality independent of multiple alternative human constructionsThere is a world of collectively shared understandings:Qualitative Science
Qualitative InquiryKnowledge of reality is mediated by our perceptions and beliefsIt is possible for knowledge to approximate closely an external realitySpencer, 2003Slide7
Qualitative Scientistsand critical appraisal“To see means”Masters of the windowClarityNeutrality
“What you see, is what you get.”
TransparancySlide8
Qualitative Scientistsand critical appraisal quality of the studyCheck the credibility of the findings in terms of an accurate display of people’s voices
Check the means to correct for the impact of the researcher on the findings
Check whether the conclusions are grounded in the data
Before using it for decision making processesSlide9
Qualitative Scientists and critical appraisal
Questions
Answers
Should
we
appraise
?YesWhich criteria should we use
?Translation of general quality concepts such as validity, generalizability, reliability, objectivityWhat approaches to quality assessment will likely be taken?Criterion based judgmentHow do researchers generally deal with the outcome of a critical appraisal?Exclude low quality studiesSlide10
Qualitative Inquiristsand critical appraisal“TO see means”MASTERS OF THE LANTERN
To step in someone else’s shoes
To explore the dark corners or gaps in our knowledge base
To go beyond what has been reported in the primary studies
To
problemize
existing literature
“Shed light where there has been no light before” ILLUMINATIONSlide11
Qualitative Inquiristsand critical appraisalQUALITY OF THE STUDY
the process of systematically examining research evidence to assess its relevance and utility for the story line to be developed
before using it to inform a decisionSlide12
Qualitative Inquirists and critical appraisal
Questions
Answers
Should
we
appraise
?Not necessarilyWhich criteria
should we use?Incisiveness, concision, coherence, generativity, illuminationWhat approaches to quality assessment will likely be taken?Overall judgment approachHow do researchers generally deal with the outcome of a critical appraisal?Use quality appraisal as a baseline measurement, rather than seeing it as a measure to include or exclude studiesOne of the major problems is that those who portray themselves as idealist researchers or
qualitative inquirists adopt the rules of qualitative scientists!Slide13
Back to the aspirin case…Who is right and who is wrong?CONFIGURATIONAGGREGATION
Metaphors honestly stolen from Gough and Thomas, 2012
Transfer of Critical
appraisal
logic Slide14
Back to the aspirin case …Who is right and who is wrong?
The moose (?)
would subscribe to the argument that the impact of the researcher on the research is inherent to the way qualitative research is conducted. They may prefer a CAI evaluating issues such as ‘
thick description
’ and ‘
the innovative nature’ or ‘value for practice’
of the findings.
Epistemological and ontological assumptions of quantitative and qualitative research are incompatible.The elephant (?) who highly values the validity of primary studies would prefer a CAI that is sensitive to aspects of validity, including criteria such as ‘all statements should be well-grounded in the data’ and ‘the impact of the investigator on the study results should be reported’. Validity and researcher bias should
be evaluated Some qualitative studies are more rigorous
than others.Slide15
A more relevant question to ask Who am I?What does this potentially
imply
for
the
choices I make?Slide16
The Third Road: Pragmatism
Meta-narrative
Critical
interpretive synthesis
Meta-ethno-
graphy
Grounded theory
Thematic synthesisThe JBI meta-aggregative approach
Framework synthesisEcological triangulationSubjective idealism
Subjective idealismObjective idealismObjective idealism
Critical realismCritical realism
Critical realismScientific realism
Idealist RealistQualitative Science
Qualitative InquiryPRAGMATISMSlide17
PragmatismThe choice for using certain critical appraisal instruments (CAI) or criteria should be based on its ‘utility’ and ‘fit for purpose’ for the studies to be included in the reviews. Reviewers should select CAIs that are suitable for the retrieved original studies.Slide18
Pragmatism Example: If you are appraising an action research design informed by critical theory, you would use an instrument including transformative- emancipatory criteria:Do the authors openly reference a problem in a community of concern? Do the authors openly declare a theoretical lens?
Were the research questions or purposes written with an advocacy stance?
Did the literature review include discussions of diversity and oppression?
Did the authors discuss appropriate labeling of the participants?
Did data collection and outcomes benefit the community?
Were the stakeholders involved in the research project?
Did the results elucidate power relationships? Did the results facilitate social change, and were the stakeholders empowered as a result of the research process? Did the authors explicitly state their use of a transformative framework?
(Saini & Shlonsky, 2012, pp. 133-135; Sweetman, Badiee, & Creswell, 2010, pp. 442-443)Slide19
PragmatismExample:If you are appraising a mixed method design, you would use criteria that allow you to assess the added value of the mixed method component:
Was it appropriate or adequate to opt for an MMR study?
Was it legitimate and if so, has the rationale been provided?
Have both strands adequately been integrated?
Did the authors provide a clear and defensible rationale for mixing the findings of studies?
Wat there an overall benefit of triangulating designs or combining quantitative and qualitative evidence?
Did the combination of quantitative and qualitative evidence minimalize bias and if so, has it clearly been documented? (
Heyvaert, Hannes & Onghena, 2016, in press).Slide20
Short summaryApart from that, decide on:1. Choosing between
:
Existing
instrument or
self-compiled
set of criteria?
Generic or design specific frameworks/instruments for quality assessment?Generic
or design specific criteria for quality assessment? Taking into account fit for purpose and expertise!2. How to use the outcome of quality assessment in your review?Determine your positionQualitative scientistQualitative inquiristPragmatist?My prediction:You will be more certain about why you do what you do (and it feels good!)You will be less
prone to external influences deciding for you (or at least be able to discuss your thoughts with those promoting other strategies).Slide21
1. Choosing between instruments and criteria
M
otivate
your
choice thoroughly!Slide22
1. Choosing between instruments and criteria
Evaluate
the
strenghts
and weaknesses of available instrumentsSlide23
1. Choosing between instruments and criteria(study based on core criteria outlined in instruments)
Criterion
JBI
CASP
Screening Q
Theoretical framework
NO
Appropriateness design
Data collection procedure
Data-analysis procedure
Findings
Context
NO
Impact of investigator
Believability
NO
Ethics
Evaluation/Outcome
NO
Value/Implication Research
NO
Both instruments have focussed on the accuracy of the audit trail = Quality of reporting.
Slide24
2. Dealing with the outcome of a critical appraisal exercise
Basis
strategies
:
Use
it to in/
exclude studiesUse it as a baseline measure for quality without excluding anythingAssign more weight to studies that score high on qualitySlide25
Assign more weight to studies that score high on qualitySensitivity analyses
Levels
of
evidence
assigned
to findingsFrequency of themes combined with quality appraisal
.Slide26
A. Sensitivity analysisSensitivity analysis involves testing how sensitive the review
findings
are to the inclusion and exclusion of
studies
of
different
quality.Question: What would happen to the results if all studies
below a certain quality threshold would systematically be excluded?Slide27
A. Sensitivity analysis9/19 studies were judged
to
be
inadequately
reported for both reviews.Dichotomization of
studies: Adequately (>2)Inadequately (<2) reportedSlide28
A. Sensitivity analysesReview 1: Sexual health studies
No
single principal
theme
was
completely dependent on data from inadequately reported studies! No data
emerged as exclusive findings from the inadequately reported studies. With the exception of two viewpoints from lower quality studies, all instances of dissonance, richness or complexity for each theme emerged from on or more of the adequately reported studies. Review 2: Online learningOverall, data derived from inadequately reported studies did little to supplement data from adequately reported studies. Some of the richness was generated from inadequately reported studies.Excluding them would have resulted in the loss of valuable data on one particular
subgroup (nurses). As a consequence differences between doctors and nurses might have been conceiled.Consistent with studies from other authors having conducted sensititivity analyses (Noyes and Popay 2007, Thomas and Harden 2008)Slide29
B. Assign levels of evidence to findings
Author
statements
in
primary studies included in the synthesis can be considered unequivocal
, credible or unsupported, based on how well they are supported with excerpts from the data collection phase.The author of a review chooses whether or not to include unsupported and credible findings.Slide30
B. Assign levels of evidence to findingsUnequivocal: Where the theme or metaphor is unequivocally supported by direct quotes from the research participants. There is a clear relationship between the author’s theme or metaphor and the participants’ expressed experiences
.
Credible
:
where the participants’ expressed experiences are less clearly related to the author’s theme or metaphor
and the author has extended beyond the expressed experiences of the participants based on the participant quotes that have been used.Unsupported: where there is no relationship between the expressed experiences of the participants and the author’s themes or metaphors, then it is clear that the author is generating findings that are unsubstantiated by the participants.
Joanna Briggs Institute, AustraliaRationale: Excluding findings may be an interesting alternative to excluding studies as a whole.Slide31
B. Assign levels of evidence to findingsSlide32
C. Weighing the evidence: FrequenciesSlide33
C. Weighing the evidence: FrequenciesCounting the frequency
of a
theme
in an
included
article
Combine the frequency with the weight of quality
appraisal done by expert judgement (EJ)Combine the frequency with the weight of quality appraisal done by checklists (CA) Average score of 6.8 and Inter rater reliability=0.88 Average score of 6.9 and Inter rater reliability of 0.94Lower scores for criteria related to validity of the study, reporting of potential bias
, contextual info in order to evaluate transferability of the findings.Some articles scored low on the checklist but passed the expert judgement based on significance of the findings!
The evidence for high frequency themes increases, while for low frequency themes it declines!The direction of the change is the same for EJ and CA, CA has a better
differentiating ability.Slide34
C. Weighing the evidence: FrequenciesSome conclusions:When a
topic
is
frequently
studied in a methodologically sound way, there is strong
evidence for the value of the findings.Not all studies are of equal methodological quality and this should be accounted for when integrating findings.Limitation: Working with a summary score may conceal errors that can be considered fatal.Slide35
Closing remarks ...Checklists may only capture what has been reported (but reviewers can dig deeper ).To validate is to investigate, to check, to question, and to theorize. All of these activities are integral components of qualitative inquiry that insure rigor (Morse, 2002).In evaluating validity at the end of a study (post hoc), rather than focusing on processes of verification during the study we run the risk of missing serious threats to validity until it is too late to correct them. The process of inquiry is where the real verification happens. Slide36
Karin.hannes@kuleuven.beThank you!Questions?