Enhancing The Quality and Credibility of Qualitative Studies
Enhancing The Quality and Credibility of Qualitative Studies
Enhancing The Quality and Credibility of Qualitative Studies
QUALITATIVE STUDIES
NCM 111- NURSING RESEARCH 1 (MA’AM COJUANGCO)
January 6, 2021
4 Distinct Elements of Credibility in Qualitative Inquiry Analytic Induction: Hypothesis Testing With Negative
Cases
Systemic, in-depth fieldwork
Systematic and conscientious analysis of data Emphasizes giving special attention to negative
Credibility of the researcher or deviant cases for testing propositions
Readers’ and users’ philosophical belief in the Works through one case at a time
value of qualitative inquiry No case can be ignored
Conscientious search for alternative themes Qualitative methods are not weaker or softer as
Divergent patterns compared to quantitative methods, it is just
Rival explanations DIFFERENT.
1189
1190 HSR: Health Services Research 34:5 Part II (December 1999)
Triangulation
The term "triangulation" is taken from land surveying. Knowing a single land-
mark only locates you somewhere along a line in a direction from the land-
mark, whereas with two landmarks you can take bearings in two directions
and locate yourself at their intersection. The notion of triangulating also works
metaphorically to call to mind the world's strongest geometric shape-the
triangle (e.g., the form used to construct geodesic domes a la Buckminster
Fuller). The logic of triangulation is based on the premise that no single
method ever adequately solves the problem of rival explanations. Because
each method reveals different aspects of empirical reality, multiple methods
of data collection and analysis provide more grist for the research mill.
Triangulation is ideal, but it can also be very expensive. A researcher's
limited budget, short time frame, and narrow training will affect the amount
of triangulation that is practical. Combinations of interview, observation, and
document analysis are expected in much fieldwork. Studies that use only
one method are more vulnerable to errors linked to that particular method
(e.g., loaded interview questions, biased or untrue responses) than are studies
that use multiple methods in which different types of data provide cross-data
validity checks.
Enhancing Quality and Credibility 1193
situation. Moreover, few researchers are equally comfortable with both types
of data, and the procedures for using the two together are still emerging.
The tendency is to relegate one type of analysis or the other to a secondary
role according to the nature of the research and the predilections of the
investigators. For example, observational data are often used for generating
hypotheses or describing processes, while quantitative data are used to make
systematic comparisons and verify hypotheses. Although it is common, such
a division of labor is also unnecessarily rigid and limiting.
Given the varying strengths and weaknesses of qualitative versus quan-
titative approaches, the researcher using different methods to investigate the
same phenomenon should not expect that the findings generated by those
different methods will automatically come together to produce some nicely
integrated whole. Indeed, the evidence is that one ought to expect that initial
conflicts will occur between findings from qualitative and quantitative data
and that those findings will be received with varying degrees of credibility.
It is important, then, to consider carefully what each kind of analysis yields
and to give different interpretations the chance to arise and be considered
on their merits before favoring one result over another based on methodo-
logical biases. 4
In essence, triangulation of qualitative and quantitative data is a form
of comparative analysis. "Comparative research," according to Fielding and
Fielding (1986: 13), "often involves different operational measures of the
'same' concept, and it is an acknowledgement of the numerous problems
of 'translation' that it is conventional to treat each such measure as a separate
variable. This does not defeat comparison, but can strengthen its reliability."
Subsequently, deciding whether results have converged remains a delicate
exercise subject to both disciplined and creative interpretation. Focusing on
what is learned by the degree of convergence rather than forcing a dichotomous
choice-the different kinds of data do or not converge-typically yields a more
balanced overall result.
That said, it is worth noting that qualitative and quantitative data can be
fruitfully combined when they elucidate complementary aspects of the same
phenomenon. For example, a community health indicator (e.g., teenage preg-
nancy rate) can provide a general and generahizable picture, while case studies
of a few pregnant teenagers can put faces on the numbers and illuminate the
stories behind the quantitative data; this becomes even more powerful when
the indicator is broken into categories (e.g., those under age 15, those 16 and
over) with case studies illustrating the implications of and rationale for such
categorization.
Enhancing Quality and Credibility 1195
and validity of their data analysis by having the people described in that data
analysis react to what is described. To the extent that participants in the study
are unable to relate to the description and analysis in a qualitative evaluation
report, it is appropriate to question the credibility of the report. Intended
users of findings, especially users of evaluations or policy analyses, provide
yet another layer of analytical and interpretive triangulation. In seriously
soliciting users' reactions to the face validity of descriptive data, case studies,
and interpretations, the evaluator's perspective is joined to the perspective
of the people who must use the information. House (1977) suggests that the
more "naturalistic' the evaluation, the more it relies on its audiences to reach
their own conclusions, draw their own generalizations, and make their own
interpretations. House is articulate and insightful on this critical point:
Unless an evaluation provides an explanation for a particular audience, and
enhances the understanding of that audience by the content and form of the
argument it presents, it is not an adequate evaluation for that audience, even
though the facts on which it is based are verifiable by other procedures. One
indicator ofthe explanatory power of evaluation data is the degree to which the
audience is persuaded. Hence, an evaluation may be "true" in the conventional
sense but not persuasive to a particular audience for whom it does not serve as
an explanation. In the fullest sense, then, an evaluation is dependent both on
the person who makes the evaluative statement and on the person who receives
it (p. 42).
Theory Triangulation
A fourth kind of triangulation involves using different theoretical perspectives
to look at the same data. A number of general theoretical frameworks derive
from divergent intellectual and disciplinary traditions, e.g., ethnography,
symbolic interaction, hermeneutics, or phenomenology (Patton 1990: ch. 3).
More concretely, there are always multiple theoretical perspectives that can
be brought to bear on substantive issues. One might examine interviews with
therapy clients from different psychological perspectives: psychoanalytic,
Gestalt, Adlerian, rational-emotive, and/or behavioral frameworks. Obser-
vations of a group, community, or organization can be examined from a
Marxian or Weberian perspective, a conffict or functionalist point of view.
The point of theory triangulation is to understand how findings are affected
by different assumptions and fundamental premises.
A concrete version of theory triangulation for evaluation is to examine
the data from the perspective of various stakeholder positions with different
theories of action about a program. It is common for divergent stakeholders
to disagree about program purposes, goals, and means of attaining goals.
Enhancing Quality and Credibility 1197
three weeks during which time the evaluator had severe diarrhea. Did that
affect the highly negative tone of the report? The evaluator said it didn't,
but I'd want to have the issue out in the open to make my own judgment.)
Background characteristics of the researcher (e.g., gender, age, race, ethnicity)
may also be relevant to report in that such characteristics can affect how the
researcher was received in the setting under study and related issues.
In preparing to interview farm families in Minnesota I began building
up my tolerance for strong coffee a month before the fieldwork was scheduled.
As ordinarily a non-coffee drinker, I knew my body would be jolted by 10-12
cups of coffee a day doing interviews in farm kitchens. These are matters of
personal preparation, both mental and physical, that affect perceptions about
the quality of the study. Preparation and training are reported as part of the
study's methodology..
Researcher Training, Experience, and Preparation
Every student who takes an introductory psychology or sociology course
learns that human perception is highly selective. When looking at the same
scene, design, or object, different people will see different things. What people
"see" is highly dependent on their interests, biases, and backgrounds. Our
culture tells us what to see; our early childhood socialization instructs us in
how to look at the world; and our value systems tell us how to interpret
what passes before our eyes. How, then, can one trust observational data? Or
qualitative analysis?
In their popular text, Evaluating Information: A Guide for Users ofSocial
Science Research, Katzer, Cook, and Crouch (1978) titled their chapter on
observation "Seeing Is Not Believing." In that chapter they tell an oft-repeated
story meant to demonstrate the problem with observational data:
Once at a scientific meeting, a man suddenly rushed into the midst of one of the
sessions. He was being chased by another man with a revolver. They scuffled in
plain view of the assembled researchers, a shot was fired, and they rushed out.
About twenty seconds had elapsed. The chairperson ofthe session immediately
asked all present to write down an account ofwhat they had seen. The observers
did not know that the ruckus had been planned, rehearsed, and photographed.
Of the forty reports turned in, only one was less than 20-percent mistaken about
the principal facts, and most were more than 40-percent mistaken. The event
surely drew the undivided attention of the observers, was in full view at close
range, and lasted only twenty seconds. But the observers could not observe all
that happened. Some readers chucdkled because the observers were researchers,
but similar experiments have been reported numerous times. They are alike for
all kinds of people. (pp. 21-22)
1200 HSR: Health Services Research 34:5 Part II (December 1999)
The third concern about evaluator effects has to do with the extent to
which the predispositions or biases of the evaluator may affect data analysis
and interpretations. This issue involves a certain amount of paradox because,
on the one hand, rigorous data collection and analytical procedures, like trian-
gulation, are aimed at substantiating the validity of the data; on other hand, the
interpretive and social construction underpinnings of the phenomenological
paradigm mean that data from and about humans inevitably represent some
degree of perspective rather than absolute truth. Getting close enough to
the situation observed to experience it firsthand means that researchers can
learn from their experiences, thereby generating personal insights, but that
closeness makes their objectivity suspect.
This paradox is addressed in part through the notion of "emphatic
neutrality," a stance in which the researcher or evaluator is perceived as caring
about and being interested in the people under study, but as being neutral
about the findings. House suggests that the evaluation researcher must be
seen as impartial:
The evaluator must be seen as caring, as interested, as responsive to the relevant
arguments. He must be impartial rather than simply objective. The impartiality
of the evaluator must be seen as that of an actor in events, one who is responsive
to the appropriate arguments but in whom the contending forces are balanced
rather than nonexistent. The evaluator must be seen as not having previously
decided in favor of one position or the other. (House 1977: 45-46)
Neutrality and impartiality are not easy stances to achieve. Denzin
(1989) cites a number of scholars who have concluded, as he does, that
all researchers bring their own preconceptions and interpretations to the
problem being studied, regardless of the methods used:
All researchers take sides, or are partisans for one point of view or another.
Value-free interpretive research is impossible. This is the case because every
researcher brings preconceptions and interpretations to the problem being
studied. The term "hermeneutical circle or situation" refers to this basic fact
of research. All scholars are caught in the circle of interpretation. They can
never be free of the hermeneutical situation. This means that scholars must state
beforehand their prior interpretations of the phenomenon being investigated.
Unless these meanings and values are clarified, their effects on subsequent
interpretations remain clouded and often misunderstood. (p. 23)
Debate about the research value of qualitative methods means that
researchers must make their own peace with how they are going to describe
what they do. The meaning and connotations of words like objectivity, subjec-
tivity, neutrality, and impartiality will have to be worked out with particular
Enhancing Qyality and Credibility 1205
audiences in mind. Essentially, these are all concerns about the extent to
which the qualitative researcher can be trusted; that is, trustworthiness is
one dimension of perceived methodological rigor (Lincoln and Guba 1985).
For better or worse, the trustworthiness of the data is tied directly to the
trustworthiness of the researcher who collects and analyzes the data.
The final issue affecting evaluator effects is that of competence. Com-
petence is demonstrated by using the verification and validation procedures
necessary to establish the quality of analysis, and it is demonstrated by
building a track record of fairness and responsibility. Competence involves
neither overpromising nor underproducing in qualitative research.
Intellectual Rigor
The thread that runs through this discussion of researcher credibility is the
importance of intellectual rigor and professional integrity. There are no simple
formulas or clear-cut rules to direct the performance ofa credible, high-quality
analysis. The task is to do one's best to make sense out of things. A qualitative
analyst returns to the data over and over again to see if the constructs,
categories, explanations, and interpretations make sense, if they really reflect
the nature of the phenomena. Creativity, intellectual rigor, perseverance,
insight-these are the intangibles that go beyond the routine application of
scientific procedures. As Nobel prize winning physicist Percy Bridgman put
it: "There is no scientific method as such, but the vital feature of a scientist's
procedure has been merely to do his utmost with his mind, no holds barred
(quoted in Mills 1961: 58, emphasis in the original).
But it is this aspect of paradigms that constitutes both their strength and their
weakness-their strength in that it makes action possible, their weakness in
that the very reason for action is hidden in the unquestioned assumptions of
the paradigm.
Beyond the Numbers Game
Thomas H. Kuhn is a philosopher of science who has extensively studied the
value systems of scientists. Kuhn (1970: 184-85) has observed that "the most
deeply held values concern predictions." He goes on to observe that "quantita-
tive predictions are preferable to qualitative ones." The methodological status
hierarchy in science ranks "hard datae above "soft data" where "hardness"
refers to the precision of statistics. Qualitative data, then, carry the stigma of
being "soft."
I've already argued that among leading methodologists in fields like
evaluation research, this old bias has diminished or even disappeared. But
elsewhere it prevails. How can one deal with a lingering bias against qualita-
tive methods?
The starting point is understanding and being able to communicate the
particular strengths of qualitative methods and the kinds of empirical ques-
tions for which qualitative data are especially appropriate. It is also helpful to
understand the special seductiveness of numbers in modern society. Numbers
convey a sense of precision and accuracy even if the measurements that
yielded the numbers are relatively unreliable, invalid, and meaningless. The
point, however, is not to be anti-numbers. The point is to be pro-meaningfulness.
Moreover, as noted in discussing the value of methods triangulation, the issue
need notbe quantitative versus qualitative methods, but rather how to combine
the strengths of each in a multimethods approach to research and evaluation.
Qualitative methods are not weaker or softer than quantitative approaches;
qualitative methods are diferent.
the early evaluation literature, the debate between qualitative and quantitative
methodologists was often strident. In recent years it has softened. A consensus
has gradually emerged that the important challenge is to match methods ap-
propriately to empirical questions and issues, and not to universally advocate
any single methods approach for all problems.
In a diverse world, one aspect of diversity is methodological. From time
to time regulations surface in various federal and state agencies that prescribe
universal, standardized evaluation measures and methods for all programs
funded by those agencies. I oppose all such regulations in the belief that
local program processes are too diverse and client outcomes too complex
to be fairly represented nationwide, or even statewide, by some narrow set
of prescribed measures and methods-regardless whether the mandate be
for quantitative or for qualitative approaches. When methods decisions are
based on some universal, political mandate rather than on situational merit,
research offers no challenge, requires no subtlety, presents no risk, and allows
for no accomplishment.
REFERENCES
Denzin, N. K 1989. Interpretive Interactionism. Newbury Park, CA: Sage.
. 1978. The Research Act: A Theoretical Introduction to Sociological Methods. New
York: McGraw-Hill.
Fielding, N.J., andJ. L. Fielding. 1986. Linking Data. Qualitative Methods Series No.
4. Newbury Park, CA: Sage Publications.
House, E. 1977. TheLogic ofEvaluativeArgument. CSE Monograph Series in Evaluation
no. 7. Los Angeles: University of California, Los Angeles, Center for the Study
of Evaluation.
Katzer,J., K Cook, and W Crouch. 1978. Evaluating Information: A Guidefor Users of
Social Science Research. Reading, MA: Addison-Wesley.
Kuhn, T. 1970. The Structure ofScientific Revolutions. Chicago: University of Chicago
Press.
Lang, K, and G. E. Lang. 1960. "Decisions for Christ: Billy Graham in New York
City." In Identity and Anxiety, edited by M. Stein, A. Vidich, and D. White. New
York: Free Press.
Lincoln, Y, and E. G. Guba. 1985. Naturalistic Inquiry. Newbury Park, CA: Sage.
Mills, C. W 1961. The Sociological Imaination. New York: Oxford University Press.
Patton, M. Q 1999. Grand Canyon Cekbration:AFather-SonJourny ofDiscovery. Amherst,
NY: Prometheus Books.
. 1997. Utilization-FocusedEvaluation: The New Century Text. Thousand Oaks, Ca:
Sage Publications.
. 1990. Qualitative Evaluation and Research Methods. Thousand Oaks, CA: Sage
Publications.