deGruyter-Medical Errors and Patient Safety
deGruyter-Medical Errors and Patient Safety
deGruyter-Medical Errors and Patient Safety
Volume 1
Jay Kalra
DE GRUYTER
Author
Jay Kalra, MD, PhD, FRCPC, FCAHS
Department of Pathology and Lab Medicine
College of Medicine
University of Saskatchewan
Royal University Hospital
103 Hospital Drive
Saskatoon
Saskatchewan S7N 0W8
Canada
Kalra, Jay.
Medical errors and patient safety / by Jay Kalra.
p. ; cm. — (Patient safety)
ISBN 978-3-11-024949-1 (alk. paper)
1. Medical errors—United States. 2. Patient safety. 3. Patients—Safety measures—
United States. I. Title. II. Series: Patient safety.
[DNLM: 1. Medical Errors. 2. Safety Management. WB 100]
2+
gDC
© 2011 Walter de Gruyter GmbH & Co. KG, Berlin/New York. The publisher, together with the
authors and editors, has taken great pains to ensure that all information presented in this work (pro-
grams, applications, amounts, dosages, etc.) reflects the standard of knowledge at the time of publi-
cation. Despite careful manuscript preparation and proof correction, errors can nevertheless occur.
Authors, editors and publisher disclaim all responsibility and for any errors or omissions or liability
for the results obtained from use of the information, or parts thereof, contained in this work.
The citation of registered names, trade names, trade marks, etc. in this work does not imply, even in
the absence of a specific statement, that such names are exempt from laws and regulations protect-
ing trade marks etc. and therefore free for general use.
Printed in Germany
U Printed on acid-free paper.
www.degruyter.com
Dedicated to
My wife and children, for teaching me the value of family
Contents
Acknowledgments ............................................................................................... ix
About the author ................................................................................................. xi
Abbreviations ...................................................................................................... xiii
This book comes out of ongoing work at the University of Saskatchewan and Royal
University Hospital, Saskatoon Health Region, Saskatoon, Saskatchewan, and related
projects devoted to addressing medical errors, patient safety, and quality programs
across the fields of medicine and health care.
I wish to express my sincere gratitude to my friends and colleagues at the Department
of Pathology and Laboratory Medicine at the University of Saskatchewan and Saska-
toon Health Region who have offered support and valuable assistance to this project.
This project has benefited from helpful discussions at various scientific meetings, panel
presentations, and colloquiums and from the thoughts of students in different classes.
Gratitude must be extended to Dr. H. E. Emson, Professor (Emeritus) and former Depart-
ment Head, for his constant guidance, encouragement, and personal concern during
the early stages of my career. I wish to thank Diane Collard, Annamae Giles, Cheryl
Booth, Lyndon Entwistle, Maxim Gertle-Jaffe, Ajay Nayar, and Amith Mulla for their
helpful suggestions, as well as for their interest in this work. The author would like to
take this opportunity to thank Heather Neufeld, and Lisa Worobec for their professional
work diligence in reviewing the complete manuscript at different stages and in offer-
ing critical suggestions and comments. My thanks are also due to Bill Gray and Todd
Reichert for their assistance in the art work and graphic design.
About the author
The prevalence of medical errors in health care systems has generated immense
interest in recent years, despite being ever-present in health care delivery to some
degree. The research has consistently revealed high rates of adverse events in hos-
pitalized populations. Some of these adverse events result from medical errors, and
a majority of these errors may be preventable. These errors can occur anywhere and
at any time in health care processes. The consequences of these errors may vary
from little or no harm to patient fatalities. It is important to recognize that a degree
of error is inevitable in any human task, and human fallibility in health care should
be accepted. The underlying precursors for many of these human errors may be
latent systemic factors inherent in today’s increasingly complex health care system.
The focus of adverse event analyses on individual shortcomings without appropri-
ate attention to system issues leads to ineffective solutions. The cognitive influence
on medical decision making and error generation is also significant and should not
be discounted.
1.1 Introduction
Health care processes are increasingly being implicated in causing harm to patients.
Medical errors and adverse events are primarily responsible for this harm. These errors,
which may occur at every level of the system, are both common and diverse in nature.
The magnitude and potential for errors is enormous and is increasingly threatening to
become a regular feature of the health care system.
It was precisely with these concerns in mind that Hippocrates coined the maxim
“First, do no harm.” Though it was phrased more than 2,000 years ago, can anyone
question its wisdom and relevance today? A little more than a decade ago, it was gener-
ally perceived that adverse outcomes from medical care or interventions were relatively
rare and isolated incidents. Patients enjoyed a perception that health care professionals
performed nearly perfectly in a perfect health care system.
In recent times, however, this sense of perceived safety within the confines of health
care processes has been challenged. In a survey commissioned by the National Patient
Safety Foundation at the American Medical Association (AMA), the general public per-
ceived the health care environment as “moderately” safe. The survey further indicated
that the health care environment was rated a 4.9 on a scale of 1 to 7, where 1 is not safe
at all and 7 is very safe. The health care environment fared badly in comparison with
airline travel and workplace safety, which were rated higher. There are other reports in
the medical literature that indicate that every patient subjected to medical intervention
may not necessarily face a beneficial outcome (Brennan et al., 1991; Kohn et al., 2000;
Vincent et al., 2001; Wilson et al., 1995). This finding is not new and was reported
close to four decades ago by Schimmel (1964). In 2000, the Institute of Medicine (IOM)
2 1 An overview and introduction to concepts
report To Err Is Human: Building a Safer Health System went to the extent of terming
medical errors a public health risk (Kohn et al., 2000).
action may be labeled erroneous if it results in an adverse outcome. This dilemma may
not have simple answers, nor is answering this question the principal objective; the
objective instead is reducing the incidence of adverse outcomes by practicing more
patient-safe medicine.
Because of the complexity of health care and the current rate of adverse events, it
must be mentioned that health care cannot be compared with other high-risk industries
(see fFig. 1.1), such as aviation and nuclear energy, both of which boast lower levels
of risk (Amalberti et al., 2005; Law, 2004; Leape, 1994). However, because of the safer
outcomes of these other sectors, they may be useful industries to borrow from in terms
of safety programs and policies. Reason (1990) noted that health care is being delivered
in an environment in which there are complex interactions among many variables, such
as the disease process itself, the medical staff and equipment, the infrastructure, and
organizational policies and procedures. Continuous advances in the field require that
medical professionals constantly familiarize themselves with new equipment, pharma-
ceuticals, and methods. Moreover, each component of the medical process is intrinsi-
cally complicated at the same time that the performance of each part is affected by the
functioning of other component parts within the system. Thus, even small errors can
result in larger system consequences. Unlike many other industries, health care sectors
do not enjoy the luxury of well-defined processes. Health care professionals function in
a dynamic environment. Some health care services, such as the emergency department
(ED), where aggressive interventions are often necessary, suffer the most from these
liabilities. Some of the key decisions made in EDs are taken in split seconds and are
frequently based on little or no prior medical information. It is therefore not surprising
that the rates of error are the highest in such challenging environments.
HIGH RISK
(>1/1000)
MODERATE
RISK
Lives Lost/ Year
HEALTH
CARE
MINIMAL
Driving RISK
(<1/100,000)
Bungee Chemical Commercial
jumping Manufacturing Aviation
Nuclear Power
Number of Encounters
The questions remaining to be answered are, in essence, these: what are the causes
for these errors, and what can be done to prevent them? The promotion of patient safety
may require greater integration of concepts evolving from cognitive psychology, human
and systems re-engineering, bioethical viewpoints, social perception, and legislative
actions. This book intends to address such topics.
Thomas et al. 2000 15,000 non- 2.9 ± 0.2% 6.6 ± 1.2% Medication and
psychiatric Mean ± 2SD Mean ± 2SD surgical
discharges procedures
in Utah and
Colorado states
8%, respectively. In New Zealand hospitals, an adverse event was recorded in at least
10.7% of the 1,326 medical records reviewed (Davis et al., 2001). Although the results
were generated by an audit study of three public hospitals for admissions in 1995, these
results were comparable with the QAHC study. In a study by O’Hara and Carson (1997)
done in Australia on routine patient data collection, adverse events were reported in
5% of separations, with the incidence increasing with patient age. Schioler et al. (2001)
studied the incidence of adverse events in Denmark and reported that 9% of all admis-
sions were associated with adverse events. Though the adverse events recorded in the
study caused a prolonged hospital stay, most resulted in minor and transient disabilities,
while a few deaths or permanent disability were recorded. Another study based on
data from New Zealand hospitals reported that the proportion of hospital admissions
associated with an adverse event was nearly 13% (Davis et al., 2002).
Another study evaluated the effects of adverse events on patients after discharge from
the general medical service of a tertiary-care academic hospital (Forster et al., 2003).
Forster et al. (2003), in a prospective cohort study, interviewed a total of 400 patients
an average 24 days postdischarge. The authors reported that at least 76 patients (19%)
had symptoms from medical care. They further classified adverse events into two types:
preventable events caused by an error and ameliorable events, in which the severity
of the event could have been decreased. Each of these subtypes accounted for 6% of
the total incidence reported. The authors emphasized the importance of focusing on
ameliorable events, as these events, though unpreventable, had the potential of being
less severe provided early corrective actions were initiated. It was also suggested that
these ameliorable events were relevant to patient safety, especially in the postdischarge
period.
The adverse events recorded in the previously discussed studies are widely distrib-
uted and are not restricted to any particular specialty or subspecialty. A careful review
of the literature, however, indicates that procedure-related and drug-related events were
the most risky for patients. The Harvard Medical Practice Study II reviewed the nature of
adverse events in hospitalized patients in New York State in 1984 (Leape et al., 1991).
The authors of this study reported that nearly half of all adverse events were associated
with an operation. The authors also reported that drug-related complications were the
single most common type of adverse event. The QAHC study reported that more than
half of all the adverse events recorded were associated with an operation and nearly
11% of adverse events were attributed to drug-related complications. Other studies
have reported similar findings (Davis et al., 2001; Thomas et al., 2000; Vincent et al.,
2001). Gawande et al. (1999), however, observed that adverse events were as common
in nonsurgical care as in surgical care. Though 66% of all adverse events in the study
by Gawande et al. were related to surgical care, it was suggested that this might be the
result of a high number of surgical encounters or consultations in admitted patients.
Another study done with the sole objective of studying adverse event rates for surgical
patients in Australia found a rate of approximately 22% for surgical admissions, with
13% of these patients suffering permanent disability and 4% of the events having fatal
consequences (Kable et al., 2002).
In Canada, studies conducted by Baker et al. (2004) and Forster et al. (2004) detected
that 7.5% and 12.7% of hospitalized patients were affected by an adverse event, re-
spectively, and they deemed 2.8% and 4.8%, respectively, of these adverse events to
have been preventable. In Baker et al.’s study, the most common causes of medical error
1.3 Magnitude and epidemiology of health care errors 7
were delayed diagnoses of cancer and heart disease, drug overdoses, communication
errors, and operative errors. According to Forster et al. (2004), the principal causes were
erroneous delivery of medication, operative complications, and nosocomial infections.
There are two other areas on which the epidemiological research on adverse events
and medical error agree: that these events are preventable and that the frequency of
adverse events rises with advanced patient age. With respect to preventability of the
adverse events, a majority of the authors suggested that the adverse events were indeed
preventable (Gawande et al., 1999; Leape et al., 1991; Vincent et al., 2001; Wilson
et al., 1995). Davis et al. (2001), however, had a different finding. They reported that
for 60% of adverse events the evidence for preventability was either low or nonexis-
tent. The other wide point of consensus in the literature on adverse events and patient
safety is that advanced age serves as a risk factor (Brennan et al., 1991; Kable et
al., 2002; O’Hara and Carson, 1997; Wilson et al., 1995). Only one study failed to
demonstrate this risk (Davis et al., 2002). Thomas and Brennan (2000) exclusively stud-
ied the incidence and types of preventable adverse events in elderly patients. These
investigators estimated that adverse events occurred in 5.29% of all of elderly patients
discharged in the two states of Utah and Colorado but in only 2.8% of nonelderly
patients. Moreover, the results of the study revealed that preventable adverse events
were more common in the elderly than in their younger counterparts. Many factors
contribute to this increased incidence of adverse events among the elderly – chiefly
their comorbid illnesses, complexity of care, and increased exposure to risk.
Although the reported studies may have applied rigorous methodologies and precisely
defined adverse events, various shortcomings in their study methods can be identified.
One of the larger studies (Thomas et al., 2000) and some smaller studies (Davis et al.,
2002; O’Hara and Carson, 1997) that estimated rates of adverse events remained silent
on the important factor of preventability. It would have been appropriate for the stud-
ies to include well-defined criteria for which adverse events were considered “highly
preventable,” “moderately preventable,” and “totally unpreventable” in their report of
their findings. These categories should have been assessed for inter-reviewer reliability
to minimize the subjective bias inherent in many of these studies. It is important to note
that only some studies included assessment of preventability of adverse events as part of
the original study design (Davis et al., 2001; Schioler et al., 2001; Vincent et al., 2001;
Wilson et al., 1995). In at least two of these studies, more than half of the adverse events
were judged to be preventable (Vincent et al., 2001; Wilson et al., 1995). One of the
studies, though, as acknowledged by the authors themselves, was a small pilot study
made up of cases mainly from high-risk surgical specialties, which may have higher
rates of adverse events than other specialties (Vincent et al., 2001). The HMP Study II
data were analyzed for preventability of adverse events, and the authors suggested that
many of the adverse events were neither preventable nor predictable based on the
current state of medical knowledge (Leape et al., 1991).
Sox and Woloshin (2000) also expressed doubts about the methods used by Thomas
et al. (1999) in classifying nearly half the adverse events as preventable as their study
reviewed the summary of the events and not the entire medical record. The methods
used to determine the effect of adverse events on eventual patient outcomes are also
questionable. The IOM report based its findings on two studies, the HMP study and
the study by Thomas et al. (2000). The data obtained in both studies were derived from
hospitalized patients, who were at greater risk of health problems or death than the
8 1 An overview and introduction to concepts
nonhospitalized population. Also, some studies (Brennan et al., 1991; Wilson et al.,
1995) have reported large fatality rates due to adverse events but failed to explain
whether the patients would have lived longer in the absence of adverse events.
Hayward and Hofer (2001), whose study reported a 6% fatality rate due to subopti-
mal care, recently evaluated this aspect. They noted, however, that, after consideration
of the prognosis and other factors in these patients, they estimated that only 0.5% of the
deceased would have gone on to live for three months or more. These findings can be
questionable on an ethical basis. The high figures cited in the IOM report have also been
previously questioned and termed unsubstantiated (Sox and Woloshin, 2000). These
studies failed to achieve conclusiveness in demonstrating that the adverse events resulted
directly from the error. The generalization of the rate of adverse events across North
America may also be impractical, particularly in view of recent findings that sociode-
mographic variables may modify the risk for complications of medical care (Villanueva
and Anderson, 2001). The other limitation in many of these studies is that their data
were primarily based on hospitalized care and adverse events that were documented
only in medical records. Many other patient care areas, including physician offices and
outpatient departments, have not been considered.
It is unfair to suggest solely on the basis of rates of adverse events that patient safety
has deteriorated over the past few decades. If the rate of adverse events is an indica-
tor of patient safety, then the indications are that patient safety is steadily attaining
greater levels. This can be assumed because the HMP Study I, which reviewed more
than 30,000 hospital records from 1984, reported an adverse event incidence of 3.7%,
whereas older studies gave much higher rates, varying from 20% in the 1960s (Schim-
mel, 1964) to 4.6% of all hospitalizations in the 1970s (California Medical Association,
1977). There may be some limitations to the HMP Study I, however, as it accounted only
for adverse events that occurred in hospitalized care and those documented in medical
records. Some have questioned the actual number of deaths due to adverse events and
substituted their own estimates (Hayward and Hofer, 2001). Such statistics demonstrate
a need for improvement – despite their uneven conclusions – and reflect an increase in
interest and attention to patient safety.
1.4 Conclusion
Setting statistics and figures aside, the core of the issue remains that patient safety is at
peril in today’s health care system. Protecting patients from iatrogenic harm is not only
the principal duty of every health care professional but also an ethical and moral re-
sponsibility. The responsibility is owed simply because of the trust placed by patients in
health care professionals. This is a period of crisis and horrendous problems for health
care. The debate is neither about the number of fatalities or injuries nor about the cost
estimates of adverse events. It should instead be focused on how best to usher in a
culture of safety by changing mindsets, understanding error mechanisms, and devoting
resources towards improvements in this area.
Present and past approaches to patient welfare have demonstrated serious over-
sight in, and a blindness of faith toward, medical processes and professionals. Across
contexts – nations and communities – it is apparent that individuals have only limited
affect upon systemic issues. These are what need to be addressed.
References 9
The reviewed literature on adverse events may not truly reflect the current state
of health care quality and patient safety issues because of some of the methodologi-
cal flaws outlined earlier. However, the core issue remains that current health care
processes are not performing at their potential best. There is tremendous scope for
quality improvement in health care delivery.
References
Amalberti R, Auroy Y, Berwick D, Barach P. Five system barriers to achieving ultrasafe health
care. Ann Intern Med 2005;142:756–764.
Baker GR, Norton PG, Flintoft V, et al. The Canadian Adverse Events Study: the incidence of
adverse events among hospital patients in Canada. CMAJ 2004;170:1678–1686.
Bates DW, Spell N, Cullen DJ, et al. The costs of adverse drug events in hospitalized patients.
Adverse drug events prevention study group. JAMA 1997;277:307–311.
Brennan TA. The institute of medicine report on medical errors – could it do harm? N Engl J
Med 2000;342:1123–1125.
Brennan TA, Leape LL, Laird NM, et al. Incidence of adverse events and negligence in hos-
pitalized patients. Results of the Harvard Medical Practice Study I. N Engl J Med 1991;
324:370–376.
Brune v. Belinkoff, 354 Mass. 102 (1968).
California Medical Association. Report of the Medical Insurance Feasibility Study. San Fran-
cisco: The Association; 1977.
Charatan F. Clinton acts to reduce medical mistakes. BMJ 2000;320:597.
Davis P, Lay-Yee R, Briant R, Ali W, Scott A, Schug S. Adverse events in New Zealand public
hospitals I: occurrence and impact. N Z Med J 2002;115:U271.
Davis P, Lay-Yee R, Schug S, et al. Adverse events regional feasibility study: indicative find-
ings. N Z Med J 2001;114:203–205.
Elder NC, Dovey SM. Classification of medical errors and preventable adverse events in pri-
mary care: a synthesis of the literature. J Fam Pract 2002;51:927–932.
Forster AJ, Clark HD, Menard A, et al. Adverse events among medical patients after discharge
from hospital. CMAJ 2004;170:345–349.
Forster AJ, Murff HJ, Peterson JF, Gandhi TK, Bates DW. The incidence and severity of
adverse events affecting patients after discharge from the hospital. Ann Intern Med
2003;138:161–167.
Gawande AA, Thomas EJ, Zinner MJ, Brennan TA. The incidence and nature of surgical ad-
verse events in Colorado and Utah in 1992. Surgery 1999;126:66–75.
Hayward RA, Hofer TP. Estimating hospital deaths due to medical errors: preventability is in
the eye of the reviewer. JAMA 2001;286:415–420.
Hofer TP, Kerr EA, Hayward RA. What is an error? Eff Clin Pract 2000;3:261–269.
Kable AK, Gibberd RW, Spigelman AD. Adverse events in surgical patients in Australia. Int J
Qual Health Care 2002;14:269–276.
Kohn LT, Corrigan JM, Donaldson MS, editors. Committee on Quality of Healthcare in Amer-
ica, Institute of Medicine. To Err Is Human: Building a Safer Health System. Washington,
DC: National Academy Press; 2000.
Law R. Safety of dietary supplements: Figure 5: How safe is safe enough? http://www.laleva.
cc/petizione/english/ronlaw_eng.html. Accessed November 1, 2010.
Leape LL, Error in medicine. JAMA 1994;272:1851–1857.
Leape LL, Bates DW, Cullen DJ, et al. Systems analysis of adverse drug events. ADE prevention
study group. JAMA 1995;274:35–43.
10 1 An overview and introduction to concepts
Leape LL, Brennan TA, Laird N, et al. The nature of adverse events in hospitalized patients.
Results of the Harvard Medical Practice Study II. N Engl J Med 1991;324:377–384.
O’Hara DA, Carson NJ. Reporting of adverse events in hospitals in Victoria, 1994–1995. Med
J Aust 1997;166:460–463.
Oxford Dictionary of Current English. Thompson D, editor. New revised ed. Oxford: Oxford
Univ. Press; 1998:294.
Reason JT. Human Error. Cambridge: Cambridge Univ. Press; 1990.
Schimmel EA. The hazards of hospitalization. Ann Intern Med 1964;60:100–110.
Schioler T, Lipczak H, Pedersen BL, et al. Incidence of adverse events in hospitals. A retrospec-
tive study of medical records. Ugeskr Laeger 2001;163:5370–5378 [Article in Danish].
Sox Jr. HC, Woloshin S, How many deaths are due to medical error? Getting the number right.
Eff Clin Pract 2000;3:277–283.
Steel K, Gertman PM, Crescenzi C, Anderson J. Iatrogenic illness on a general medical service
at a university hospital. N Engl J Med 1981;304:638–642.
Thomas EJ, Brennan TA. Incidence and types of preventable adverse events in elderly patients:
population based review of medical records. BMJ 2000;320:741–744.
Thomas EJ, Studdert DM, Burstin HR, et al. Incidence and types of adverse events and negli-
gent care in Utah and Colorado. Med Care 2000;38:261–271.
Thomas EJ, Studdert DM, Newhouse JP, et al. Costs of medical injuries in Utah and Colorado.
Inquiry 1999;36:255–264.
Villanueva EV, Anderson JN. Estimates of complications of medical care in the adult US popu-
lation. BMC Health Serv Res 2001;1:2–10.
Vincent CA. Research into medical accidents: a case of negligence? BMJ 1989;299:1150–1153.
Vincent CA, Neale G, Woloshynowych M. Adverse events in British hospitals: preliminary
retrospective record review. BMJ 2001;322:517–519.
Wilson RM, Runciman WB, Gibberd RW, Harrison BT, Newby L, Hamilton JD. The quality in
Australian health care study. Med J Aust 1995;163:458–471.
2 Perceptions of medical error and adverse events
2.1 Introduction
Even if perceptions of medical error differ, one may wonder why it is important to
understand how perceptions differ among participants in the health care system. One
might think that a medical error is a medical error, no matter how it is perceived. This,
however, is not strictly true. How people perceive something and what they believe to
be true impacts how they interact with, apply, and are affected by something. No matter
how theorists define medical error, perceptions play a critical role in creating its “real-
ity,” so, in practice, the characteristics and impacts of medical error are shaped by how
medical error is perceived. How medical error is perceived, in turn, is constantly being
redefined by what is identified as an error and how it is explained in the medical field.
For this reason, perception of medical error can determine the efficacy of initiatives
that are implemented to improve patient safety. Bosk (2005) suggests that the IOM’s
report completely ignores a certain social-science approach to analyzing error, one
that looks how workers define error “on the shop floor.” Bosk argues that the strategy of
understanding perceptions is absent from current research; without it, no policymaker
can hope to reduce the frequency of medical error. By understanding the ways in which
perceptions of medical error can differ, researchers and policymakers can become more
aware and thus create better policy.
Another reason why it is important to understand perceptions of medical error is
that doing so can help researchers develop an awareness of the barriers to build-
ing a culture of safety, where the minimization of medical errors is the top priority.
Interpreting the perspectives of front-line workers with direct but different experiences
can help to pinpoint problems and possible solutions to them that might otherwise
go unremarked. As Edwards et al. (2008), citing Nieva and Sorra (2003), point out,
“assessing staff perceptions of safety culture is a first step to identify areas in need
of improvement, and then, follow-up improvements can show the effectiveness of
changes made to address initial improvement needs.” Furthermore, knowing how
12 2 Perceptions of medical error and adverse events
perceptions of medical error differ among physicians, nurses, and medical students
can help researchers to understand the context in which safety initiatives are being
implemented and the mindset of the people who will be performing them. Moreover,
because perceptions are not static, Edwards et al. (2008) conclude that researchers
can effectively understand the efficacy of different safety initiatives in terms of moving
medical environments toward safety cultures by analyzing changes in perceptions
over time. The importance of this move is highlighted by recently published studies
that show that physicians are still uncomfortable with reporting medical errors, even
though a large part of the overall push in the medical community toward improved
patient safety has been the establishment of nonpunitive reporting systems (Anderson
et al., 2009). Without having effective reporting systems in place, it is impossible to
accurately measure and address medical error. Therefore, researchers need to have a
detailed understanding of the effectiveness of patient safety initiatives, rather than just
relying on medical error statistics. Measuring efficacy in terms of the propagation of
a safety culture, which builds on and develops the perceptions and, eventually, the
practices of medical workers, provides researchers with more nuanced information
than they would have if they were just to measure the efficacy of different safety
initiatives in terms of a strict reduction in medical errors.
If researchers want to truly understand medical error, they must first have an under-
standing of the different ways in which medical error is perceived by those who, through
their actions, are constantly redefining the part medical error plays in the medical and
health care field. If policymakers want to put into place effective policies for combating
medical error, they must first have an understanding of the different ways in which medi-
cal error is perceived by those who will enact and be affected by those policies. Essen-
tially, by striving to understand the perception of those who must effectively negotiate the
reality of medical error, they are doing no less than enhancing their own perception.
actions of an individual, then physicians are more likely to consider it an error (Elder
et al., 2004).
The findings of Elder et al. (2004) draw attention to three discrepancies between the
generally accepted definitions of medical error and the way in which physicians define
an error. The Institute of Medicine (IOM) definition identifies “errors of execution” and
“errors of planning” (Kohn, 2000). Neither of these types of error depends on outcome;
each is defined solely on the basis of what action was taken. If a physician perceives
only those medical events that are harmful as error, she is unlikely to also recognize ac-
tions that could be potentially harmful in the future. The IOM definition of medical error
also potentially differs from the perception of physicians in that physicians are more
likely to define those events that are rare occurrences as an error, whereas the IOM does
not distinguish between rare and common occurrences. Furthermore, physicians are
more likely to consider an event as an error if they perceive it to be the responsibility
of an individual, whereas the IOM draws attention to the fact that errors are most com-
monly perpetuated through systemic causes. These two things may be related in that
errors caused by the system are generally less recognizable and therefore may be seen
as common, unavoidable occurrence, whereas individual errors are easier to identify
and therefore stand out when they do occur.
The danger in these discrepancies is that an error, whether or not it is harmful,
whether it is common or rare, and whether it results from individual or systemic cause,
ultimately leads to suboptimal health delivery at the very least; at the most, it can lead
to the occurrence of large serious errors. This notion is explored further in chapter 3.
Furthermore, given that physician reporting is the foundation of many medical error pre-
vention initiatives, it is important that the physician perception of medical error match
more complex formulations of medical error. While a physician might have difficulty in
addressing systemic errors alone, the effectiveness of a larger task force in addressing
these systemic errors depends on the physician’s ability and willingness to recognize
and report those errors that come from the system (Hobgood et al., 2005).
emotional and psychological harm to the patient and his family (Mazor et al., 2010).
Therefore, whether or not a medical system-defined error has actually taken place,
perceived medical error is an important part of determining how effective or satisfac-
tory medical care has been. It follows, then, that educating patients and their families
as to what to expect during their experience with medical services – whether or not
they experience an actual medical system-defined error – can help prevent unsatisfac-
tory medical experiences. Further, educating patients about how to observe their own
bodies can help to lessen medical error by helping patients become more part of the
monitoring and reporting team. This may have the side benefit of making patients feel
more responsible and in control of their medical experience, thus, increasing patient
satisfaction.
The patient perspective demonstrates how understanding differing perspectives of
medical error must be developed if all parties involved in health services are to work to-
gether to improve patient safety. Even patients have a role to play in lessening the occur-
rence and impact of medical error. A shared understanding of what constitutes medical
error can be built through, as Burroughs et al. (2007) specifically recommend, patient
education programs “to address the fears and concerns of each patient.” It is crucial,
however, to understand why patients’ perceptions of medical error are so different from
those of medical practitioners.
Public perception of the safety of the medical system, as influenced by media, friends
and family, and experiences, can lead patients who come to the hospital to expect to face
a medical error. Even if no error takes place during a patient’s medical experience, if he
perceives an error to have taken place, his overall satisfaction with the experience will
be altered. Some researchers argue, however, that, rather than including everything the
patient considers to be medical error in a standard definition of medical error, medical
caregivers must search for the root cause of why patients are so broadly defining medical
error. Patti Muller-Smith, in an interview with Hospital Peer Review, draws attention to
why this is important when she says that it would be a mistake to expand the definition
of medical error to fit that of patients (Hospital Peer Review, 2007). She contends that
doing so would obscure the focus on making changes that would prevent significant
medical errors, particularly those that have an impact on patient safety. Still, it is hard
simply to maintain that tight focus on medical system-defined error without understand-
ing patients’ perception of their experience, as this is dependent upon communication.
Postmedical-experience surveys of patients already assist in measuring whether patients
feel that their needs and concerns were adequately addressed. It is more difficult, how-
ever, to know whether patients feel that they are part of a satisfactory communication
process. Even if the patient does not realize it, she might not be receiving the communi-
cation necessary to understand what exactly to expect of her hospital experience.
The solution, then, is to work to ensure the strength and effectiveness of patients’
communication with their caregivers; in particular, patients need to be educated so
that their expectations of the medical process match those of their caregivers, enabling
them to comprehend more clearly the medical system understandings of medical error.
Furthermore, there must be processes during a patient’s medical experience through
which they can communicate whether they feel they are being adequately cared for
and, if not, why they feel unsafe. This way, health service providers can work to address
widespread sources of patient anxiety and increase the chances that patients will feel
satisfied after the experience.
2.4 Perceptions by health care staff 15
Ultimately, the best way to ensure that patients are not anxious about the possibility
that a medical error will occur is to lower the occurrence of medical errors. Previous
experience with medical error, from either personal experience or from exposure to the
media, can intensify their expectation and assessment of error. Coverage of the most
extreme medical errors in the media, for example, can lead patients who come into the
hospital to expect to face an error. If patients expect an error, they are more likely to see
an error in anything that seems abnormal to them.
Several researchers have shown how previous exposure to medical error can result in
patient and public anxiety. Northcott et al. (2008) report that those people who believe
that either they themselves or a family member has experienced a preventable medical
error are more likely to overestimate the frequency with which medical errors occur,
have less trust that their doctor would report a medical error, and have a lower opinion
of the health care system overall. Mazor et al. (2010) found that perceived medical
errors have strong impacts on patients and their families, and Mira et al. (2009) conclude
that “risk perception for adverse events increases after having suffered such an event.”
There is the potential, then, for a somewhat self-perpetuating cycle in which, as pa-
tients become increasingly anxious about the medical system, they are more likely to
perceive a situation as being part of a medical error, therefore feeding their anxiety. This
cycle might not be as vicious if patients were shown the gravity with which the medical
field treats medical error. Mazor et al. (2010) found that, in the event of an error, the in-
volved medical practitioner’s acknowledgment of that error often tempered the patient
and family’s subsequent view of the medical system, especially if it was accompanied
by an apology. Patient and public perception shows that those in medical services not
only must work to reduce actual medical errors but also must work to rebuild the public
image of their field so that their patients once again recognize it as consistently and
reliably safe.
Besides addressing the impact of perceived medical error, respecting patient and fam-
ily perspectives could actually assist in avoiding those medical errors that fit theorists’
definitions. Mazor et al. (2010) suggest that parents and other family members might
have a better understanding of the patient and might even help prevent medical errors,
for example, by questioning misdiagnoses. By regarding patient perception as valuable
and worthy of consideration, medical practitioners, policymakers, and theorists can all
take steps to lessen the instances and effects of error.
their high level of exposure to patients, nurses have the potential to play an important role
in the identification of medical errors and the promotion of a culture of safety. Promoting
a culture of safety is seen as one of the most effective ways of preventing medical error.
A culture of safety depends on a group of professionals who, both individually and as a
group, are willing to dedicate themselves to prioritizing the prevention of medical error as
part of their pursuit of improved patient safety. Part of this group dedication includes the
ability and the willingness to learn from each other. Nurses often work alongside other
medical professionals, particularly physicians, and the cooperative and supplementary
nature of their work puts them in an ideal position to observe and provide feedback for
those other professionals working with them.
The ideal advantage that nurses may enjoy in terms of preventing medical error may
conflict with the real pressures nurses face in their work. As Elder et al. (2008) discov-
ered, for example, nurses perceive that there are often too many barriers for them to
have a useful function in reducing medical error. Despite the fact that, as described
earlier, medical error reduction is often promoted as a product of teamwork, nurses
often feel that they are at risk of straining their relationships with their coworkers if they
are discovered to have reported a medical error (Elder et al. 2008).
These barriers to medical error reporting mean that a nurse might not report a medical
error if no harm actually befalls the patient (Elder et al. 2008). One can see a similarity
between a nurse’s reluctance to report a medical error that does not result in patient
harm and the way in which physicians decide whether or not an event constitutes a
medical error on the basis of whether or not harm occurs (Elder et al. 2004). Harm
occurrence, as a limiting factor to whether or not a nurse perceives medical error as
worth the effort and the cost to work relationships that would come from her reporting
it, means that medical errors that are part of the system are unlikely to be identified until
they do result in serious harm. Even though medical error reporting has a potential to be
proactive, this limitation often relegates medical error to the purely reactive.
Besides recognizing and reporting errors made by those with whom they work, nurses
also have to recognize and decide when to report their own errors. One can see how
self-reporting could be a stressful process to go through. Elder et al. (2008) describes
nurses who see making an error as a deeply shameful thing to do. However, they also
describe nurses who hold a great deal of trust and respect for those fellow nurses, whom
they see as being honest and trustworthy enough to report their own errors.
The latter perspective especially makes sense in the context of medical error, given
that the majority of medical errors have been attributed to systemic causes. Self-
reporting should not be seen as shameful for two reasons. First, if medical errors are
caused by a systemic flaw, then the vast majority of time an error cannot be attributed
to an individual. If a medical error should bring shame on anyone, more often than not
that shame should accrue to the entire system. The second reason that self-reporting
should not be shameful is that doing so shows a prioritization of the success of the entire
medical system, since a report of an error associated with a particular person can assist
in the identification of systemic problems. Both of these reasons can also apply to why
a nurse’s colleague should not be embarrassed or angry if that nurse reports one of his
errors; that error likely is not the individual’s fault, and identifying that error is a boon
that will help all medical professionals in the system do their jobs better.
If nurses feel uncomfortable pointing out errors directly to a peer, an anonymous
reporting system can assist in avoiding this discomfort. On the other hand, sometimes
2.5 Perceptions by medical students 17
even an anonymous report can be easily traced back to a single nurse, in cases, for
example, where that nurse was the only other person present when the error took place.
In these cases, it is important to have not only an anonymous reporting system but also
a culture in which the nurses share a perspective that medical error reporting is a lauda-
tory process in which to participate. Such a perspective becomes even more important
in situations where nurses identify the errors of those who they might feel are higher in
the hierarchy.
Despite the fact that nurses often work closely with physicians and residents, Elder
et al. (2008) found that nurses rarely report a physician’s or resident’s medical error.
Nurses describe having to find ways to draw physician’s attention to errors without
embarrassing them by forthrightly pointing them out. Again, and again problemati-
cally, nurses found it worth the risk of a negative reaction from a physician only if they
saw that a patient was or had been immediately in harm’s way. While there is a well-
documented hierarchal tension between nurses and physicians, one hopes that it can
be overcome by awareness that medical error should be considered irrelevant to status
struggles between individuals; it is the product of a faulty system.
In the section of this chapter on physicians, we suggested that physicians are less
likely to identify medical error that stems from systemic causes than that which results
from individual error. Nurses, however, have been shown to recognize the relationship
between organization and patient safety and therefore have the potential to identify
systemic causes of medical error. Ramanujam et al. (2008) conclude that nurses’ per-
ceptions of patient safety conditions in their workplaces change as their work demands
change. As the demands on nurses increase, their perceptions of how safe their work-
places are go down. Nurses see higher work demand as coming from work overload,
coupled with exhaustion and the inability to have personalized relationships with pa-
tients. Higher work demand is perceived by nurses as correlated with decreased patient
safety (Ramanujam et al. 2008). Given that exhaustion and overload are two commonly
identified issues in workplace organization and structure (see chapter 3), it appears that
nurses’ perceptions could be valuable in flagging certain systemic circumstances that
could lead to medical error.
As Edwards et al. (2008), citing Ralston, et al. (2005), point out, a medical error is
often only apparent to those who are providing the care during which the event occurs,
and whether or not the error is reported depends on the people who committed it.
and may well be more likely to eventually innovate new creative solutions to patient
safety problems. Medical students represent the leading edge of the medical field. A
change in medical students’ perception of error would be indicative of the beginning of
a shift in perception in the larger field.
Even as early as 2003, Schenkel et al. showed that residents have some understanding
of what constitutes medical error. In their study, residents in the emergency department
attributed responsibility for medical error primarily to the overall emergency depart-
ment team (Schenkel et al., 2003). While they also identified the actions of individuals
as a contributing factor, their primary focus on the entire department displays an under-
standing of the systemic nature of error. Schenkel et al. (2003) noted that residents were
more likely to recognize diagnostic errors, which they suggest might be a result of the
educational focus on differential diagnosis, a technique where physicians make a list of
diagnoses and then eliminate diagnoses from the list until one remains. If none remain,
it suggests that a misdiagnosis may have occurred. In other words, residents may have
had limited awareness of medical error because of standard curricula, but the residents
did not yet have an indication that patient safety–based education had been effective.
Medical education is the most obvious way students’ perceptions of medical error
are shaped. Muller and Ornstein (2007) found that students with clinical training were
more likely to say that their views of medical error had changed over time than were
students without such training, though it is unclear how these perceptions had changed.
What this indicates is that educators have to be prepared to adapt the way in which
they teach students to integrate a certain view of medical error into their perspective
depending on whether they are in the classroom or in the field.
At the same time, it is insufficient to regard medical education as the sole vari-
able affecting student perspective on medical error. While medical education formally
stresses the importance of error reporting, among other patient safety initiatives, stu-
dents, like physicians and nurses, are subject to a number of contradictory pressures.
In particular, students are participating in a field that is notoriously competitive, where
those who can withstand the demands of the field and manage to succeed in deliver-
ing quality care stand to gain valued recognition, whether it is in the form of a better
residency or better job placement, respect from their peers, or, potentially, a higher
salary. Under certain conditions, it might be more convenient for students to deflect
blame for error that is of uncertain origin. Students might attempt to capitalize on the
focus in the field on systemic error and use it as an opportunity to take advantage of
plausible deniability, shifting attention away from their own potential responsibility.
What is more likely is that in most circumstances there are multiple causes for error;
in different cases, one or another might be given more weight, but generally all have
some degree of effect.
By studying medical students, researchers can examine variables that affect physi-
cians and other medical staff in some of their most intense forms. Given that one of the
primary causes of medical error is fatigue (see chapter 3), those concerned with patient
safety look to work-hour overload as a source of increased error. When it comes to
potential fatigue due to increased work hours, residents represent the extreme in the
medical field. In an effort to give them the maximum opportunity for hands-on train-
ing possible, medical students in residence are often expected to work intense hours,
sometimes more than 80 hours a week. In 2003, the American Accreditation Council
for Graduate Medical Education (ACGME) set a limit on the number of hours that all
2.6 A sociological perception of medical error 19
residents could work: a maximum of 80 hours total a week and a maximum of 30 hours
in continuous shifts (Philibert, 2002, cited in Jagsi et al., 2008). Some researchers, how-
ever, have expressed concern that the effectiveness of these limited hours in reducing
medical error will be counteracted by increased patient workload (Okie, 2007, cited in
Jagsi et al., 2008).
Jagsi et al. (2008) investigated the effect of reduced hours on residents’ percep-
tions of patient safety. They sent surveys to residents in all ACGME-accredited train-
ing programs at two hospitals in Boston, Massachusetts, between 2003 and 2004
and obtained 1,770 responses across the spectrum of specialties (Jagsi et al., 2008).
What they found was that, between 2003 and 2004, when the ACGME’s work-hour
limits were set, a significantly lower number of residents reported committing medical
errors, particularly those caused by “working too many hours” and “carrying or admit-
ting too many patients” (Jagsi et al., 2008). Students whose hours were reduced were
also less likely to identify fatigue as a factor that frequently affected the quality of care
they provided (Jagsi et al., 2008). Their findings show that not only did those residents
whose hours were reduced not have increased workloads, but also they reported fewer
medical errors. The work of Jagsi et al. (2008) is an example of using the perceptions
of medical practitioners as a way of measuring changes to patient safety. Their findings
indicate a degree of success in identifying patient safety initiatives that can be used to
guide further research and policy for others in the medical field.
It is difficult, however, to identify a single variable that affects medical error and
then to generalize it to the broader field of patient safety. Jagsi et al. (2008) also found
that one reason that residents’ patient loads were not increased when their hours were
limited is that those hospitals that had to decrease their residents’ hours also took steps
to ensure that patient overload would not become a problem, such as taking on ad-
ditional residents and physicians. That these hospitals had the foresight to do this is an
example of making organizational adjustments in anticipation of new opportunities for
systemic errors. It illustrates the importance of considering the entirety of the system
when making changes.
What further problematizes the search for particular identifiable perspectives on medi-
cal error is that the students’ perspectives cannot be distilled into one single position.
While there might be observable trends across all students, given that all students follow
similar educational paths, certain core demographic differences can alter perceptions
of medical error. While Muller and Ornstein (2007) found that there is no significant
difference in how male and female students define medical errors or in their willingness
to report errors, they reported that female students are more likely to experience feelings
of guilt after committing a medical error and more likely to feel stress about potential
consequences.
about medical error as early as 1979 in his book Forgive and Remember: Managing
Medical Failure, published a paper called “Continuity and change in the study of medical
error: The culture of safety on the shop floor.” While his article (2005) does not disagree
with the need for a culture of safety and even begins by praising the 1999 IOM report for
its role in bringing public attention to the issue of medical error, Bosk does believe that
the report focuses on one way of understanding error while ignoring another.
Bosk’s description of how the IOM interprets medical error should be familiar after
reading chapter 1. He says that IOM uses “normal accident theory, a blend of organiza-
tional theory, cognitive psychology, and human factors engineering” to explain medical
error (Bosk, 2005). In normal accident theory, systems are “error-prone,” accidents are
“normal,” and individuals do not fail but instead are the victims of inevitable failings in
the system. “Normal accidents” occur because each part of the system is individually
complex while also having an effect on the complex working of the overall system;
small errors therefore have consequences for the whole system (Bosk, 2005).
In “normal accident theory,” error reduction occurs through better system design.
Bosk (2005), however, says that the IOM completely ignores a second social-science
approach to analyzing error, one that looks to understand the workers’ perceptions
of error. The IOM, therefore, is “unmindful of the cultural context of the workplace”
(Bosk, 2005). In other words, the IOM fails to propose solutions that deal with the
cultural realities of the field. Through his work, Bosk is actually calling on researchers,
theorists, policymakers, and, in particular, the Institute of Medicine to change their
perceptions.
2.7 Conclusion
Medical error does not exist outside society. To understand medical error, one must
know not just how it is caused but also how it perceived, especially by those who have
an effect on the prevalence and prevention of error, as well as by those who are affected
by it. Perceptions can help explain why certain patient safety initiatives are or are not
effective, why patient satisfaction is not directly linked to how the health system rates
patient safety, and even how health researchers interpret and apply their findings. It
is not enough to put in place systems for reporting errors without understanding that
all those involved in medical processes have different views on what constitutes an
error. If the IOM simply calls for policy to address error that is systems based, they are
bound to be ineffective, because they fail to address the fact that, in practice, nurses
and physicians tend to attribute errors to individuals. It is insufficient for researchers and
policymakers to study how or how much medical error occurs. For their efforts to be
relevant and valuable, they must study how medical error is perceived. It is a difficult
task, but a necessary one.
References
Anderson B, Stumpf PG, Schulkin J. Medical error reporting, patient safety and the physician.
Journal of Patient Safety 2009;5:176–179.
Bosk, CL, Continuity and Change in the Study of Medical Error: The Culture of Safety on the
Shop Floor. Princeton, NJ: Institute of Advanced Study, School of Social Science, University
of Pennsylvania, School of Social Science. Occasional Papers 2005; Paper 20.
References 21
Bosk C. Forgive and Remember: Managing Medical Failure. Chicago: University of Chicago
Press; 1979.
Burroughs TE, Waterman AD, Gallagher TH, et al. Patients’ concerns about medical errors
during hospitalization. Jt Comm J Qual Patient Safety 2007;33:5–14.
Edwards PJ, Scott T, Richardson P, et al. Using staff perceptions on patient safety as a tool
for improving safety culture in a pediatric hospital system. Journal of Patient Safety
2008;4(2):113–118.
Elder NC, Vonder Meulen MB, Cassedy A. The identification of medical errors by family physi-
cians during outpatient visits. Annals of Fam Med 2004;2(2):125–129.
Elder NC, Brungs SM, Nagy M, Kudel I, Render ML. Nurses’ perceptions of error communica-
tion and reporting in the intensive care unit. Journal of Patient Safety 2008;4(3):162–168.
Hobgood C, Eaton J, Weiner BJ. Identifying medical errors: developing consensus on clas-
sifications and consequences. J Patient Safety 2005;1(3):138–144.
Hospital Peer Review. Patients may define medical errors differently than you. Hospital Peer
Review 2007;32(4):41–42, 47–48.
Institute of Medicine. Keeping Patients Safe. Transforming the Work Environment of Nurses.
Washington, DC: National Academies Press; 2004.
Jagsi R, Weinstein DF, Shapiro J, Kitch BT, Dorer D, Weissman JS. The Accreditation Council
for Graduate Medical Education’s limits on residents’ work hours and patient safety. Arch
Intern Med 2008;168(5):493–500.
Kohn LT, Corrigan JM, Donaldson MS, editors. Committee on Quality of Healthcare in Amer-
ica, Institute of Medicine. To Err Is Human: Building a safer Health System. Washington,
DC: National Academy Press; 2000.
Mazor KM, Goff SL, Dodd KS, Velten SJ, Walsh KE. Parents’ perceptions of medical errors.
J Patient Safety 2010;6(2):102–107.
Mira J, Lorenzo S, Vitaller J, et al. Hospital clinical safety from the patient’s point of view.
validation of a safety perception questionnaire. Rev Med Chile 2009;137:1441–1448.
Muller D, Ornstein K. Perceptions of and attitudes towards medical error among medical
trainees. Medical Education 2007;41:645–652.
Nieva VF, Sorra J. Safety culture assessment: a tool for improving patient safety in healthcare
organizations. Qual Saf Health Care 2003;12: ii17–ii23.
Northcott H, Vanderheyden L, Northcott J, et al. Perception of preventable medical error in
Alberta, Canada. Int J Qual Health Care 2008;20(2):115–122.
Okie S. An elusive balance: residents’ work hours and the continuity of care. N Engl J Med
2007;356(26):2665–2667.
Philibert I, Friedmann P, Williams WT. New requirements for resident duty hours. JAMA
2002;288(9):1112–1114.
Ralston J, Larson E. Crossing to safety: transforming healthcare organizations for patient safety.
J Postgrad Med 2005;51(1):61–67.
Ramanujam R, Abrahamson K, Anderson JG. Influence of workplace demands on nurses’
perception of patient safety 2008;10:144–150.
Ramsey G. Nurses, medical errors, and the culture of blame. Hastings Rep 2005;35:20–21.
Schenkel SM, Khare RK, Rosenthal MM, Sutcliffe KM, Lewton EL. Resident perceptions of
medical errors in the emergency department. Acad Emerg Med 2003;10(12):1318–1324.
3 Causes of medical error and adverse events
The complexity of the health care delivery system may be primarily responsible
for increased error rates. A large proportion of the blame lies not with individual
failures but with system complexities. The second IOM report, Crossing the Quality
Chasm: A New Health System for the 20th Century, states that safety is a systems
property and that patients should be safe from injury caused by the care system
(2001). The report suggests that better reporting of systemic errors will lead to better
system design, minimizing latent system defects and, in turn, medical errors. The
report goes on to elaborate that reducing risk and ensuring safety require greater at-
tention to systems that help prevent and mitigate errors. It is suggested that little can
be achieved in improvements to patient safety unless safer systems are designed
around human factors. Health care systems should be held accountable, rather
than just individuals. The health care system requires coordinated improvements to
overcome problems in system design that threaten quality of care.
3.1 Introduction
It has long been recognized that health care systems are far from safe. To offset the
sources of these deficiencies, health care facilities have employed safety programs since
the beginning of the 20th century. Physicians created the initial model for external qual-
ity oversight in the United States in 1917 (Roberts et al., 1987). This was soon followed
by the establishment of the hospital standardization program of the American College
of Surgeons, which was the forerunner in the United States of both the National Joint
Commission on Accreditation of Health Care Organizations (JCAHO), established in
1951, and the federal and state regulatory models now in place for all types of health
care organizations. The commitment to quality and patient safety by JCAHO has been
periodically strengthened ever since.
The foremost requirement of designing systems in a way that can eventually reduce
the possibility of errors is to allow the identification of errors before significant dam-
age occurs by integrating disciplines, ideas, and various models and theories. Health
care has liberally borrowed concepts from many sectors, particularly industry and so-
cial psychology. Creating safer systems requires more such knowledge transfers with
modifications to suit our needs. It is our belief that for quality improvements and pa-
tient safety initiatives to occur, we will need to absorb various ideas and theories from
human factors reengineering, information technology, organizational and management
sciences, cognitive psychology, and bioethics.
Research has described two kinds of errors, active and latent (Rasmussen and Peder-
sen, 1984). Active errors are those that produce immediate events and involve operators
of complex systems; examples are errors made by pilots in aviation and by health care
professionals in medicine. Latent errors may result from factors that are inherent in the
24 3 Causes of medical error and adverse events
ERROR
HUMAN
Mistakes
Violations
Negligence
Incompetence
WORK ENVIRONMENT
'#$ ' $#
'W! 'ysical agents SYSTEM
ORGANIZATIONS
' $# '!!chy
incentives
' #Policies
ERROR
Fig. 3.1: Influence of system factors on human deficiencies in causing an error (from Cook
and Woods, 1994, modified).
3.1 Introduction 25
system that represents only a small part of the overall problem; therefore, measures
aimed at this end can be minimally effective. Maximum efforts have to be directed at
the latent deficiencies in the system, the base of the pyramid, which will ultimately yield
more positive results than can be obtained by minimizing the active errors at the sharp
end. It is generally agreed that the most immediate and proximate error appears to be
human error, while the background and provoking error are well beyond an individual’s
control (Leape, 1994).
The conventional methodology of error investigations, in which, after a careful ret-
rospective analysis, errors are primarily classified as human failure and the majority
of blame is placed on either human error or mechanical error, must change. These
investigations lead to such shortsighted measures as revising procedures, taking admin-
istrative and regulatory actions, and introducing new rules that add to the complexity
of the process. Thus, increased complexity serves to introduce new types of failures that
present new opportunities for error generating behavior (see fFig. 3.2). It is, therefore,
suggested that health care errors be viewed from a systems perspective to help bring
about effective changes in delivering safe health care.
Attitude and stress are two important system factors that play a major role in influenc-
ing the active errors and that deserve further elaboration. A certain part of responsibil-
ity for the high incidence of errors in health care may be attributed to the attitude of
health care professionals toward error. It has been suggested that physicians tend to
overestimate their ability to function under conditions of stress, such as fatigue and
high anxiety (Helmreich, 2000). This belief among health care professionals that they
ADVERSE
OUTCOME
INCREASED
OPPORTUNITIES RETROSPECTIVE
FOR ERRORS ANALYSIS OF CRITICAL
INCIDENT
NEW TYPES OF
FAILURES
CRITICAL INCIDENT
CLASSIFICATION
INTRODUCTION OF
NEWER COMPLEXITIES
IN THE PROCESS REMEDIAL )
""!"
ACTIONS )% ""!"
PROPOSED )(#$""!"
)
"!%"&#!
) #$"$& $! #
) $!
) $"!%$! ! '$ !!y
can perform flawlessly even at times of extreme stress may prove to be their undo-
ing. Health care professionals are only human and are as susceptible to error as any
others, but there seems to be a prevailing consensus that perfection is their forte and
mistakes are unacceptable. This perception seems to be prevalent not only among the
general public but also within the health care fraternity. It is therefore appropriate that
in his editorial remarks, Blumenthal (1994) calls for a total attitude change toward errors
from both physicians and the general public. Blumenthal also suggests that such efforts
require abandoning traditional cherished myths. Health care, just like any other human-
controlled process, is not 100% safe. It is rather a very risky business, simply because it
is not delivered by a machine: a majority of health care processes are run by, controlled
by, and dependent on human function. It is appropriate for both health care providers
and health care seekers to accept this reality.
It has been more than a half-century since it was reported that stress directly influ-
ences an individual’s thought processes and that, as stress increases, an individual’s
thought processes and attention span narrow (Combs and Taylor, 1952; Easterbrook,
1959). The stressful conditions can due to mental or physical factors. Stress and fatigue
have recently been documented to be the principal contributors to many accidents
(Dinges et al., 1997; Sandal, 1998). These authors also suggest that, under stressful
conditions, even well-rehearsed actions are poorly performed. However, medical staffs
are more likely to deny the effects of stress and fatigue on their performance (Sexton
et al., 2000).
Sleep deprivation has been identified as one of the main concerns in the practice of
medicine. Long, continuous working hours and sleep deprivation are incompatible with
enhancing quality care and patient safety (Shine, 2002). Medical trainees, who share
a great part of the health care burden, have an immense workload and suffer from the
accompanying sleep deprivation and fatigue (Schwartz et al., 1992). Sleep deprivation
is considered the prime cause of many health care errors committed by trainees (Gaba
et al., 1994). With fatigue and associated stress topping the concerns of the sleep-
deprived health care professional, provision of quality care and patient safety may not
be a priority. It is again the system at play. The transportation and aviation industries have
long recognized the disastrous consequences of fatigue for safety and have implemented
strict regulations (Hopkins, 1971; Office of Technology Assessment, 1991).
Health care is the only high-risk setting that may have overlooked the fatigue factor
and therefore failed to implement any regulations limiting professional staff’s work-
ing hours. Resident physicians seem to be suffering the most from these unregulated
work hours. In July 2003, the Accreditation Council for Graduate Medical Education
(ACGME) in the United States mandated an 80-hour workweek for all medical residents
and fellows in accredited training programs. Resident physicians provide a source of
expertise at less cost than that associated with more experienced doctors; in addition,
the training years are crucial for honing residents’ clinical skills. It is therefore believed
that the more time residents spend training, the better the skills that they acquire – but,
in the meantime, patient safety may suffer.
Concerns have also been raised that the recent work-hour restrictions imposed by
the ACGME have undercut the importance of continuity of care in surgical practice
(Fischer, 2004). This argument is validated by the finding that a substantial number of
residents reported negative impacts on surgical training and quality and continuity of
patient care due to work-hour limitations in New York State (Whang et al., 2003). In
3.2 The cognitive influence on error-generating behavior 27
another, similar survey, fewer than half of residents expected the work-hour reforms to
have a positive impact on patient care. Romanchuk (2004), however, suggested that the
new regulations would result in a happier work setting that would facilitate improved
patient care. The impact of the changes should be continuously monitored over time
before researchers reach any conclusions. It is also interesting to note that the new
work-hour limits imposed by the ACGME exceed the allowable number of work hours
in many other hazardous industries. Also, an arbitrarily selected 80-hour workweek
may not necessarily reduce the occurrence of medical errors by trainees. In this context,
the findings of Davydov et al. (2004) are worth consideration. These authors reported
that they found no correlation between the number of hours worked and the number
of prescribing errors made by medical house staff. An important limitation of this study,
as the authors themselves acknowledged, was the short time period assessed, which
makes it difficult to draw any definitive conclusions. The short study period also made it
impossible to consider the cumulative effect on patient care of fatigue over weeks and
months. Concerns about health care professionals’ attitudes toward errors, teamwork,
and the effects of stress and fatigue on performance are prime areas for further research
and improvements in health care (Sexton et al., 2000).
students know how to make decisions. Recently, however, innovative training strategies
in cognitive skills development through strategic management simulation techniques
have been proposed (Satish and Streufert, 2002).
Cognitive decision-making errors are encountered in complex setups like health care
because of biases. Leape (1994) suggested that in clinical situations that do not follow
a normal pattern and that require rule-based or knowledge-based solutions, humans
are biased to search for a “prepackaged solution,” that is, a rule, before resorting to
more strenuous “knowledge-based functioning.” Bornstein and Emler (2001) noted the
unavailability of any perfect method for eliminating biases in clinical decision making
but suggested employing evidence-based medicine approaches and incorporation of
formal decision analytical tools to improve the quality of the clinical reasoning. There
can also exist a mindset that is based on a misunderstanding of the situation. There can
be a disconnect between what someone believes a situation to be and what the situa-
tion really is or how it has developed (Dekker, 2002). Similarly, people can continue
with their original plan despite indications that the plan should be changed. Use of
modern information technology tools to improve quality and patient safety to aid in
complex medical decision-making may be a feasible approach (Bates and Gawande,
2003; Kushniruk, 2001).
3.3 Conclusion
The decline in quality of care and the increase in the number health care errors require
appropriate action. The solutions may be both sophisticated and simple. Accepting
human fallibility in health care systems can be a sound beginning that allows us to
address the need for safer system designs and human factor re-engineering. Health care
organizations should urgently implement the best-known practices in patient safety and
error reduction and simultaneously focus research strategies to attain the next level of
knowledge in error prevention techniques. At the same time, adherence to basics and
interdisciplinary experience, sharing, and learning should be encouraged. Health care
needs to adopt systematic approaches to absorb and eliminate errors. Technical com-
petence is not the key, but system re-engineering to enhance quality care should be the
prime objective. If errors cannot be reduced to zero, the system must at the least aim to
reduce to zero the number of instances when error harms a patient (Nolan, 2000). It is
our hope that the study of health care errors will indeed be the discipline through which
health care will advance and achieve its quality goals.
References
Bates DW, Gawande AA. Improving safety with information technology. N Engl J Med
2003;348:2526–2534.
Blumenthal D. Making medical errors into “medical treasures.” JAMA 1994;272:1867–1868.
Bornstein BH, Emler AC. Rationality in medical decision making: a review of the literature on
doctors’ decision-making biases. J Eval Clin Pract 2001;7:97–107.
Combs A, Taylor C. The effect of the perception of mild degrees of threat on performance.
J Abnorm Soc Psychol 1952;47:420–424.
Cook RI, Woods DD. Operating at the sharp end: the complexity of human error. In: Bogner
MS, ed. Human Error in Medicine. Hillsdale, NJ: Erlbaum; 1994:255–310.
References 29
Croskerry P. The cognitive imperative: thinking about how we think. Acad Emerg Med
2000a;7:1223–1231.
Croskerry P. The feedback sanction. Acad Emerg Med 2000b;7:1232–1238.
Davydov L, Caliendo G, Mehl B, Smith LG. Investigation of correlation between house-staff
work hours and prescribing errors. Am J Health Syst Pharm 2004;61:1130–1134.
Dekker, S. The Field Guide to Human Error Investigations. Aldershot: Ashgate; 2002.
Dinges DF, Pack F, Williams K, et al. Cumulative sleepiness, mood disturbance, and psycho-
motor vigilance performance decrements during a week of sleep restricted to 4–5 hours
per night. Sleep 1997;20:267.
Easterbrook JA. The effect of emotion on cue utilization and the organization of behavior.
Psychological Rev 1959;66:183–201.
Fischer JE. Continuity of care: a casualty of the 80-hour work week. Acad Med 2004;79:
381–383.
Gaba DM, Howard SK, Jump B. Production pressure in the work environment. California
anesthesiologists’ attitudes and experiences. Anesthesiology 1994;81:488–500.
Helmreich RL. On error management: lessons from aviation. BMJ 2000;320:781–785.
Hopkins GE. The Airline Pilots: A Study in Elite Unionization. Cambridge, MA: Harvard Univ.
Press; 1971:116–141.
Institute of Medicine. Crossing the Quality Chasm: A New Health System for the 21st Century.
Washington, DC: National Academy Press; 2001.
Kassirer JP. Teaching problem-solving – how are we doing? N Engl J Med 1995;332:1507–1509.
Kushniruk AW. Analysis of complex decision-making processes in health care: cognitive ap-
proaches to health informatics. J Biomed Inform 2001;34:365–376.
Leape LL. Error in medicine. JAMA 1994;272:1851–1857.
Nolan TW. System changes to improve patient safety. BMJ 2000;320:771–773.
Office of Technology Assessment. Biological Rhythms: Implications for the Worker. Washing-
ton, DC: Government Printing Office; 1991. Report No. OTA-BA-463.
Rasmussen J, Jensen A. Mental procedures in real life tasks: a case study of electronic trouble
shooting. Ergonomics 1974;17:293–307.
Rasmussen J, Pedersen OM. Human Factors in Probabilistic Risk Analysis and Risk Manage-
ment. Operational Safety of Nuclear Power Plants, vol. 1. Vienna: International Atomic
Energy Agency; 1984.
Reason J. Human Error. Cambridge: Cambridge University Press; 1990.
Roberts JS, Coale JG, Redman RR. A history of the joint commission on accreditation of
hospitals. JAMA 1987;258:936–940.
Romanchuk K. The effect of limiting residents’ work hours on their surgical training: a Cana-
dian perspective. Acad Med 2004;79:384–385.
Sandal GM. The effects of personality and interpersonal relations on crew performance during
space simulation studies. Life Support Biosph Sci 1998;5:461–470.
Satish U, Streufert S. Value of a cognitive simulation in medicine: towards optimizing decision
making performance of healthcare personnel. Qual Saf Health Care 2002;11:163–167.
Schwartz RJ, Dubrow TJ, Rosso RF, Williams RA, Butler JA, Wilson SE. Guidelines for surgi-
cal residents’ working hour. Intent vs. reality. Arch Surg 1992;127:778–782 [discussion
782–783].
Sexton JB, Thomas EJ, Helmreich RL. Error, stress, and teamwork in medicine and aviation:
cross sectional surveys. BMJ 2000;320:745–749.
Shine KI. Health care quality and how to achieve it. Acad Med 2002;77:91–99.
Whang EE, Mello MM, Ashley SW, Zinner MJ. Implementing resident work hour limitations:
lessons from the New York State experience. Ann Surg 2003;237:449–455.
4 Medical error and strategies for working
solutions in clinical diagnostic laboratories
and other health care areas
The IOM report (2001) stated that the prevalence of medical errors is high in today’s
health care system. Some specialties in health care are more risky than others.
A varying blunder/error rate of 0.1–9.3% in clinical diagnostic laboratories has
been reported in the literature. Many of these errors occur in the preanalytical and
postanalytical phases of testing. It has been suggested that the number of errors that
occur in clinical diagnostic laboratories is smaller than the number of errors that
occur elsewhere in a hospital setting. However, given the volume of laboratory tests
used in health care, even this low rate may reflect a large number of errors. The
surgical specialties, emergency rooms, and intensive care units have been previ-
ously identified as areas of risk for patient safety. Though the nature of work in these
specialties and their interdependence on clinical diagnostic laboratories present
abundant opportunities for error-generating behavior, many of these errors may be
preventable. Appropriate attention to system factors involved in these errors and the
use of intelligent system approaches may help control and eliminate many of these
errors in health care.
4.1 Introduction
Patient safety is a current public and health care concern. It has received substantial
attention following the release of two recent reports from the IOM (2001; Kohn et al.,
2000). Both of these reports have placed special emphasis on enhancing patient safety
in today’s health care system. Much of the problem in the provision of better and safer
health care is a result of preventable adverse events due to medical errors (Kohn et
al., 2000). We have previously highlighted some of the key concepts in medical error
generation and have suggested a systems approach in medical error reduction (Kalra,
2004).
Some specialties in health care are more prone to errors than others. The notability of
surgical specialties with respect to the rate of adverse outcomes and errors in surgical
practice has been outlined earlier (Leape et al., 1991). The Quality of Australian Health
Care Study (QAHCS) reported that more than half of all the adverse events recorded in
the study were associated with a surgical operation (Wilson et al., 1995). The Harvard
Medical Practice (HMP) Study II identified certain sites of health care that are high-risk
zones for patient safety. In their analysis, the authors suggested that among these sites
are operating rooms, in-patient rooms, emergency rooms, and intensive care units. The
authors did not clearly delineate the adverse events occurring in the operating rooms,
and presumably these may be a result of either anesthetic or surgical interventions.
32 Medical error and strategies for working solutions in clinical diagnostic
However, the authors noted the low preventability of such incidents compared to events
occurring in other high-risk zones, such as emergency departments (EDs) and intensive
care units (ICUs), where both preventability and the potential for disability resulting
from the adverse events are high. It is these high-risk zones, where preventability of ad-
verse events is high, that are prime areas of concern and that we will discuss in greater
detail. Apart from the specialties of emergency and intensive care medicine, which are
highly interdependent on clinical laboratories, we also review the impact of errors in
clinical diagnostic laboratories.
reported a blunder rate of 0.1% for laboratory requests. The low error rates in the study
may have resulted from the fact that the authors examined only errors detected or re-
ported in the final checking-out phase. The errors in the preanalytical, analytical, and
final validation stages of the laboratory process were not included in the study. The
authors also acknowledged that some errors might have gone unnoticed, as the authors
did not make any special efforts to scrutinize reports intensely. Another study identified
a crude rate of 1.1 problems per 1,000 patient visits in laboratory testing in primary care
physicians’ offices (Nutting et al., 1996).
Khoury et al. (1996) studied the error rates in Australian chemical pathology labora-
tories and reported error rates as high as 39% for transcription and 26% for analytical
results, with the best laboratory performing error-free business only 95% of the time.
Khoury et al. surveyed clinical pathology laboratories of five Australian states to com-
pile their data. A transcription error rate as high as 39% of all requests has the potential
to seriously compromise patient identification data. The authors utilized the National
Quality Award Criteria for Benchmark Indicators to identify errors in the laboratory
process (Australian Quality Council, 1994). Lord (1990) underlined the importance of
managing risk in clinical pathology laboratories by ensuring proper identification of
specimens and prevention of adverse patient care-related incidents.
Recently, we studied sample labeling errors, utilizing a retrospective audit of all inci-
dents reported at the front end of a clinical microbiology laboratory during the period
from May 1996 to December 2002 (Kalra et al., 2004a). We observed that unlabeled
and mislabeled samples contributed to the majority of risk problems in the study period.
However, after implementation of a zero tolerance policy to address the labeling errors
in 1999, there was a gradual decrease in these events from 100% (43/43) in 1998 to 64%
(16/25) of the total risk problem in 2002. Along with this, there was a gradual increase
in sample rejections, which decreased the overall probability of risk events. Though our
study observed only one specific aspect of laboratory testing, it underlines the effect
of category-specific risk management policies in addressing the risk issues of clinical
laboratories. The mislabeling rate at Johns Hopkins Hospital was found to be 1.4%, or 1
in 71 specimens (Lumadue et al., 1997). A mislabeled sample is a received sample that
does not meet the hospital’s criteria for sample acceptance. Medical semiautomated
patient identification systems have had radical success in reducing mislabeling rates to
about 0.01%, or 1 in 10,000 specimens (Astion, 2010). Implementation, however, took
around seven years at Valley Hospital in New Jersey.
Hofgartner and Tait (1999), in a study based on error rates in clinical genetic testing
laboratories over a 10-year period, detected error rates of 0.38% in the inspected laborato-
ries. In this study, an onsite review of two participating laboratories was carried out, along
with an extensive review of problem cases that occurred during the study period. Though
the study utilized laboratory data from a wide range of medical specialties, the study
design had its own limitations. The error detection strategies were far from perfect, and, as
a result, some analytical errors may have been ignored if the test results were consistent
with other available clinical data. It should also be emphasized that the reviewed studies
had different study designs and were therefore incomparable in many aspects. Also, with
no universal definition of laboratory error/blunder and imperfect error detection methods,
not all laboratory errors may have been accounted for in these studies.
Bonini et al. (2002) studied and reported on the contrasting error rates in inpatients
and outpatients subjected to a laboratory test. The authors suggested an error rate of
34 Medical error and strategies for working solutions in clinical diagnostic
0.60% and 0.039%, respectively, in the two categories and attributed the large differ-
ence to human factors related to skill in drawing blood and the sheer amount of labora-
tory usage for inpatients. Another important factor contributing to erroneous laboratory
reporting, which may compromise quality care and hamper patient safety, is analytical
immunointerference in laboratory assays. This problem has been highlighted in two
recent reports and can lead to unnecessary consultations and other interventions to
patients (Ismail et al., 2002; Marks, 2002). The importance of information and adequate
communication between the clinician and laboratory technologists in such instances
cannot be overemphasized.
Goldschmidt and Lent (1993) observed that 12.5% of laboratory errors lead to an
erroneous medical decision, 75% of the results were within normal reference limits,
and in the remaining 12.5% the results were too absurd to be considered for clinical
decisions. Plebani and Carraro (1997) suggested that 74% of the laboratory mistakes
did not influence patients’ outcomes and 19% resulted in increased costs to the patient
because they led to further inappropriate investigations. The authors suggested that
6.4% of laboratory mistakes caused inappropriate care or inappropriate modification of
therapy to the patient. fFig. 4.1 summarizes the frequency of errors observed in clinical
laboratories (Boone, 1990; Chambers et al., 1986; Lapworth and Teal, 1994; Nutting
et al., 1996; Plebani and Carraro, 1997).
If one individual is frequently committing more errors than his peers without any
obvious impairment or recklessness or his behalf, then he will have to be moved to a
different service or terminated. If the errors are due to known external factors, such as
bereavement, then a temporary leave from duties of patient care is necessary.
The various types and the rate of errors occurring at these three stages of a laboratory
process as suggested by two studies (Plebani and Carraro, 1997; Wiwanitkit, 2001) are
summarized in fFig. 4.2. It has been suggested that a majority of errors occurring in a
laboratory process are due to preanalytical factors (Hofgartner and Tait, 1999; Khoury
et al., 1996; Plebani and Carraro, 1997; Wiwanitkit, 2001). The high rates of pre- and
postanalytical errors undermine the quality performance of the analytical process and
necessitate the active involvement of nonlaboratory personnel, particularly phleboto-
mists and clinicians, in improving the quality of laboratory reporting. The high error
rates in the nonanalytical stage, where human involvement is maximal, reaffirm the
susceptibility of a manual process to error.
Consider, for instance, the case of a 36-year-old female patient who is experiencing
her first pregnancy. A blood test shows that she has low levels of protein-S, and her
physician warns her about the significant risk of thrombosis – a rare blood disease in
which blood clots form within a vein, causing significant risk to the patient. Fearing
for her safety, the woman terminates the pregnancy. In this case, while the preanalyti-
cal and analytical phases have proceeded without error – the blood sample was prop-
erly prepared and tested – the physician is unaware that low protein-S values in most
pregnancies do not increase the risk of thrombosis. This is a postanalytical error, as the
correct results have been incorrectly analyzed.
Such mistakes are costly. Patients suffer harmful physical and emotional conse-
quences from receiving misdiagnoses and undergoing unnecessary or even dangerous
procedures. This, in turn, increases the cost of care in the form of additional lab tests
Equipment
Malfunction
Reporting
Insufficient Sample or
Sample Condition Analysis
Fig. 4.2: Types and rates of error in the three stages of the laboratory testing process.
36 Medical error and strategies for working solutions in clinical diagnostic
and professional labor to rectify the resulting medical errors, as well as costing valuable
physician and laboratory staff time in the corrective action taken after incorrectly
interpreted tests, as well as the interpretation itself.
Laboratory medicine has high quality standards. This is ensured by a number of regu-
latory authorities, including the National Joint Commission on Accreditation of Health
Care Organizations (JCAHO)) and the College of American Pathologists (CAP), that peri-
odically outline the quality and accrediting standards required for practice. In addition,
quality improvement initiatives, such as the Q-probes quality improvement program of
the CAP (Bachner and Howanitz, 1991), instituted to establish provisional benchmarks
for measuring systematic quality improvement efforts, are also in place. The efforts of
such initiatives are paying high dividends in areas that have been identified for quality
improvement and enhancing patient safety, such as wristband patient identification and
laboratory specimen acceptability (Howanitz et al., 2002; Zarbo et al., 2002). Other
professional societies are also involved in sponsoring a number of educational pro-
grams that focus on quality issues. The professional boards responsible for certification
processes have ensured the competency in practice of quality laboratory medicine over
the years. These measures have helped laboratory medicine to function at the level of
reliability that is expected from any clinical practice involved in health care. However,
the focus of these certification and accreditation processes is most often on laboratories’
general performance, rather than particularly on errors (Bonini et al., 2002) and other
related issues closely linked to quality.
Two particularly important quality issues confronting modern laboratory services are
the problems of weak control over affairs in the preanalytical stage and the identifica-
tion of errors and subsequent reporting. The preanalytical phase is clearly the most
important part of the laboratory testing process. What a laboratory analyzes and ulti-
mately reports is a direct result of the quality of the specimen or sample delivered for
testing. However, it is suggested that the clinical laboratories assume responsibility for
the whole cycle of the testing process, from the physician’s ordering a laboratory investi-
gation to the recognition of the significance of the reported result in the management of
the patient. This responsibility requires that the testing cycle be completed by involving
other professionals in the quality loop. The benefits achieved by such efforts have previ-
ously been demonstrated. Renner et al. (1993) reported the influence of phlebotomists
on improving quality and patient safety in transfusion practice by monitoring the level
of conformity of wristbands to quality standards.
case of interobserver variation. However, such variations can have far-reaching con-
sequences for a patient because of the different diagnoses and subsequent case man-
agement. There are some other common uncertainties in a laboratory testing process
over which the clinical laboratories have limited control. These may be in the form of
biological variation, systematic errors, analytical variation, and interfering substances
that are not controlled even by the addition of blocking agents.
Many of the quality improvement and patient safety measures used in different sectors
of health care may also apply widely to the practice of laboratory medicine. We have
previously suggested a “no-fault” error reporting model for pathology and laboratory
medicine, which emphasizes the importance of risk-free voluntary error reporting sys-
tems for health care practitioners to develop a safe health care system (Kalra et al., 2003).
Further elaborating on this model, we have also integrated an incident reporting sys-
tem in our clinical chemistry laboratory. The incident reporting system includes various
types of errors in the clinical laboratory and encourages voluntary reporting of critical
laboratory incidents on a nonpunitive basis. The implemented model also includes an
educational program and a quality care committee to monitor the critical incidents and
medical errors in the clinical laboratory (Kalra et al., 2004b). The quality care committee
reviews the reported errors periodically and recommends appropriate action aimed at
preventing the occurrence of similar events. This model, along with the laboratory quality
assurance program, forms an integral part of clinical diagnostic/chemistry laboratories.
The model is in its infancy, and its impact is being continuously assessed.
Certain other measures with particular relevance to laboratory medicine initiatives in
improving quality care are reviewed in this section. Apart from making patient safety
a priority and a departmental goal, laboratory medicine should dedicate resources for
education and research to address the unique concerns in practices like those discussed
in the preceding paragraph. Research initiatives should be directed toward establishing
the best evidence-based practices and diagnostic gold standards. Studies should also be
directed toward eliminating biases in decision making related to diagnostic activities.
Clinical laboratories should identify areas where human involvement in the process can
be minimized and increased possibilities for automation and robotics exist. Laboratory
automation provides for standardized workflow and helps eliminate many error-prone
steps undertaken by humans. Automation may decrease opportunity for errors caused
by active human factors like stress, fatigue, negligence, and cognitive impairment.
Health care is an intensive and highly skilled profession, and clinicians are required
to work long hours. Moreover, the work situation itself is particularly prone to fatigue-
associated error. For example, clinicians are required to detect rare but important sig-
nals; for example, they must notice error flags in a system of automated tests. This is in
addition to performing tasks that require multitasking and creative troubleshooting, all
of which are susceptible to fatigue (Astion, 2009).
There are, however, both temporary and long-term ways to reduce fatigue. Stimulants,
such as coffee; brighter lights in work areas; and conversation and physical activity
among staff are all ways to temporarily reduce fatigue (Astion, 2009). In the long term,
staggered shifts should be implemented, rather than abrupt shift changes. The workers
who have just begun their shift should be made to handle the hardest work available.
Staggered shifts allow for better hand-offs of workloads than abrupt shifts. When clini-
cians are transferring their workload at the end of their shift, they’re not only tired but
also eager to depart from the workplace. As such, the quality of information received
38 Medical error and strategies for working solutions in clinical diagnostic
by the incoming clinician isn’t as high as it would be were the departing colleague
less tired, and communication breakdowns would be less common. Furthermore, there
should exist a permanent night shift so that workers can develop regular sleep cycles.
The emergency departments and intensive care units depend heavily on clinical
laboratories. This dependence involves but is not restricted to therapeutic drug monitor-
ing, interpretation of laboratory results in view of interfering substances, and laboratory
monitoring or surveillance to predict signs of clinically unrecognized toxicity. The other
critical area where the clinical laboratories and EDs are integrated to achieve effective
clinical decision making is diagnosis. This issue, in the context of acute myocardial in-
farction (AMI), is discussed in the following section. With greater integration and effec-
tive information sharing between clinical laboratories and these units, the medication
and diagnostic errors occurring in the EDs and intensive care units may be reduced to
a large extent.
The reliability of the figures quoted by studies to suggest the error rates in EDs can
itself be questioned. Emergency medicine is a relatively new specialty, and the figures
quoted by the HMP study II and the IOM report are derived from fairly old data, data
that were gathered before trained and board-certified ED physicians were managing
and providing care in EDs. Times have changed. ED personnel are now professionally
staffed and managed, but one aspect of EDs remains the same – the time-pressured
environment. This environment is a fertile breeding ground for errors. Hence, it may not
be appropriate to assume that the statistics on error rates in EDs are exaggerated and
that EDs are safe. The reality is otherwise.
Emergency medicine was one of the first specialties to realize the potential of errors
to compromise patient safety and quality of care. The emergency medicine specialty
was the pioneer in patient safety initiatives. The wide varieties of presentations and
diagnoses in EDs have led to focused research and effective interventions. Rosen et al.
(1983) were among the first researchers to study errors in emergency medicine, and
they suggested innovative approaches to help eliminate such errors. Karcz et al. (1996)
reviewed malpractice claims filed against emergency physicians in Massachusetts over
an 18-year period, 1975–1993. The authors identified chest pain, abdominal pain, and
fractures as areas where physicians were at high risk of making errors.
Chest pain as a result of acute cardiac ischemia is both a common and a frequently
missed diagnosis in EDs. Approximately 2% of patients with AMI were misdiagnosed,
with 25% of these patients having a fatal outcome, in a multicenter study in a coronary
care unit (McCarthy et al., 1993). The authors employed a case control study design with
a sample size of 1,050 patients with AMI. They compared patients with a missed AMI di-
agnosis to two control groups, patients admitted with AMI and a group discharged with-
out AMI. At least 50% of misdiagnoses could have been prevented if the ST elevations
of electrocardiographs (ECGs) had been recognized or patients who were recognized as
having ischemic heart disease by the physicians in the ED had been admitted. However,
the authors suggested that the patients with missed AMI were significantly less likely
to have ECG changes than patients admitted with AMI. Chan et al. (1998), by means
of a retrospective analysis, estimated that at least 27% of AMI were missed in the EDs
and cited the absence of chest pain and the lack of ST elevations in ECGs as the main
predisposing factors for misdiagnosis. The relative higher rate of missed AMI diagnoses
in this study than in the earlier study by McCarthy et al. (1993) may be attributed to a
smaller sample size (159 patients) in the Chan et al. study and the use of a different
study design. Another reason for the difference may be the demographic and regional
variations evident in the two study designs. The former was conducted in North America
and the latter in Asia. The preventability rate in the study by Chan et al. was 35%, and
the errors were mostly attributed to inaccurate interpretation of ECGs. However, the
competence of emergency physicians in ECG interpretation of patients with chest pain
and ST elevation has been proved (Brady et al., 2000). Brady et al. also suggested that
the low rate of ECG misinterpretation was of minimal clinical consequence.
In 2005, the Canadian Patient Safety Institute (CPSI) launched a grassroots campaign
named Safer Healthcare Now! (SHN). The campaign aims to promote patient safety by
endorsing six targeted, evidence-based interventions by health care professionals (CPSI,
2007). One of the promoted interventions is improved care for AMI in Canada. CPSI
aims to help hospitalized patients by ensuring the reliable delivery of evidence-based
40 Medical error and strategies for working solutions in clinical diagnostic
care, including a focus on correct diagnosis of the illness. A National Steering Committee
officially leads the SHN campaign, with support from the CPSI.
Pope et al. (2000) found a correlation between racial, gender, and clinical factors
and failure to hospitalize patients presenting with acute cardiac ischemia. The authors
suggested that nonwhites, women less than 55 years old, patients presenting chiefly
with shortness of breath, and patients with nondiagnostic ECGs were ideal candidates
for nonhospitalization. All of these findings point to other extraneous sources apart from
human factors in erroneous decision making. These factors are more important as they
are unidentified and hence have the potential to recur and to lead to more catastrophic
consequences in the future. Clinical laboratories have a limited but specific role in min-
imizing these missed diagnoses. The judicious use of new-generation cardiac troponin
tests based on evidence-based medicine may be a solution. In establishing the cut-off
levels for therapeutic interventions, the clinical laboratories have to play a crucial role
along with EDs and cardiology units. Clear communication and interaction between
clinical laboratories and clinicians are essential in maximizing the availability and ef-
fectiveness of these and other tests in helping them to make better decisions. Clinical
laboratories should also focus on decreasing the turnaround times (TAT) for diagnosti-
cally important tests like those used for AMI. The effective use of clinical laboratories,
together with ECGs and other useful risk factor information, may serve to enhance
clinical decision making and eliminate errors in the diagnosis of AMI.
Another subgroup of diagnoses that is often misdiagnosed in EDs is acute surgical
emergencies, particularly appendicitis. This condition often poses problems for accu-
rate diagnosis and prompt treatment because of the differing clinical features in every
patient. Rothrock et al. (1991) compared the clinical presentations of children with
misdiagnosed appendicitis and those in whom the appendicitis was correctly diagnosed
the first time. These investigators concluded that clinical presentations in the two groups
were different. Reynolds (1993) retrospectively reviewed hospital ED records and re-
ported that at least 7% of the patients were seen twice in their EDs before the diagnosis
of appendicitis was finally made. Similar results were obtained in a retrospective case
series study on nonpregnant women of childbearing age, which showed that 33% of the
women with appendicitis were initially misdiagnosed (Rothrock et al., 1995). One of
the main reasons for the misdiagnoses in the two studies was the absence of the typical
findings that are normally associated with appendicitis. The consequences of such a
delay in diagnosis can be grave. The most common complication of delayed diagnosis
or misdiagnosed appendicitis is perforation. The rate of perforation can vary between
71% (Cappendijk and Hazebroek, 2000) and 91% (Rusnak et al., 1994). Other com-
plications include the need for more extensive surgical procedures, more postoperative
complications, and longer hospital stays. The problem lies not only in underdiagnosis
but also in overdiagnosis of appendicitis in EDs, which results in negative appendec-
tomies, surgeries performed as a result of presumed appendicitis. The frequency with
which misdiagnosis leads to unnecessary appendectomies has not changed very much
in spite of modern imaging techniques like CT scans (Flum et al., 2001). An estimated
$741.5 million in total hospital costs was attributed to this presumed diagnosis of
appendicitis in 1997 (Flum and Koepsell, 2002).
Another area of concern in EDs that needs further improvements in quality to ben-
efit patient safety is radiograph interpretation. Emergency medicine professionals are
often required to interpret radiographs in the absence of trained radiologists. About
4.6 Errors in intensive care medicine 41
an incident report filed during their ICU admission was 41%, whereas the rate for all
ICU patients was 21%. A prospective two-center study of adult ICUs reported iatrogenic
complications in 31% of the total of 400 admissions, and 13% of these admissions had
major complications as a result of the adverse event (Giraud et al., 1993). The study
also reported that the mortality risk was close to twice as high in patients with a major
iatrogenic complication as in other patients without any iatrogenic complications. An
important limitation of this study was that it excluded medication-related adverse events.
Exclusion of this important category of adverse events, as discussed later, may have led
the researchers to underestimate the adverse event rates in the study. Donchin et al.
(2003) investigated the nature and causes of human errors in ICUs through a concurrent-
incident study. The study suggested that an estimated 1.7 errors occurred per patient
per day, and for the whole ICU it reported that a severe or potentially detrimental error
occurred, on average, twice a day. The authors cautioned about generalizing from these
results as the study adopted a direct-observation methodology, which may have pre-
vented some of the error-generating behavior of the staff. Another limitation of the study
was that it was conducted in a unit that was grossly understaffed and that also served as
a training facility. These factors may have influenced the results obtained by the study.
A high proportion of errors occurring in ICUs may be associated with medications
(Flaatten and Hevroy, 1999). An earlier prospective study in a NICU reported a total of
313 medication error incidents in 23,307 patient days, and these errors were respon-
sible for a majority of all serious incidents (Vincer et al., 1989). However, negligence
related to medication schedules and failure to regulate intravenous infusions were
responsible for the majority of these events. Another interesting study compared the
rates of preventable adverse drug events in ICUs with general care units (Cullen et al.,
1997). The study revealed that the rate of preventable adverse drug events was twice
as high in ICUs as in non-ICUs. The investigators conducted structured interviews that
indicated that there were almost no differences between ICUs and general care units
on many characteristics of the patient and caregiving teams. The authors also failed to
notice any correlation between adverse drug events in the units and the working envi-
ronment of the staff. An error rate of 6.6% for all drug administration events has been
reported in a prospective pharmacist-performed observation study (Tissot et al., 1999).
Dosing errors accounted for the majority of detected errors, and, though there were no
fatal errors, at least 20% of the errors were potentially life threatening. However, the
authors of the study could not correlate the high incidence of preventable adverse drug
events with increased workload or a stressful environment. Nevertheless, the authors
identified deficiencies in the overall organization of the hospital medication track, in
patient follow-up, and in staff training as reasons for the errors. The study, like some of
the others cited, suffered the limitation of observer bias that is inherent in some directly
observed studies (Donchin et al., 1995).
Another observational study on medication errors of adult patients in ICUs found
more promising results than those reported in earlier studies. The study reported a rather
low 3.3% incidence rate for medication errors (Calabrese et al., 2001). The authors
identified wrong infusion rates as the most common type of error. However, it must be
mentioned that the different methodologies of various authors, their varying definitions
of errors, and the different settings make fair comparison of results difficult, if not impos-
sible. Allan and Barker (1990) suggest that the “disguised observation technique” used
by some authors (Calabrese et al., 2001; Tissot et al., 1999) may be a more efficient
4.6 Errors in intensive care medicine 43
method for medication error research than the retrospective incident report analysis
used by other authors (Abramson et al., 1980). Some of the reviewed studies suffer from
limitations in the form of study design, as their focus was restricted to drug administra-
tion errors (Tissot et al., 1999; Calabrese et al., 2001) or nursing staff errors (Tissot et al.,
1999). The specific focus on certain events or groups may have led researchers to miss
other critical incidents in different settings. Also, the error detection techniques utilized
by the authors themselves may not be perfect, allowing many medication errors to
escape scrutiny. Though the error rates reported in the literature vary, the consequences
remain the same – patient harm.
Some of the common factors promoting errors have been identified as a lack of stan-
dardization, insufficient labeling of medications, improper documentation, and, most
important, poor communication (Gopher et al., 1989). Clear communication is a princi-
pal component in providing quality care in ICUs. Fassier (2010) demonstrated that poor
communication underlies many conflicts in the ICU. Studies indicate that it is generally
believed that communication practiced in ICUs is suboptimal and that ICU profession-
als poorly understand the ultimate goals of care for any particular day (Pronovost et al.,
2003). Pronovost et al. reported that fewer than 10% of ICU residents and nurses under-
stood the goals of care. The authors implemented a daily goals form in the ICUs and ob-
served that, after its implementation, more than 95% of nurses and residents understood
the goals of care for that day. This intervention resulted in a reduction in patients’ total
length of stay and significant improvements in the provision of quality care.
Human involvement in errors is very high in ICUs, and at times these errors may
have catastrophic consequences. Wright et al. (1991) attributed 80% of critical events
to human error. Earlier, Abramson et al. had reported that 63% of incidents were due
to human error. Other studies have periodically reported similar findings (Giraud et al.,
1993; Stambouly et al., 1996). Bracco et al. (2001) studied human errors prospectively
in a multidisciplinary ICU and found that 31% of the critical incidents reported in the
study were due to human errors. A majority of the errors in the study occurred while a
clinician was executing an action, but planning errors were primarily responsible for
significant complications. These findings suggest that human errors are a major cause of
critical incidents, threatening quality care and patient safety in an ICU environment.
Another critical area that needs attention if we are to make ICUs safer is staff workload
and training. Understaffing and a decreased staff-to-patient ratio are burning concerns
not only in ICUs but also health care in general. This problem with respect to ICUs
deserves special mention because of the growing number of procedural interventions
and the critical condition of the patients in ICUs. This shortage of staff threatens quality
and service levels in ICUs. The research in this area has shown that mortality, adjusted
for all other factors, was more than twice as high in patients exposed to ICUS with the
greatest workloads as those where workers had lower ICU workloads (Tarnow-Mordi,
2000). It has also been reported that patients in ICUs with fewer nurses have an in-
creased risk of postoperative respiratory complications (Dimick et al., 2001a; Pronovost
et al., 2001a). Many of the workload-associated problems could be offset to an extent
by employing nurses without ICU experience who could assist the regular ICU nursing
staff by providing basic nursing care, thereby relieving the ICU staff’s workload while
not compromising the quality of care (Binnekade et al., 2003).
The other area that needs to be addressed if we are to enhance quality of care and
patient safety in ICUs is staff training. Present-day ICUs commonly follow one of the
44 Medical error and strategies for working solutions in clinical diagnostic
three staffing models. The most common model is the open-model ICU, where the ad-
mitting physician is in charge of the care of their patients. The next most common is the
closed-model ICU, which is managed by trained critical-care specialists, or intensivists,
who are responsible for the full care of the patients in ICUs. The third model is the semi-
closed model, in which the admitting physician coordinates with the intensivists who
manage the care of the patient. ICUs have been slow to convert from the open model
to the closed model, though the benefits outweigh the costs and risks. An estimated
53,000 lives and $5.4 billion in costs could be saved annually if intensivist physician
staffing were implemented in the nonrural United States (Pronovost et al., 2001b). The
intensivists bring with them advanced training in critical-care medicine and technical
skills, which immensely benefits the patient. The evidence suggests that ICUs staffed
with intensivists are more likely to produce the desired results for patient outcomes
(Dimick et al., 2001b). The organizational change from open to closed ICUs improves
clinical outcomes, and it has been reported that nurses had more confidence in the
clinical judgment of the physician in the closed ICU than in the open ICU (Carson et
al., 1996). And it is not only physicians’ expertise, training, and experience that benefits
quality of care and patient safety. Morrison et al. (2001) suggested that nursing staff
inexperience contributed to adverse patient outcomes in ICUs. Amid all these disturb-
ing trends in ICUs, there seems to be promise. Much of that promise derives from the
premise that continuous research and incident analysis by means of critical incident
reporting system will improve patient safety.
The reporting of incidents, including both adverse events and near-mistakes caught
just before they actually took place, is an essential component of improving patient
safety. In intensive care, the Australian Incident Monitoring System (AIMS-ICU) has
pioneered efforts in this direction (Beckmann, 1996). AIMS-ICU is a national critical
error reporting system set up in 1993 to assess the impact of an anonymous, volun-
tary reporting system on quality of care and patient safety. It has been demonstrated
that critical incident reporting is effective in revealing the latent errors present in the
system, the role of human error, and the generation of these critical incidents (Buckley
et al., 1997). Additionally, this reporting system is a useful way to highlight problems
that have been previously undetected by the routine quality assurance programs. Frey
et al. (2002) reported on the benefits of critical incident reporting systems on medication
error prevention in ICUs. In a study of two methods, incident monitoring and retrospec-
tive chart review, it was found that incident monitoring offered more contextual infor-
mation about the incidence of such errors and identified a larger number of preventable
problems than did medical chart reviews (Beckmann et al., 2003). However, the authors
of the study suggested that incident monitoring identified fewer iatrogenic infections
and advised supplementing this process with selective audits and focused medical chart
reviews to detect those problems not reported well by incident monitoring alone.
Patient monitoring is critical in ICU environments. The purpose of monitoring equip-
ment is to provide timely and maximum information. Good cognitive ergonomic
design of monitoring equipment should serve to reduce errors caused by human fac-
tors. Instrument-redesigning research is as important in reducing human errors as is
focusing on organizational issues related to work. Innovative, targeted interventions for
ICU alarms and their warnings have been previously suggested. McIntyre and Nelson
(1989) suggested that human voice messages were far superior to nonverbal signals
emanating from ICU monitoring devices. The investigators suggested that such warning
4.7 Conclusion 45
signals might influence humans to make fewer errors. Schoenberg et al. (1999) devised
a module that assists in eliminating false and insignificant alarms in the ICU setting and
improves the efficacy of these devices.
The high rate of medication errors in ICUs is a primary area that needs improvement.
Landis (1999) reported that putting a pharmacist on the ICU team could cut ordering er-
rors by 66%. This area requires innovative approaches so that basic procedures such as
ordering, following the orders, and drug administrations are made error-proof and any
potential for erring is completely eliminated. All ICUs should adopt critical incident re-
porting as a routine, and these reporting systems should be incorporated in the policy of
continuous quality improvements. Such a system would form a vital link in identifying
deficiencies in the system. Correction of these deficiencies will lead to a reduction in
future incidence of these critical incidents and will promote overall patient safety.
4.7 Conclusion
Efforts to reduce medical errors and enhance patient safety can be directed toward
many health care processes. Clinical diagnostic laboratories, EDs, and ICUs are recog-
nized risk zones with high potential for error generation. It is prudent that maximum
efforts be directed toward preventing errors in areas of high preventability. The human
factors involved in generating errors in these specialties deserve priority, and appropri-
ate technology may play a vital role in mitigating many of these factors. Staffing short-
ages in some of these risk zones may be an individual factor affecting the quality of
care provided. Regulatory and legislative bodies have a central role in addressing many
of these patient safety issues. Adopting intelligent systems approaches to promoting
efficiency and enhancing team coordination to facilitate optimal outcomes in patient
care are a necessity. The design and development of adequate programs to achieve a
culture of safety in all health care processes will be a continuing challenge that needs
to be appropriately addressed.
References
Abramson NS, Wald KS, Grenvik AN, Robinson D, Snyder JV. Adverse occurrences in inten-
sive care units. JAMA 1980;244:1582–1584.
Allan EL, Barker KN. Fundamentals of medication error research. Am J Hosp Pharm 1990;
47:555–571.
Astion M. Fatigue and error: an interview with Dr. Matthew B. Weinger. Clin Lab News
2009;35:10–11.
Astion M. Lean principles and patient safety overview: principles and solutions in using LEAN
to reduce errors in patient ID and specimen collection. Presentation. 2010.
Australian Quality Council. Australian Quality Awards Foundation. Assessment Criteria and
Application Guidelines. Sydney: Australian Quality Council; 1994.
Bachner P, Howanitz PJ. Using Q-Probes to improve the quality of laboratory medicine: a
quality improvement program of the College of American Pathologists. Qual Assur Health
Care 1991;3:167–177.
Beckmann U, Bohringer C, Carless R, et al. Evaluation of two methods for quality improve-
ment in intensive care: facilitated incident monitoring and retrospective medical chart
review. Crit Care Med 2003;31:1006–1011.
46 Medical error and strategies for working solutions in clinical diagnostic
Beckmann U, West LF, Groombridge GJ. The Australian incident monitoring study in intensive
care: AIMS-ICU. The development and evaluation of an incident reporting system in inten-
sive care. Anaesth Intensive Care 1996;24:314–319.
Binnekade JM, Vroom MB, de Mol BA, de Haan RJ. The quality of intensive care nursing
before, during, and after the introduction of nurses without ICU-training. Heart Lung
2003;32:190–196.
Bonini P, Plebani M, Ceriotti F, Rubboli F. Errors in laboratory medicine. Clin Chem 2002;
48:691–698.
Boone DJ. Comment on “Random errors in haematology tests.” Clin Lab Haematol 1990;
12(Suppl. 1):169–70.
Bracco D, Favre JB, Bissonnette B, et al. Human errors in a multidisciplinary intensive care
unit: a 1-year prospective study. Intensive Care Med 2001;27:137–145.
Brady WJ, Perron A, Ullman E. Errors in emergency physician interpretation of ST-segment eleva-
tion in emergency department chest pain patients. Acad Emerg Med 2000;7: 1256–1260.
Brunswick JE, Ilkhanipour K, Seaberg DC, McGill L. Radiographic interpretation in the emer-
gency department. Am J Emerg Med 1996;14:346–348.
Buckley TA, Short TG, Rowbottom YM, Oh TE. Critical incident reporting in the intensive care
unit. Anaesthesia 1997;52:403–409.
Calabrese AD, Erstad BL, Brandl K, Barletta JF, Kane SL, Sherman DS. Medication administra-
tion errors in adult patients in the ICU. Intensive Care Med 2001;27:1592–1598.
Canadian Patient Safety Institute. Safer Healthcare Now! Program Review. July 23, 2007.
Cappendijk VC, Hazebroek FW. The impact of diagnostic delay on the course of acute ap-
pendicitis. Arch Dis Child 2000;83:64–66.
Carson SS, Stocking C, Podsadecki T, et al. Effects of organizational change in the medical
intensive care unit of a teaching hospital: a comparison of “open” and “closed” formats.
JAMA 1996;276:322–328.
Chambers AM, Elder J, O’Reilly DS. The blunder-rate in a clinical biochemistry service. Ann
Clin Biochem 1986;23:470–473.
Chan WK, Leung KF, Lee YF, Hung CS, Kung NS, Lau FL. Undiagnosed acute myocardial in-
farction in the accident and emergency department: reasons and implications. Eur J Emerg
Med 1998;5:219–224.
Chin MH, Wang LC, Jin L, et al. Appropriateness of medication selection for older persons in
an urban academic emergency department. Acad Emerg Med 1999;6:1232–1242.
Cullen DJ, Sweitzer BJ, Bates DW, Burdick E, Edmondson A, Leape LL. Preventable adverse
drug events in hospitalized patients: a comparative study of intensive care and general care
units. Crit Care Med 1997;25:1289–1297.
Dimick JB, Pronovost PJ, Heitmiller RF, Lipsett PA. Intensive care unit physician staffing is
associated with decreased length of stay, hospital cost, and complications after esophageal
resection. Crit Care Med 2001a;29:753–758.
Dimick JB, Swoboda SM, Pronovost PJ, Lipsett PA. Effect of nurse-to-patient ratio in the inten-
sive care unit on pulmonary complications and resource use after hepatectomy. Am J Crit
Care 2001b;10:376–382.
Donchin Y, Gopher D, Olin M, et al. A look into the nature and causes of human errors in the
intensive care unit. 1995. Qual Saf Health Care 2003;12:143–147.
Eng J, Mysko WK, Weller GE, et al. Interpretation of emergency department radiographs: a
comparison of emergency medicine physicians with radiologists, residents with faculty,
and film with digital display. AJR Am J Roentgenol 2000;175:1233–1238.
Fassier T, Azoulay E. Conficts and communication gaps in the intensive care unit.Curr Opin
Crit Care 2010;16:1–12.
Flaatten H, Hevroy O. Errors in the intensive care unit (ICU). Experiences with an anonymous
registration. Acta Anaesthesiol Scand 1999;43:614–617.
References 47
Flum DR, Koepsell T. The clinical and economic correlates of misdiagnosed appendicitis:
nationwide analysis. Arch Surg 2002;137:799–804.
Flum DR, Morris A, Koepsell T, Dellinger EP. Has misdiagnosis of appendicitis decreased over
time? A population-based analysis. JAMA 2001;286:1748–1753.
Frey B, Buettiker V, Hug MI, et al. Does critical incident reporting contribute to medication
error prevention? Eur J Pediatr 2002;161:594.
Gatt ME, Spectre G, Paltiel O, Hiller N, Stalnikowicz R. Chest radiographs in the emergency
department: is the radiologist really necessary? Postgrad Med J 2003;79:214–217.
Giraud T, Dhainaut JF, Vaxelaire JF, et al. Iatrogenic complications in adult intensive care
units: a prospective two-center study. Crit Car Med 1993;21:40–51.
Goldschmidt HMJ, Lent RW. Gross errors and work flow analysis in the clinical laboratory.
Klin Biochem Metab 1995;3:131–140.
Gopher D, Olin M, Badihi Y. The Nature and Causes of Human Errors in a Medical Intensive
Care Unit. Proceedings of the 33rd Annual Meeting of the Human Factors Society. Denver:
Human Factors Society; 1989:956–960.
Gratton MC, Salomone III JA, Watson WA. Clinically significant radiograph misinterpretations
at an emergency medicine residency program. Ann Emerg Med 1990;19: 497–502.
Hofgartner WT, Tait JF. Frequency of problems during clinical molecular-genetic testing. Am J
Clin Pathol 1999;112:14–21.
Howanitz PJ, Renner SW, Walsh MK. Continuous wristband monitoring over two years de-
creases identification errors: a College of American Pathologists Q-Tracks study. Arch Pathol
Lab Med 2002;126:809–815.
Institute of Medicine. Crossing the Quality Chasm: A New Health System for the 21st Century.
Washington, DC:National Academy Press; 2001.
Ismail AA, Walker PL, Barth JH, Lewandowski KC, Jones R, Burr WA. Wrong biochemistry
results: two case reports and observational study in 5,310 patients on potentially mis-
leading thyroid-stimulating hormone and gonadotropin immunoassay results. Clin Chem
2002;48:2023–2029.
Kalra J, Medical errors:an introduction to concepts. Clin Biochem 2004:37:1043–1051.
Kalra J, Neufeld H, Knight I, et al. Risk management at the front end of a clinical laboratory:
a retrospective study [Abstract]. Clin Biochem 2004a;37:724.
Kalra J, Saxena A, Mulla A, Neufeld H, Qureshi M, Massey KL. Medical error: a clinical
laboratory approach in enhancing quality care [Abstract]. Clin Biochem 2004b;37:
732–733.
Kalra J, Saxena A, Mulla A, Neufeld H, Qureshi M, Sander R. Medical error and patient safety:
a model for error reduction in pathology and laboratory medicine [Abstract]. Clin Invest
Med 2003;26:215.
Karcz A, Korn R, Burke MC, et al. Malpractice claims against emergency physicians in Mas-
sachusetts: 1975–1993. Am J Emerg Med 1996;14:341–345.
Kazmierczak SC, Catrou PG. Laboratory error undetectable by customary quality control/
quality assurance monitors. Arch Pathol Lab Med 1993;117:714–718.
Khoury M, Burnett L, Mackay MA. Error rates in Australian chemical pathology laboratories.
Med J Aust 1996;165:128–130.
Klein EJ, Koenig M, Diekema DS, Winters W. Discordant radiograph interpretation between
emergency physicians and radiologists in a pediatric emergency department. Pediatr Emerg
Care 1999; 15:245–248.
Kohn LT, Corrigan JM, Donaldson MS, editors. Committee on Quality of Healthcare in Amer-
ica, Institute of Medicine. To Err Is Human: Building a Safer Health System. Washington,
DC:National Academy Press; 2000.
Landis NT. Pharmacist on ICU team cuts ordering errors by 66%. Am J Health Syst Pharm
1999;56:1700.
48 Medical error and strategies for working solutions in clinical diagnostic
Lapworth R, Teal TK. Laboratory blunders revisited. Ann Clin Biochem 1994;31:78–84.
Leape LL. Striving for perfection. Clin Chem 2002;48:1871–1872.
Leape LL, Brennan TA, Laird N, et al. The nature of adverse events in hospitalized patients.
Results of the Harvard medical practice study II. N Engl J Med 1991;324:377–384.
Levey S, Jennings ER. The use of control charts in the clinical laboratory. Am J Clin Pathol
1950;20:1059–1066.
Lord JT. Risk management in pathology and laboratory medicine. Arch Pathol Lab Med 1990;
114:1164–1167.
Lumadue JA, Boyd JS, Ness PM. Adherence to a strict specimen labeling policy decreases
the incidence or erroneous blood grouping of blood bank specimens. Transfusion
1997;37:1169.
Mann FA, Danz PL. The night stalker effect: quality improvements with a dedicated night-call
rotation. Invest Radiol 1993;28:92–96.
Marks V. False-positive immunoassay results: a multicenter survey of erroneous immunoassay
results from assays of 74 analytes in 10 donors from 66 laboratories in seven countries. Clin
Chem 2002;48:2008–2016.
McCarthy BD, Beshansky JR, D’Agostino RB, Selker HP. Missed diagnoses of acute myocar-
dial infarction in the emergency department:results from a multicenter study. Ann Emerg
Med 1993; 22:579–582.
McIntyre JW, Nelson TM. Application of automated human voice delivery to warning devices
in an intensive care unit: a laboratory study. Int J Clin Monit Comput 1989;6:255–262.
McSwiney RR, Woodrow DA. Types of error within a clinical laboratory. J Med Lab Technol
1969;26:340–346.
Morrison AL, Beckmann U, Durie M, Carless R, Gillies DM. The effects of nursing staff inex-
perience (NSI) on the occurrence of adverse patient experiences in ICUs. Aust Crit Care
2001;14:116–121.
Nevalainen D, Berte L, Kraft C, Leigh E, Picaso L, Morgan T. Evaluating laboratory performance
on quality indicators with the six sigma scale. Arch Pathol Lab Med 2000;124:516–519.
Nutting PA, Main DS, Fischer PM, et al. Toward optimal laboratory use. Problems in labora-
tory testing in primary care. JAMA 1996;275:635–639.
Plebani M, Carraro P. Mistakes in a stat laboratory:types and frequency. Clin Chem 1997;
43:1348–1351.
Pope JH, Aufderheide TP, Ruthazer R, et al. Missed diagnoses of acute cardiac ischemia in the
emergency department. N Engl J Med 2000;342:1163–1170.
Pronovost P, Berenholtz S, Dorman T, Lipsett PA, Simmonds T, Haraden C. Improving com-
munication in the ICU using daily goals. J Crit Care 2003;18:71–75.
Pronovost PJ, Dang D, Dorman T, et al. Intensive care unit nurse staffing and the risk for
complications after abdominal aortic surgery. Eff Clin Pract. 2001a;4:199–206.
Pronovost PJ, Waters H, Dorman T. Impact of critical care physician workforce for intensive
care unit physician staffing. Curr Opin Crit Care 2001b;7:456–459.
Renner SW, Howanitz PJ, Bachner P. Wristband identification error reporting in 712 hospitals.
A College of American Pathologists’ Q-Probes study of quality issues in transfusion prac-
tice. Arch Pathol Lab Med 1993;117:573–577.
Reynolds SL. Missed appendicitis in a pediatric emergency department. Pediatr Emerg Care
1993;9:1–3.
Romanchuk K. The effect of limiting residents’ work hours on their surgical training: a Cana-
dian perspective. Acad Med 2004;79:384–385.
Rosen P, Markovchick V, Dracon D. Normative and technical error in the emergency depart-
ment. J Emerg Med 1983;1:155–160.
Rothrock SG, Green SM, Dobson M, Colucciello SA, Simmons CM. Misdiagnosis of appendi-
citis in nonpregnant women of childbearing age. J Emerg Med 1995;13:1–8.
References 49
Rothrock SG, Skeoch G, Rush JJ, Johnson NE. Clinical features of misdiagnosed appendicitis
in children. Ann Emerg Med 1991;20:45–50.
Rusnak RA, Borer JM, Fastow JS. Misdiagnosis of acute appendicitis: common features discov-
ered in cases after litigation. Am J Emerg Med 1994;12:397–402.
Schoenberg R, Sands DZ, Safran C. Making ICU alarms meaningful: a comparison of tradi-
tional vs. trend-based algorithms. Proc AMIA Symp 1999:379–383.
Shirm SW, Graham CJ, Seibert JJ, Fiser D, Scholle SH, Dick RM. Clinical effect of a quality
assurance system for radiographs in a pediatric emergency department. Pediatr Emerg Care
1995;11:351–354.
Sirota RL. The Institute of Medicine ’s report on medical error. Implications for pathology. Arch
Pathol Lab Med 2000;124:1674–1678.
Stambouly JJ, McLaughlin LL, Mandel FS, Boxer RA. Complications of care in a pediatric
intensive care unit: a prospective study. Intensive Care Med 1996;22:1098–1104.
Tarnow-Mordi WO, Hau C, Warden A, Shearer AJ. Hospital mortality in relation to staff work-
load: a four-year study in an adult intensive-care unit. Lancet 2000;356:185–189.
Thomas EJ, Studdert DM, Burstin HR, et al. Incidence and types of adverse events and negli-
gent care in Utah and Colorado. Med Care 2000;38:261–271.
Tissot E, Cornette C, Demoly P, Jacquet M, Barale F, Capellier G. Medication errors at the
administration stage in an intensive care unit. Intensive Care Med 1999;25:353–359.
Urrutia A, Bechini J, Tor J, Olazabal A, Rey-Joly C. Assessment of thoracic X-ray readings by
emergency room physicians at a university hospital. Med Clin (Barc) 2001;117:332–333.
[in Spanish].
Vincer MJ, Murray JM, Yuill A, Allen AC, Evans JR, Stinson DA. Drug errors and incidents in a neo-
natal intensive care unit: a quality assurance activity. Am J Dis Child 1989;143:737–740.
Wilson RM, Runciman WB, Gibberd RW, Harrison BT, Newby L, Hamilton JD. The Quality in
Australian Health Care Study. Med J Aust 1995;163:458–471.
Wiwanitkit V/. Types and frequency of pre-analytical mistakes in the first Thai ISO 9002: 1994
certified clinical laboratory, a six-month monitoring. BMC Clin Pathol 2001;1:5–9.
Wright D, Mackenzie SJ, Buchan I, Cairns CS, Price LE. Critical incidents in the intensive
therapy unit. Lancet 1991;338:676–678.
Zarbo RJ, Jones BA, Friedberg RC, et al. Q-tracks: a College of American Pathologists pro-
gram of continuous laboratory monitoring and longitudinal tracking. Arch Pathol Lab Med
2002;126:1036–1044.
5 Creating a culture for medical error reduction
The issue of medical errors has received substantial attention in recent years. The
Institute of Medicine (IOM) report released in 2000 has several implications for
health care systems in all disciplines of medicine. Notwithstanding the plethora
of available information on the subject, little substantive action has been taken
toward reducing medical error. A principal reason for this may be the stigma as-
sociated with medical errors. An educational program with a practical, informed,
and longitudinal approach offers realistic solutions toward this end. Effective re-
porting systems need to be developed as a medium of learning from the errors and
modifying behaviors appropriately. The presence of a strong leadership supported
by organizational commitment is essential in driving these changes. A national,
provincial, or territorial quality care council created solely for the purpose of en-
hancing patient safety and medical error reduction may be formed to oversee these
efforts. The bioethical and emotional components associated with medical errors
also deserve attention and focus.
5.1 Introduction
In recent years, there has been a groundswell of interest from the public, the health
care community, and government agencies in addressing the issue of medical errors.
It is widely accepted that the present health care system is complex and stressed. This
system leads to unsafe conditions for patient care. Reducing the error rates in health
care is central to improving overall quality in health care delivery. The health care sys-
tem’s unique needs in quality improvement cannot be met without an effective culture
that promotes quality and patient safety. The outcomes in complex work cultures like
health care depend mainly on integration of individuals into teams as well as technical
and other organizational factors (Bogner, 1994). The reduction of health care errors
requires a concerted effort, similar to that undertaken by the commercial aviation and
nuclear power sectors. Unless there is an organizational and institutional commitment
toward achieving the objective of patient safety, any proposed solutions are futile. To
foster a culture of safety, it is necessary to have a culture that supports reporting, justice,
flexibility, and learning (Reason, 1997).
The benefits of such a culture are immense. It creates a medical system in which all
the stakeholders – patients, medical professionals, and insurers – have increased trust
in one another, making confrontations, including costly malpractice claims, less likely
to occur. It lessens the stress experienced by physicians, allowing them to take better
care of patients. It lowers costs because of the drop in malpractice claims and the
reduction in the number of patients who experience extended hospitalization due to
adverse events, raising the system’s efficiency.
52 5 Creating a culture for medical error reduction
The recent attention and initiatives toward patient safety are encouraging but are
not entirely novel. The health care sector has been continuously involved with quality
improvement activities, albeit under different names. However, previous efforts have
delivered only limited results, and tremendous potential for improvement still exists.
A renewed commitment toward quality and safety initiatives is the current need. This
poses significant challenges, and not necessarily all the factors are under the control of
the profession itself (Blumenthal, 1994). A prime role exists for the accreditation and
legislative bodies in promoting patient safety initiatives; patient safety issues should
form one of the principles behind the accreditation standards established for all health
care sectors. The other contributors to initiatives aimed at promoting patient safety
include professional associations, regulatory bodies, health care organizations, and
the educational system. In the ensuing sections, we discuss some solutions that form
the basic framework for quality improvements in health care sectors that can build
on the past efforts of quality improvement and patient safety. We have also reviewed
some essential bioethical concerns and the emotional impact of medical errors on
health care professionals, as they form an integral part of quality improvement and
patient safety initiatives.
physicians and other health care professionals are well prepared with the scientific
base of medical knowledge that equips them to provide care for their individual pa-
tients, but only a relative few develop the skills that are necessary to continuously
strive for quality improvements in everyday practice. Education in this field is essential
for health care professionals to further hone their scientific skills and, ultimately, to
improve their professional practice. Morbidity and mortality due to health care errors
are high, probably higher than they are for some pathological causes of death. These
high numbers underline the need for action. Education about these errors and their
preventability could provide us with some of the basic interventional approaches to
reducing errors and improving quality. This may assume significance when one con-
siders the prevalence of medical mistakes among students and house officers (Estrada
et al., 2000; Wu et al., 1993). Wu et al. (1993) reported that 45% of the house officers
in internal medicine admitted having made a mistake in practice. The authors, using
multivariate analysis of the data, were able to show that house officers who accepted
responsibility for their errors were more likely to make constructive changes to their
practices than were those who coped by escape or avoidance, who were more likely
to report a defensive change in practice. They suggested, among other things, that
medical educators provide specific advice about how to prevent reoccurrences of
mistakes.
We have suggested that this concept be placed in an organized and systematic for-
mat in the learning years through an educational program (Kalra and Collard, 2002).
We have previously suggested that quality care grand rounds be instituted as a forum
for discussion and evaluation of medical errors in a guilt-free and secure environment
aimed at achieving enhanced patient safety (Mulla et al., 2003) (see fFig. 5.1). We
have also suggested and implemented a “no-fault” model to encourage voluntary
anonymous reporting of all critical incidents by laboratory professionals (Kalra et al.,
2003). Our model also includes an educational component that targets all levels of
laboratory professionals, including technologists and residents. Briefing and debriefing
sessions are held periodically to continuously inculcate the culture of safety and error
prevention in laboratory practices. As a part of the process and completion of the loop,
the laboratory staff completes a feedback questionnaire at regular intervals that assesses
staff attitudes, understanding, and satisfaction with the design of the model and the
educational program. The reporting system allows for sufficient scope to review and
critically analyze the incident. It accounts for the systemic factors that contribute to
errors and deemphasizes individual blame or action. The initiative is supported by the
clinical chemistry laboratory professional staff and clearly spells out the management
team’s attitude toward error reporting and reduction.
To achieve industrial-grade quality in clinical medicine is an overwhelming task.
It calls for a total cultural revamp and interdisciplinary cooperation. Satisfactory im-
provements in achieving a culture of quality and patient safety cannot be achieved
without revolutionary changes. There are significant discontinuities in the current health
care system, and these need to be minimized through the adoption of an educational
approach. The imperative should be to improve and teach these improvements while
simultaneously attempting to reform the existing systems. Significant changes can
begin by defining a curriculum to inform and educate (Cosby and Croskerry, 2003). An
educational program may have various components (see fFig. 5.2); these components
need to be emphasized to achieve the maximum benefits from the program.
54 5 Creating a culture for medical error reduction
QUAL
ITY
IM
PR
O
S
VE
CE
M
UR
Simplifying
EN
O Standardizing
T
Complex
SA
RES
Procedures Systems
ND
ADEQUATE
RISK MANAGEM
QUALITY CARE ROUNDS
OF
of Sessions
Timely
L
EN
B I
Information
ILA
T A
A
SE
AV
X
EC
UT
IVE
GO
ALS
Communications
Systematic/
Human Bioethics
issues
PROFESSIONALISM
Error coping
Team work mechanisms
in a system that is fraught with flaws. This should not underemphasize the systemic influ-
ence on error but should emphasize human coping mechanisms that can help minimize
the impact of these deficiencies. A balance can be achieved by creating an academic
training program that works aggressively to remove certain obstacles that house officers
encounter when they look for personal causes of error (Casarett and Helms, 1999).
The third component that needs to be addressed in an educational program relates
to teamwork and work ethics. In today’s complex medical care, no single individual
achieves total patient care. It requires a team effort. The medical care team may con-
sist of one or more physicians, nursing staff, laboratory staff, pharmacists, and medical
trainees. Medicine, over the years, has developed a hierarchical structure, with the
physician on top, an elite position enjoyed by virtue of his knowledge and academic
qualifications (Lester and Tritter, 2001). With no offense intended toward the elite physi-
cian, we suggest that this hierarchical structure of the medical care team does not bode
well for enhancing patient safety. In such a structure, questioning the professional wis-
dom of the physician or reporting an obvious incompetence becomes difficult. Health
care students have to learn that safe and high-quality patient care is best achieved
by small teams, where every member of the team owes the patient the best possible
care. A team-oriented approach that allows constructive suggestions from any team
member should be encouraged in clinical decision making, and this should be a key
point in the educational program. Teamwork in health care systems offers a layered
approach through which it becomes difficult for an error to reach the patient. Similar
teamwork approaches adopted in the form of crew resource management in aviation
56 5 Creating a culture for medical error reduction
have had their desired effect and even convinced skeptics of the value of teamwork
and teamwork training (Johnson, 2001). Multidisciplinary teaching that enables medical
students to appreciate the roles and skills of other health care professionals has also
been emphasized to develop the culture of teamwork (Lester and Tritter, 2001).
Another component of an educational program to reduce health care errors is the de-
velopment of adequate communication skills. Usherwood (1993) showed the influence
of communication skills on clinical consultation behaviors such as active listening and
empathy. Good communication skills constitute a significant gain for students and are
associated with improvements in patient satisfaction (Novack et al., 1992). Developing
these skills is imperative, not only for patient contacts but also for peer interactions and
interactions with the family of the patient. The emphasis on teaching communication
skills has been clearly underlined in the General Medical Council report Tomorrow’s
Doctors (1993). Other important components that are to be addressed in any educational
program include ethical issues and emotional coping after adverse events and errors.
It is essential, too, that development of professionalism be central to the education
of medical practitioners. Physicians, for one, have long eschewed the double roles of
healer and professional (Cruess and Cruess, 1997). It has been observed that genuine
medical professionalism is at peril today (Wynia et al., 1999). The medical curriculum
has periodically renewed its focus on developing the role of a health care practitioner
as a healer, but it has largely ignored the concept of professionalism. This, however, is
changing, and professionalism as a concept is becoming respectable again (Cruess and
Cruess, 2000). A renewed commitment toward this end has been seen in the release of
a new charter for enhancing professionalism in the new millennium (ABIM Foundation,
2002). It is suggested that the concept of professionalism form an integral component
of an educational program to reduce medical errors and quality improvement in health
care processes.
There are various options when it comes to teaching about quality improvement
and error prevention techniques. It is suggested that, whatever the teaching techniques,
education be patient centered and productive. Maximum benefits can be achieved by
training health care professionals in life-like simulation exercises, as suggested for anes-
thesia (Barach, 2000). Gauging the responses of trainees to such situations and training
them on the ideal responses should offer a good learning experience for trainees. Other
innovative training techniques have also been described. Rall et al. (2002) have reported
training techniques that require theoretical medical information as well as practical
training sessions and that use realistic patient simulator systems with videotaping and
interactive debriefing following the training sessions. Other teaching approaches may
include problem-based learning, small-group learning activities, and didactic lectures.
Involvement of students in designing the core teaching topics and courses may stimulate
motivated and active participation.
This educational program should be implemented at all levels and should not con-
fine itself to undergraduate and postgraduate teaching curriculums. Practitioners also
stand to benefit immensely from such a program. The curriculum on quality improve-
ment and patient safety should form a part of continuing medical education. The regula-
tory and licensing bodies should be actively involved and should mandate adherence.
Appropriate scheduling and crediting of sessions should be encouraged. Adequate
incentives and resources should be dedicated to promoting periodic education and
attendance at the sessions.
5.3 Error reporting systems 57
The level at which the reporting systems should be implemented and operated is
highly debatable. A national reporting system of all adverse events and events averted
at the last minute has been advocated for improving patient safety (Kohn et al., 2000).
Doubts exist, however, over the feasibility of such a system, in terms of both require-
ments and economics (Leape, 2002). Local reporting systems, though highly practical
and operational, may not serve the purpose. Locally based systems may increase report-
ing and offer better possibilities for direct feedback, but these may have a smaller data-
base. Another possibility is that local reports may be restricted to only some categories
in which errors occur at such low frequencies that generating sufficient information for
prophylactic action may be impossible. Under these circumstances, the best compro-
mise may be specialty-restricted reporting systems that allow practitioners to glean the
maximum advantage of information from the data generated (Wu et al., 2002). Certain
other voluntary systems, such as the Joint Commission on Accreditation of Healthcare
Organization (JCAHO) sentinel event reporting program, the MedMARx program of the
United States Pharmacopeia, the National Nosocomial Infection Survey of the Centers
for Disease Control, and the Institute for Safe Medical Practices’ reporting program for
the prevention of medication error are already being practiced with success (Leape,
2002).
There are numerous national and local organizations that can assume leadership roles
and implement programs; alternatively, a separate commission or institute dedicated
solely to the pursuit of quality care and patient safety could be established.
The CSPI also provides assistance for patients and families, advising patients to ask
questions and to talk and listen to their team of healthcare providers to promote com-
munication and health. After all, doctors may fail to ask potentially lifesaving questions
or fully clarify medical information to patients. Patients must be aware that many of
doctors’ personal questions are legitimate, and they too should not feel inhibited about
asking questions. The only way to ensure that a doctor or health care provider does not
forget to ask important questions is often for patients to ask such questions themselves.
CPSI encourages patients, for examples, to ask questions such as “Why are you prescrib-
ing this medication, what are the side effects, how will the medication will help me,
and when should I take the medication?” (Canadian Patient Safety Institute, 2007). As
pharmacist Melanie Rantucci explains, “The patient is important to bringing the com-
munity and hospital systems together and in keeping the communication flowing during
ambulatory care. The patient is the only common denominator in the process and efforts
need to be made to inform and engage patients in medication reconciliation” (Canadian
Patient Safety Institute, 2007). This role for the patient must be supported by multifaceted
educational and informational promotions by medical professionals themselves and by
a public awareness campaign.
5.7 Conclusion
Provision of safe patient care is a major challenge confronting today’s health care
system. It is imperative that development of a culture of safety be accelerated. Well-
designed patient safety initiatives based on systematic interventions may produce the
62 5 Creating a culture for medical error reduction
greatest enhancements in the quality of health care processes. These initiatives must be
adequately integrated into organizational policies as they are developed. There is also a
pressing need for uniform, well-defined policies to address the bioethical component of
medical errors, particularly disclosure of errors and the emotional issues attached with
them. Providing education to patients to enable them to better take charge of their own
health care experience is one such necessary change. Ensuring adequate leadership
and developing this leadership are good first steps as we begin to address the gaps in
patient safety and change the current cultural climate.
References
ABIM Foundation. Medical professionalism in the new millennium: a physician charter. Ann
Intern Med 2002;136:243–246.
Alberti KG. Medical errors: a common problem. BMJ 2001;322:501–502.
Barach P. Patient safety curriculum. Acad Med 2000;75:551–552.
Bhasale A. The wrong diagnosis: identifying causes of potentially adverse events in general
practice using incident monitoring. Fam Pract 1998;15:308–318.
Blumenthal D. Making medical errors into “medical treasures”. JAMA 1994;272:1867–1868.
Bogner MS. Human Error in Medicine. Hillsdale: Erlbaum; 1994.
Bosk CL. Continuity and change in the study of medical error – the culture of safety on the
shop floor. 2005. http://www.sss.ias.edu/files/papers/paper20.pdf. Accessed July 24, 2010.
Canadian Patient Safety Institute. Annual Review 2010.
Canadian Patient Safety Institute. “Tips for patients and families: ask. Talk. Listen.” 2007. http://
www.patientsafetyinstitute.ca/English/toolsResources/patientsAndTheirFamilies/Pages/
PatientTips.aspx. Accessed July 25, 2010.
Casarett D, Helms C. Systems errors versus physicians’ errors: finding the balance in medical
education. Acad Med 1999;74:19–22.
Cohen MR. Why error reporting systems should be voluntary. BMJ 2000;320:728–729.
Cosby KS, Croskerry P. Patient safety: a curriculum for teaching patient safety in emergency
medicine. Acad Emerg Med 2003;10:69–78.
Croskerry P, Wears RL, Binder LS. Setting the educational agenda and curriculum for error
prevention in emergency medicine. Acad Emerg Med 2000;7:1194–1200.
Cruess SR, Cruess RL. Professionalism: a contract between medicine and society. CMAJ
2000;162:668–669.
Cruess SR, Cruess RL. Professionalism must be taught. BMJ 1997; 315:1674–1677.
Cullen DJ, Bates DW, Small SD, Cooper JB, Nemeskal AR, Leape LL. The incident reporting
system does not detect adverse drug events: a problem for quality improvement. Jt Comm
J Qual Improv 1995;21:541–548.
Eichhorn JH. Prevention of intraoperative anesthesia accidents and related severe injury
through safety monitoring. Anesthesiology 1989; 70:572–577.
Estrada CA, Carter J, Brooks C, Jobe AC. Reducing error, improving safety. Medical errors must
be discussed during medical education. BMJ 2000;321:507–508.
Figueiras A, Tato F, Fontainas J, Gestal-Otero JJ. Influence of physicians’ attitudes on reporting
adverse drug events: a case-control study. Med Care 1999;37:809–814.
Flanagan JC. The critical incident technique. Psychol Bull 1954;51:327–358.
Gaynes R, Richards C, Edwards J, et al. Feeding back surveillance data to prevent hospital-
acquired infections. Emerg Infect Dis 2001;7:295–298.
General Medical Council. Tomorrow’s Doctors: Recommendations on Undergraduate Medi-
cal Education. London: GMC; 1993.
References 63
Hebert PC, Levin AV, Robertson G. Bioethics for clinicians: 23. Disclosure of medical error.
CMAJ 2001;164:509–513.
Hughes EC. “Mistakes at Work” in the Sociological Eye: Selected Papers on Work, Self, and
Society. Chicago: Aldine-Atherton; 1971.
Institute of Medicine. Crossing the Quality Chasm: A New Health System for the 21st Century.
Washington, DC: National Academy Press; 2001.
Johnson D. How the Atlantic barons learnt teamwork. BMJ 2001;322:563.
Johnson N. Patient Safety Improvement Act of 2002, H.R 4889, 107th Cong.
Joint Commission on Accreditation of Health Care Organizations (JCAHO). Joint Commis-
sion president outlines top strategic priorities to aid QI efforts. Hosp Peer Rev 2001a;
26:73–76.
Joint Commission on Accreditation of Health Care Organizations (JCAHO). JCAHO patient
safety standards stress leadership. Hosp Case Manag 2001b;9:1–3 [Suppl.].
Kalra J, Collard D. Medical error: a need for an educational program [Abstract]. RCPSC An-
nals 2002;35:13 [Suppl].
Kalra J, Saxena A, Mulla A, et al. Medical error and patient safety: a model for error reduction
in pathology and laboratory medicine [Abstract]. Clin Invest Med 2003;26:15.
Kohn LT, Corrigan JM, Donaldson MS, editors. Committee on Quality of Healthcare in Amer-
ica, Institute of Medicine. To Err Is Human: Building a Safer Health System. Washington,
DC: National Academy Press; 2000.
Leape LL. Reporting of adverse events. N Engl J Med 2002;347:1633–1638.
Lester H, Tritter JQ. Medical error: a discussion of the medical construction of error and sug-
gestions for reforms of medical education to decrease error. Med Educ 2001;35:855–861.
Liu BA, Knowles SR, Mittmann N, et al. Reporting of fatal adverse drug reactions. Can J Clin
Pharmacol 2001;8:84–88.
Lowery DW, Heilpern KL, Otsuki JA. An interdisciplinary curriculum for reducing medical
errors in the academic medical center [Abstract]. Acad Emerg Med 2001;8:586–7.
Mizrahi T. Managing medical mistakes: ideology, insularity and accountability among
internists-in-training. Soc Sci Med 1984;19:135–146.
Mulla A, Massey KL, Kalra J. Does every patient benefit from care? Quality care rounds, a
patient safety initiative [Abstract]. Clin Invest Med 2003;26:215.
Newman MC. The emotional impact of mistakes on family physicians. Arch Fam Med 1996;
5:71–75.
Novack DH, Dube C, Goldstein MG. Teaching medical interviewing. A basic course
on interviewing and the physician-patient relationship. Arch Intern Med 1992;
152:1814–1820.
O’Leary DS. Accreditation’s role in reducing medical errors. West J Med 2000;172:357–358.
Pierce Jr. EC. The 34th Rovenstine Lecture. 40 years behind the mask: safety revisited. Anes-
thesiology 1996;84:965–975.
Pilpel D, Schor R, Benbassat J. Barriers to acceptance of medical error: the case for a teaching
program (695). Med Educ 1998;32:3–7.
Rall M, Schaedle B, Zieger J, et al. Innovative training for enhancing patient safety. Safety
culture and integrated concepts. Unfallchirurg 2002;105:1033–1042. [in German]
Reason J. Managing the Risks of Organizational Accidents. Aldershot: Ashgate; 1997.
Reinertsen JL. Let’s talk about error. BMJ 2000;320:730.
Stewart DW. Conflict in fiduciary duty involving health care error reporting. Medsurg Nurs
2002;11:187–1191.
Uribe CL, Schweikhart SB, Pathak DS, et al. Perceived barriers to medical-error reporting: an
exploratory investigation. J Healthc Manag 2002;47:263–279.
Usherwood T. Subjective and behavioural evaluation of the teaching of patient interview
skills. Med Educ1993;27:41–47.
64 5 Creating a culture for medical error reduction
The majority of errors made in the health care industry are systemic and not in-
dividual in nature, but, unless the errors are brought to a supervisor’s attention, it
will be difficult to correct a flaw in the system. We also recognize that many health
care professionals, despite being safe from legal action, will be hesitant to admit
errors that could damage their careers and reputations unless the disclosure policy
is also to a large extent nonpunitive. This allows physicians to take leadership roles
and accept responsibility without fear of undue reprisal. The correction of flawed
medical systems and the subsequent protection of patients’ health should be the
industry’s top priority, rather than subjecting physicians who make inevitable errors
to punitive measures. Inappropriate blame attribution to serve regulatory needs will
merely alienate professionals and discourage them from participating in system
improvement.
6.1 Introduction
Clinical diagnostic laboratories play a critical role in the diagnosis of many human
diseases; they have been doing so for decades now (Kotlartz, 1998). Laboratory testing
influences a majority of clinical decision making. With laboratories having such a high
degree of influence, it is easy to see the importance of quality in laboratory testing.
In today’s health care environment of managed care and cost-containment processes,
laboratorians have to work collaboratively with other health care professionals, with
the sole focus being improvement in medical outcomes. It has been suggested that the
importance of laboratory scientists must be proven in order to guarantee the quality of
tests and thereby improve the quality of laboratory services (Plebani, 1999). In a health
care delivery system that is interdependent with other departments, where the quality
of one department has an effect on another, it becomes essential for clinical diagnostic
laboratories to set high standards for other departments to follow.
The solution to medical error requires a systemwide overhaul, which should also ad-
dress the tendency to blame individuals’ actions. Yet, when errors do occur, procedures
must be in place to ensure that valuable information about what went wrong can be
used to make processes safer, instead of being concealed for fear of liability. Doing so
would require a system of assurance for health professionals, covering both educational
and legal aspects. Errors continue to happen in all three stages of analysis and include
human and technical factors. Already, proficiency testing agencies, one of the best ex-
ternal quality assessment methods, are delivering standards for accuracy, though all
assessment methods are not of equal value and all have limitations. We will first address
the development of quality management programs and the extent of error in clinical
diagnostic laboratories, followed by a discussion of the other quality assurance systems;
finally, we will propose a no-fault model for clinical laboratories.
66 6 Improving quality in clinical diagnostic laboratories
attributed the large difference in the two settings primarily to the lower skill level of
ward staff in blood drawing, the greater complexity of tests performed, and the higher
frequency of blood drawings for inpatients.
More recently, Ismail et al. (2002) reported a total error rate of 0.53% in the analytical
phase of common immunoassay tests for thyroid stimulating hormones and gonadotro-
pins. The authors concluded that these errors were a result of analytical interference and
stressed the importance of early interference in cases where results were not compatible
with the clinical scenario. Marks (2002) studied the influence of analytical interference
on assays of 74 analytes. A total of 66 laboratories across seven countries participated
in the study. Marks found that 8.7% of the results were erroneous, and 49% of these
erroneous results were not corrected even by the addition of a blocking reagent.
Tab. 6.1: Review of error rates in clinical diagnostic laboratories represented in DPM
opportunities.
Notes:
NA = Not Applicable.
1
Different authors have used different terminologies (i.e., errors, mistakes, blunders, outliers,
problems, unacceptable results).
2
Inspected laboratories.
3
Data collected through a survey questionnaire.
68 6 Improving quality in clinical diagnostic laboratories
though the criteria they used to classify errors could be termed rather strict. Wiwanitkit
studied the types and frequency of preanalytical errors in a large hospital laboratory and
found that approximately 85% of errors occurred in the preanalytical stage, whereas a
mere 4.35% of errors occurred in the analytical stage.
The quality cycle in a laboratory is not dependent on the control of analytical pro-
cesses alone. The precision and accuracy of reported laboratory results are also de-
pendent on the accuracy of the preanalytical and postanalytical stages of the testing
process. The high rates of pre- and postanalytical errors necessitate the involvement of
nonlaboratory personnel, including clinicians, in order to improve the quality of labora-
tory results. The heavy balance of errors occurring in the pre- and postanalytical stages
of a laboratory testing process reconfirms the susceptibility of the analytical process to
human error. It has been estimated that up to 97% of mistakes occurring in laboratory
processes result from human error (Goldschmidt et al., 1995). It is therefore suggested
that clinical laboratories employ automation and robotics and minimize human involve-
ment in the process wherever possible. Laboratory automation provides for standardized
workflow and helps eliminate many error-prone steps undertaken by humans. In doing
so, it provides an opportunity for processes to reduce the influence of human factors
such as stress, fatigue, negligence, and cognitive impairment. In addition, it enhances
the quality of laboratory results and reduces the turnaround times for results.
In order to properly screen and diagnose patients, individual results must be ref-
erenced against values or cut-off values. Any purely analytical results do not provide
all the required information. To monitor patient situations, however, the follow-up of
results can be accomplished only via analytic and intra-individual information. This
turnaround time is an important factor in patient care.
Quality Care
Council/
Quality Care
Quality Control/ Committee
Quality Improvement
Program APPROPRIATE
Division Head/ Physician ACTION
Lab Manager
RECOMMENDED
Departmental/
Divisional
Quality Care
Committee
Volunteer Error
Incident Report Lab Staff
ERROR
Fig. 6.1: No-fault model for error reduction and patient safety in laboratories.
6.6 Conclusion 73
involved in the error are not subjected to punitive measures (expectations will be made
in special cases of gross negligence, as when a health care professional’s exceptional
failure to use reasonable care results in damage or injury to another, and in these cases
the error report will be kept confidential).
The report is disclosed to the division head, who takes the appropriate action with
laboratory staff, physicians, and the quality care committees that were previously dis-
cussed in this chapter. The division head then contacts the laboratory advisory commit-
tee, which remains in communication with the quality care committee, as well as the
medical advisory committee. The medical advisory committee records and analyses
the error and then recommends actions both downwards, to the laboratory advisory
committee and the national patient safety and quality care council, and upwards, to
the regional quality care council, the hospital board or appropriate regional authority,
and the national patient safety and quality care council committee. This final commit-
tee is responsible for acting as an interface between the government and health care
providers. It promotes research and education, offers evolving policies and guidelines
for quality control standards, and recommends and implements action. The Canadian
Patient Safety Institute, discussed in chapter 5, performs a job similar to the proposed
national council.
Such a model necessitates the integration of certain laboratory activities. First, it
requires collaboration between laboratories and industry, professional, accreditation,
and regulatory bodies. Second, it calls for clinical laboratory science professional edu-
cation in error prevention and reporting. Third, it depends on clear and effective com-
munication practices among the all members of the laboratory. Moreover, it requires
an emphasis on increased automation and robotics in order to reduce human error.
Ultimately, in addition to a model that can address fault and blame, proficiency testing
should take a central role in laboratories.
6.6 Conclusion
Clinical laboratories should be able to attain peak performance levels in terms of both
medical error reduction and cost effectiveness. This requires modern quality manage-
ment techniques focused on enhancing patient safety. Management must adopt an ap-
proach of designing safer systems, evaluating the success and benefits of the systems via
proficiency testing programs, and practicing these systems efficiently. Quality assurance
management programs, which found their footing in clinical diagnostic laboratories,
have developed and improved since the 1950s. Yet, studies continue to show that great
numbers of errors are occurring, for a variety of reasons, in all stages of the analysis
process. Programs such as EQA and PT are promising starts to achieve greater safety,
but they can be much improved with the addition of no-fault programs that better deal
with errors if they do occur.
References
Berry K, Krizek B. Root cause analysis in response to a “near miss.” J Healthc Qual
2000;22:2–12.
Bonini P, Plebani M, Ceriotti F, Rubboli F. Errors in laboratory medicine. Clin Chem
2002;48:691–698.
74 6 Improving quality in clinical diagnostic laboratories
Chambers AM, Elder J, O’Reilly DS. The blunder rate in a clinical biochemistry service. Ann
Clin Biochem 1986;23:470–473.
Clinical Laboratory Improvement Act, 1988. Association of American Physicians and Sur-
geons. http://www.aapsonline.org/msas/clia.php. Accessed July 26, 2010.
Fahey JL, Aziz N, Spritzler J, et al. Need for external proficiency testing program for cy-
tokines, chemokines, and plasma markers of immune activation. Clin Diag Lab Immun
2000;7:540–548.
Goldschmidt HMJ, Lent RW. Gross errors and workflow analysis in the clinical laboratory.
Klin Biochem Metab 1995;3:131–140.
Grannis GF, Grumer HD, Lott JA, Edison JA, McCabe WC. Proficiency evaluation of clinical
chemistry laboratories. Clin Chem 1972;18:222–236.
Hofgartner WT, Tait JF. Frequency of problems during clinical molecular-genetic testing. Am J
Clin Pathol 1999;112:14–21.
Ismail AA, Walker PL, Barth JH, Lewandowski KC, Jones R, Burr WA. Wrong biochemistry
results: two case reports and observational study in 5,310 patients on potentially mis-
leading thyroid-stimulating hormone and gonadotropin immunoassay results. Clin Chem
2002;48:2023–2029.
Kazmierczak SC, Catrou PG. Laboratory error undetectable by customary quality control/
quality assurance monitors. Arch Pathol Lab Med 1993;117:714–718.
Khoury M, Burnett L, Mackay MA. Error rates in Australian chemical pathology laboratories.
Med J Aust 1996;165:128–130.
Kotlarz VR. Tracing our roots: the broadening horizons of clinical laboratory practice
(1945–62). Clin Lab Sci 1998;11:339–345.
Lapworth R, Teal TK. Laboratory blunders revisited. Ann Clin Biochem 1994;31:78–84.
Levey S, Jennings ER. The use of control charts in the clinical laboratory. Am J Clin Pathol
1950; 20:1059–1066.
Libeer JC. Role of external quality assurance schemes in assessing and improving quality in
medical laboratories. Clin Chim Acta 2001;309:173–177.
Marks V. False-positive immunoassay results: a multicenter survey of erroneous immunoassay
results from assays of 74 analytes in 10 donors from 66 laboratories in seven countries. Clin
Chem 2002;48:2008–2016.
McSwiney RR, Woodrow DA. Types of error within a clinical laboratory. J Med Lab Technol
1969;26:340–346.
Miller WG. Specimen materials, target values and commutability for external quality assess-
ment (proficiency testing) schemes. Clinica Chimica Acta 2003;327:25–37.
Nevalainen D, Berte L, Kraft C, Leigh E, Picaso L, Morgan T. Evaluating laboratory performance
on quality indicators with the six sigma scale. Arch Pathol Lab Med 2000;124:516–519.
Plebani M. The changing face of clinical laboratories. Clin Chem Lab Med 1999;37:711–717.
Plebani M, Carraro P. Mistakes in a stat laboratory: types and frequency. Clin Chem
1997;43:1348–1351.
Rex JH, Turnbull JE, Allen SJ, vande Voorder K, Luther K. Systemic root cause analysis of
adverse drug events in an tertiary referral hospital. J Qual Improv 2000;26:563–575.
Spath P. Medical errors: root cause analysis. Or Manager 1998; 14(9):38–41.
Spath P. Your RCA: how do you select events for analysis? Hosp Peer Rev 1999:97–99.
Stöckl D, Reinauer H. Candidate reference methods for determining target values for cho-
lesterol, creatinine, uric acid, and glucose in external quality assessment and internal ac-
curacy control. Clin Chem 1993;39:993–1000.
US Code of Federal Regulations. Public Health-Laboratory Requirements, Title 42, Part
493.801. Condition: enrollment and testing of samples, laboratory requirements (Subpart
H), 2010. http://ecfr.gpoaccess.gov/cgi/t/text. Accessed July 26, 2010.
References 75
Westgard JO, Klee GG. Quality Assurance. In: Tietz NW, ed. Fundamentals of Clinical Chem-
istry. 3rd ed. Philadelphia: W. B. Saunders; 1987:238–253.
Wilson RM, Runciman WB, Gibberd RW, Harrison BT, Newby L, Hamilton JD. The Quality in
Australian Health Care Study. Med J Aust 1995;163:458–471.
Witte DL, VanNess SA, Angstadt DS, Pennell BJ. Errors, mistakes, blunders, outliers, or unac-
ceptable results: how many? Clin Chem 1997;43:1352–1356.
Wiwanitkit V. Types and frequency of pre-analytical mistakes in the first Thai ISO 9002: 1994
certified clinical laboratory, a 6-month monitoring. BMC Clin Pathol 2001;1:5–9.
7 Barriers to open disclosure
7.1 Introduction
In this chapter we offer recommendations for how to initiate and successfully utilize
disclosure to improve communication while minimizing negative legal or liability re-
percussions. Failure to disclose adverse events to a patient harms patients in two distinct
ways. First, nondisclosure compromises the patient’s ability to undergo corrective medi-
cal treatment to mitigate the error. If error is not disclosed to patients in a timely fashion,
certain injuries sustained due to medical error may no longer be reversible, resulting
in permanent disability or even death. Second, it means that patients are unaware of
compensation that they may be entitled to, whether through civil action or a compensa-
tion package designed by health services for such an event. Patients should not have to
pay full fees for medical treatment gone awry, nor should they have to pay for corrective
action necessitated by medical error. Nondisclosure may force patients to do both.
around 5% of the examined cases, although the other mistakes did not impact patient
treatment. Both affected patients were immediately notified, and parts of the completed
report were later made public (Canadian Broadcasting Corporation, 2008).
Disclosure is a process; the discloser must avoid speculating as to how the error
occurred and simply state what is known at the time. The nature, severity, and cause of
the unanticipated outcome should be presented in a straightforward and nonjudgmen-
tal manner: individual blame should not be meted out. The discloser should express
regret about the unanticipated result on behalf of the entire medical team and should,
moreover, educate the patient, as well as the patient’s family, about the clinical implica-
tions of the unanticipated outcome and the medical team’s plan for the patient’s future
medical care. Any questions or concerns put forth by the patient should be answered to
the best of the discloser’s ability.
It is also important that the disclosure be followed up with a second conversation with
the patient, preferably conducted by the same individual who performed the original
disclosure. The patient should be given all available information on the adverse event
investigation and told of the progress of any new medical treatment. The patient and fam-
ily should be offered an opportunity to discuss the issue with other relevant health care
professionals, who can offer additional information as well as a second opinion on the
matter. Moreover, all previously queries left unanswered because of a lack of contextual
knowledge on the discloser’s part should be responded to.
Wu et al. (1991) reported that nearly 76% of house officers have chosen not to disclose
a serious error to their patients.
Though research on error disclosure is limited, there is little evidence to suggest that
the phenomenon of not disclosing errors to patients has changed over the years. Blendon
et al. (2002) reported that only 30% of the patients who had suffered from an error ac-
knowledged having been informed about the error by their physician. Not disclosing an
error during the course of patient care may amount to compromising patients’ informed
consent. Robertson (1987) reported on the necessity of providing information regarding
an error in order before a patient can give consent for treatment of injury caused by an
error. This lapse on the part of the health care professional breaches an unspoken under-
standing between the professional and her patients to act in the patient’s best interests.
Professional ethics aside, Hebert et al. (2001) commented that patients have a right to
information about errors by virtue of being respected as a person. Patients have expecta-
tions for their physicians regarding disclosure (Gallagher et al., 2003). Patients expect
information about what happened, why the error occurred, how the consequences will
be managed, and how a recurrence of the error will be prevented.
Health care errors take a severe toll on patients and families. If information about
the adverse event is withheld, it only adds to patients’ distress and agony. The failure to
disclose information on medical mistakes adversely affects patient’s decision making,
impairs the patient’s trust in his physician, and increases the chances for a malpractice
suit (Braddock III et al., 1999; Hikson et al., 1992; Witman et al., 1996). In fact, 24%
of those who filed malpractice suits did so because they believed that their physician
was dishonest and had covered up important information (Hickson et al., 1992). As an
incentive to health care professionals, evidence exists that punitive actions may not
necessarily be initiated against them if their errors are honestly acknowledged, and
appropriately disclosed (Heilig, 1994; Witman et al., 1996).
A noteworthy example is Wighton v. Arnot, an Australian court case from 2005 (New
South Wales Supreme Court, 2005). A patient’s nerve was severed during an operation,
but no investigation or disclosure of this adverse event, which was suspected by the
surgeon, ever took place. It was not the treatment that the plaintiff received during the
operation – including severing the nerve – however, that was the basis of the negligence
assertion. The defendant was alleged to have been negligent in the sense that he had
failed to do three things: inform the patient of his suspicion that the nerve had been
served; perform an investigation of the suspected incident; and assist the patient in
receiving timely corrective surgery. The irony, as Madden and Cockburn (2005) write, is
that “the doctor would not have been found liable in negligence had he disclosed the
adverse event to the patient.”
Although a risk management policy that provides patients with the greatest possible
amount of information should also enable patients to maximize liability claims, the
Lexington hospital’s payouts were moderate when compared to those of similar facili-
ties analyzed by Kraman and Hamm (1999). The authors contribute this primarily to a
strengthened fiduciary patient-physician relationship, under which plaintiffs are more
willing to negotiate a settlement on the basis of damages incurred by the hospital and
choose not to further financially punish their health care providers. Settlements reached
out of court are also less costly than the litigation process, which can stretch over many
years and involve many expenses, including fees for expert witnesses and litigation
teams and incidental expenses.
claims that “the malpractice system performs reasonably well in its function of separat-
ing claims without merit from those with merit and compensating the latter,” with only
16% of claims determined to be valid going uncompensated (Studdert et al., 2006).
Nevertheless, the overhead cost of litigation remains high. The total cost of litigating the
claims was 54% of the compensation paid to plaintiffs (Studdert et al., 2006). The fact
that the majority of these costs were incurred in resolving claims in which medical error
did transpire requires that the system be made more effective if costs are to be reduced;
the average claim was resolved five years after the initial injury occurred. This appears
to be an area that needs improvement, rather than seeking tort reforms to block frivolous
claims. Although some industry leaders call for a cap on financial damages awarded
to patients, such a policy would do nothing to ensure fair compensation to patients. It
would also fail to improve their safety, which must remain the fundamental issue.
Former senators Hillary Clinton and Barack Obama (2006) identify four goals that an
effective tort system would achieve. It would reduce the rates of preventable adverse
events; improve the fiduciary patient-physician relationship by promoting open commu-
nication; provide fair compensation for deserving victims of medical error; and reduce
liability insurance premiums for health care institutions. If more effective patient safety
initiatives can be developed, then fewer malpractice cases will have the opportunity to
go to court, reducing the risk of liability and thus driving down insurance premiums.
Mandating open and complete disclosure concerning important medical information
will also decrease the number of lawsuits filed.
7.8 Conclusion
The culture of malpractice suits continues to grow. Suits filed solely for monetary con-
siderations abuse the tort system and set an unacceptable trend (Kalra et al., 2004).
Blame and retribution may have their place, but society’s interests are best served by
creating a trusting environment that promotes honest disclosure of error. To restore trust
and perhaps to lower malpractice claims, both the public and health care providers
must avoid the “shame and blame” game. Another challenge lies in achieving a bal-
ance between a nonpunitive approach to error and the need for a process that includes
accountability and suitable compensation for patients. We suggest that a system-based
error disclosure program can achieve this balance.
References
Blendon RJ, DesRoches CM, Brodie M, Benson, et al. Views of practicing physicians and the
public on medical errors. N Engl J Med 2002;347:1933–1940.
Braddock III CH, Edwards KA, Hasenberg NM, Laidley TL, Levinson W. Informed decision
making in outpatient practice: time to get back to basics. JAMA 1999;282:2313–2320.
Breen, K, N.L. Breast-cancer patients reach $17.5M settlement over botched tests. St. John’s Tele-
gram, October 31, 2009. http://www.canada.com/health/breast+cancer+patients+reach+
settlement+over+botched+tests/2168107/story.html. Accessed August 6, 2010.
Canadian Broadcasting Corporation. Review finds pathologist’s rate of errors within average.
June 26, 2008. http://www.cbc.ca/canada/manitoba/story/2008/06/26/pathologist-review.
html#socialcomments. Accessed August 6, 2010.
Clinton HR, Obama B. Making patient safety the centerpiece of medical liability reform.
N Eng J Med 2005;354:2205–2208.
References 85
Cohen JR. Apology and organizations: exploring an example from medical practice. Fordham
Urban Law 2000;27:1447–1482.
Crawshaw R, Rogers DE, Pellegrino ED, et al. Patient–physician covenant. JAMA 1995;
273:1553.
Davis P, Lay-Yee R, Briant R, Ali W, Scott A, Schug S. Adverse events in New Zealand public
hospitals I: occurrence and impact. N Z Med J 2002;115(1167):U271.
Donaldson, L. The HSJ interview: Professor Sir Liam Donaldson. Rock and role. Interview by
Alastair McLellan. Health Serv J. 2003;113:20–21.
Fla Stat §90.4026.
Gallagher TH, Waterman AD, Ebers AG, Fraser VJ, Levinson W. Patients’ and physicians’ at-
titudes regarding the disclosure of medical errors. JAMA 2003;289:1001–1007.
Gorovitz S, MacIntyre A. Toward a theory of medical fallibility. Hastings Cent Rep
1975;5:13–23.
Hebert PC, Levin AV, Robertson G. Bioethics for clinicians: 23. Disclosure of medical error.
CMAJ 2001;164:509–513.
Heilig S. Honest mistakes. From the physician father of a young patient. Camb Q Healthc
Ethics 1994;3:636–638.
Hickson GB, Clayton EW, Gethens PB, Sloan FA. Factors that prompted families to file medi-
cal malpractice claims following perinatal injuries. JAMA 1992;267:1359–1363.
Kalra J, Saxena A, Mulla A, Neufeld H, Qureshi M, Massey K. Medical error: a clini-
cal laboratory approach in enhancing quality care [Abstract]. Clin Biochem 2004;
37:732–733.
Kohn LT, Corrigan JM, Donaldson MS, editors. Committee on Quality of Healthcare in Amer-
ica, Institute of Medicine. To Err Is Human: Building a Safer Health System. Washington,
DC: National Academy Press; 2000.
Kraman SS, Hamm G. Risk management: extreme honesty may be the best policy. Ann Intern
Med 1999;131:963–967.
Liebman CB, Hyman CS. A mediation skills model to manage disclosure of errors and adverse
events to patients. Health Affairs 2004;23(4):22–32.
Localio AR, Lawthers AG, Brennan TA, et al. Relation between mapractice claims and ad-
verse events due to negligence: results of the Harvard Medical Practice Study III. NEJM
1991;325:245–251.
Madden B, Cockburn T. Duty to disclose medical error in Australia. Australian Health Law
Bulletin 2005;14:13–20.
Mass Gen Laws ch 233, §23D.
Mazor KM, Simon SR, Yood RA, et al. Health plan members’ views about disclosure of medi-
cal errors. Ann Intern Med 2004;140:409– 418.
Mello MM, Studdert DM, Brennan TA. The new medical malpractice crisis. N Engl J Med
2003;348:2281–2284.
New South Wales Supreme Court. 637 (Wighton v Arnot) 2005; BC200504663.
Paterick ZR, Paterick BB, Waterhouse BE, Paterick TE. The challenges to transparency in
reporting medical errors. J Patient Saf 2009;5(4):205-209.
Phillips RL, Dobey SM, Graham D, Elder NC, Hickner JM. Learning from different lenses:
reports of medical errors in primary care by clinicians, staff, and patients. K Patient Saf
2006;2:140 –146.
Reifsteck SW. Difficult physician-patient relationships. Med Group Manage J 1998;45:
46–50, 54.
Robertson GB. Fraudulent concealment and the duty to disclose medical mistakes. Alberta
Law Rev 1987;25:215–223.
Smith ML, Forster HP. Morally managing medical mistakes. Camb Q Healthc Ethics
2000;9:38–53.
86 7 Barriers to open disclosure
Studdert DM, Mello MM, Gawande AA, et al. Claims, errors, and compensation payments in
medical malpractice litigation. NEJM 2006; 354[19]:2024 –2033.
Tamblyn R, Abrahamowicz M, Dauphinee D, et al. Physician scores on a national clini-
cal skills examination as predictors of complaints to medical regulatory authorities. JAMA
2007;298:993–1001.
Wensing M, Elwyn G. Methods for incorporating patients’ views in health care. BMJ
2003;326:877–879.
Witman AB, Park DM, Hardin SB. How do patients want physicians to handle mis-
takes? A survey of internal medicine patients in an academic setting. Arch Intern Med
1996;156:2565–2569.
Wu AW, Folkman S, McPhee SJ, Lo B. Do house officers learn from their mistakes? JAMA
1991;265:2089–2094.
8 International laws and guidelines addressing
error and disclosure
In any health care process, some error is inevitable. As indicated in the US Institute
of Medicine’s report To Err Is Human (Kohn et al., 2000), the challenge is to cut
the rate of error to a minimum. In Canada, various strategies are being applied to
this end, and the federal government has established a Patient Safety Institute. The
UK likewise has a National Patient Safety Agency. However, in the many countries
where efforts are being made to reduce adverse events and errors, a neglected
issue is honest disclosure to the patient or family. In this chapter we examine the
central issues and dilemmas concerning “apology” and suggest how we might work
toward a systematic and effective process.
8.1 Introduction
National, intranational, and international initiatives and advances to address disclosure
following medical error show a range of attitudes toward patients’ right-to-know and
toward institutional and health care providers’ immunity and responsibility. Differences
in legal, social, economic, and health care systems, social and cultural factors, and
access to health services, which have historically differed among nations, underpin
patient experiences of care and expectations of quality and disclosure.
The rate of adverse events in hospital patients reported by studies worldwide has var-
ied from 3.7% in New York to 11% in UK hospitals and 16.6% in Australian hospitals.
In Canada, two recent papers give rates of 5% and 7.5% (Baker et al., 2004; Forster
et al., 2004) and the report Health Care in Canada 2004 states that about 5.2 million
Canadians (representing a quarter of the population) have experienced a preventable
adverse event either themselves or in a family member (Gagnon, 2004). The wide varia-
tion in reported adverse event rates is partly due to differences in study methods and
patient selection. Moreover, there is no agreement on what constitutes “preventability.”
Only a few studies looked at preventability of adverse events as part of their original
design (Forster et al. 2004; Vincent et al., 2001; Wilson et al., 1995). But there is now
a consensus that, in terms of patient safety, many health systems perform below their
potential best.
8.3.3 Australia
In 2002, a committee of the Australian Council for Safety and Quality in Health Care
offered an approach to achieving open and honest communication with patients after
an adverse event (Australian Council for Safety and Quality in Health Care, 2002),
addressing the interests of consumers, healthcare professionals, managers, and orga-
nizations. Like the policy proposed by the US Joint Commission, the draft standard is
flexible in allowing development of local policies and procedures. The unique aspect
of the Australian draft standard is the integration of disclosure into a risk manage-
ment analysis and investigation of the critical event. The level of investigation depends
on grading of the event according to the extent of injury and the likelihood of its
occurrence.
A decision by the New South Wales Supreme Court decreed that a legal duty to
disclose medical errors occurs when the knowledge is relevant to possible changes to
the patient’s care and medical outcome. As Justice J. Bryson remarked: “communication
with the patient, both before and after treatment, of the diagnosis, advice about what
treatment is proposed, and of a report of what treatment has taken place are all integral
and essential parts of treatment. They are essential where the patient is conscious and
has the capacity to participate in them, because of the nature of the patient as a person
with a right to give or withhold consent to an intervention in his body by another
person” (New South Wales Supreme Court, 1994).
The Australian tort system was also reformed to afford medical professionals some
protection when they make apologies to patients who have experienced an adverse
event. The Civil Liability Act of New South Wales (Civil Liability Act [NSW], 2002),
which is similar to those in many other provinces, reads as following:
90 8 International laws and guidelines addressing error and disclosure
A duty to disclose, however, does not exist in Australia, nor does a legal duty to
express sympathy or regret (Madden and Cockburn, 2005).
8.3.5 Canada
In 2002, the Royal College of Physicians and Surgeons of Canada called for health
care systems to promote disclosure on safety issues to all partners including patients
(National Steering Committee on Patient Safety, 2002), but no uniform Canadian guide-
lines on the subject are yet in place (see fFig. 8.1). Reviewing nationwide practices on
adverse event disclosure, we found that just a few licensing bodies had ratified policies
for disclosure and discussion of negative outcomes during patient care. The College of
Physicians and Surgeons of Saskatchewan requires physicians to disclose any adverse
events and errors to the patient or his or her representative as soon as possible during
care and provides 10 guidelines for purposeful disclosure. The College of Physicians
8.3 International progress and initiatives 91
Y.T.
N.W.T.
NU
B.C. NF
LD
AB &L
SK AB
MB
QC
ON
P.E.I.
N.S.
N.B.
and Surgeons of Manitoba requires physicians to avoid all speculations and to state
plain facts as known at the time. In 2003, after lengthy deliberation, the College of
Physicians and Surgeons of Ontario approved a policy that made disclosure of harm
to patients a standard of practice (Borsellino, 2003), even in circumstances where such
disclosure might result in a complaint or a malpractice insurance claim. A special as-
pect of the Ontario College policy is the guideline for medical trainees (i.e., students or
residents), who are advised to report adverse events either to their supervisor or to the
“most responsible physician.”
The College of Physicians and Surgeons of Quebec has no distinct policy on adverse
event disclosure to patients but synthesizes the concept of disclosure in its code of eth-
ics. The College of Physician and Surgeons of Newfoundland and Labrador enacted a
disclosure policy in 2006, emphasizing that a disclosure is not an admission of fault or
liability. The College in Nova Scotia developed a disclosure policy two years later and
also initiated a quality assurance program to complement the policy (see fTab. 8.1).
Disclosure may also be viewed as part of the informed-consent process. While some
Canadian provinces have regulatory initiatives to address disclosure and though many
of these are similar in content, they remain isolated because of their nonmandatory
nature and an absence of federal and provincial laws. It is suggested that a uniform na-
tional policy centered on addressing errors in a nonpunitive manner and respecting the
patient’s rights to an honest disclosure be implemented (Kalra et al., 2007). In Canada,
there is nothing in the nature of the United States Joint Commission initiative that makes
92 8 International laws and guidelines addressing error and disclosure
Tab. 8.1: Canadian provincial College of Physicians and Surgeons policies on disclosure.
8.4 Conclusion
Clearly, a wide range of responses to patient safety concerns exists in Western nations’
health care systems. Canada has much to learn from such progressive initiatives in
many countries. Strong examples are present that attempt to balance physician and
health care worker responsibility to the patient against unnecessary risk of legal action.
Development of international standards may be one way to encourage and promote
the continued development of policies in this area. Yet, variation within nations, such
as that among the provinces in Canada, epitomizes why regulations may be difficult to
enact for this purpose. There is a need for cross-country comparative studies to examine
differences between national policies and outcomes to determine best approaches.
References
American Medical Association Council on Ethical and Judicial Affairs (AMACEJA). Code of
Medical Ethics: Current Opinions. Chicago: American Medical Association; 2000.
Australian Council for Safety and Quality in Health Care. Draft Open Disclosure Standard.
Standards Australia, XX 1234–2002, Draft v5.2.
Baker GR, Norton PG, Flintoft V, et al. The Canadian Adverse Events Study: the incidence of
adverse events among hospital patients in Canada. CMAJ 2004;170:1678–1686.
Borsellino M. Disclosure of harm to be standard of practice. Medical Post 2003;39 (11).
Civil Liability Act 2002 (NSW), §69.
Davis P, Lay-Yee R, Fitzjohn J, Hider P, Briant R, Schug S. Compensation for medical injury in
New Zealand: does “no-fault” increase the level of claims making and reduce social and
clinical selectivity? J Health Polit Policy Law 2002;27(5):833–854.
References 93
The autopsy is revered as the gold standard in medical quality assessment, mean-
ing that there exists no better methodology for determining the cause of a patient’s
death. A physician’s original diagnosis of the patient’s disease can therefore be de-
termined to be correct if it is concordant with the postmortem diagnosis; if discor-
dant, then the physician made an error. This chapter addresses the ways in which
the autopsy contributes to modern medical knowledge and its unique contribution
to patient safety.
9.1 Introduction
Autopsy examinations have been suggested as one of the methods in quality assessment
(Anderson, 1984). Autopsy diagnoses have traditionally been used as the gold standard
for determining the cause of death and can provide an invaluable retrospective look into
the ability of the medical staff to accurately diagnose the actual cause of death. Despite
all of the technical advances in medical and diagnostic modalities, research concern-
ing medical errors in hospitalized populations has consistently revealed high rates of
misdiagnoses. The medical literature is replete with studies that suggest discrepancies
between antemortem clinical diagnoses and postmortem autopsy diagnoses. Concor-
dance rates of clinical and autopsy diagnoses have been reported to vary from 32% to
77% (see fTab. 9.1). Utilization of autopsies as part of an error reduction strategy may
be a common denominator in improving quality, one that may be used more frequently
if we put in place measures to encourage its undertaking.
that misdiagnoses or missed diagnoses occurred in 26% of the cases analyzed (Balaguer
et al., 1998).
A separate study at another Spanish hospital revealed that more than half of the risk
of death could be traced back to clinical care that resulted in adverse events (Garcia-
Martin et al., 1997). Although autopsy rates are declining, autopsies remain an important
method by which medical error can be detected. Despite the use of modern medical
technology to help medical professionals more accurately diagnose patients, includ-
ing the use MRIs and CT scans, the frequency of discordance has remained largely
unchanged for more than 40 years (Lundberg, 1998).
We have suggested a model in which the health care sector either incentivizes the
legally responsible party to allow an autopsy or promotes the benefits of autopsies
and postmortem findings (Kalra et al., 2010). After all, autopsies contribute to clinical
knowledge by extending the scientific and medical understanding of the human body
and disease processes. They also aid medical education by allowing medical students
and residents to train with greater efficiency by having access to actual human tis-
sues and organs. Lesar et al. (1990) found that there were more prescribing errors –
typically a result of inaccurate diagnosis – among first-year postgraduate residents than
among other clinicians. Wilson et al. (1998) discovered that the likelihood of error
increased in a pediatric intensive care unit when a new doctor joined the rotation. If
the quality of medical education can improve, it is likely that the number of inaccurate
diagnoses and other medical errors made by new clinicians will in turn decrease.
if detected, would not have led to a change in management. Classes III and IV refer to
missed minor diagnoses involving diseases that were not directly related to the death
of the patient. Gibson et al. (2004) added a fifth discordance category in their study of
postmortem discrepancies: overdiagnosis, that is, instances in which patients are diag-
nosed with additional diseases, either principal or secondary, whose treatment would
not have altered the outcome in anyway.
It is important to note that misdiagnoses do not always represent errors. They may
instead reflect acceptable limits of antemortem diagnoses or atypical clinical presenta-
tions. When such atypical presentations occur, however, the autopsy becomes extremely
important, as it can reveal the clinical course of rare or unrecorded diseases.
1996). Pediatric technology has also identified the delineation of Reyes syndrome and
amoebic meningoencephalitis (Khong and Arbuckle, 2002).
Third, autopsies allow for performance checks, both within the hospital itself and
within the health care system overall. If the individual physician in charge of the diag-
nosis greatly influences misdiagnosis rates at a given hospital, then it’s an indication that
corrective action must be taken. A study undertaken by Battle et al. (1987), however,
indicated that discordance rates are not dependent upon the individual physician. Yet,
higher-than-average discrepancy rates in individual hospitals within a health care region
are also a cause for concern.
Fourth, autopsies can be used to instruct medical professionals. They can be part
of a process to teach anatomy, macroscopic pathology, skills in clinicopathological
correlation, the fallibility of medicine, medical bioethics, the process of dying and han-
dling the dead, invasive clinical procedures, medical law, and the importance of health
and safety at work (Burton, 2003). The autopsy satisfies these functions because of its
visual nature and because of the systems-based, problem-oriented approach adopted
by pathologists during the autopsy procedure. Better training produces better doctors,
which improves patient safety across the board.
which only blind or targeted percutaneous needle biopsies are taken and only examina-
tions that can be done by endoscope or laparoscope are performed. An autopsy that
does not penetrate the body in any way, known as a virtopsy, can be performed via the
use of MRI and multislice CT scans. In such a procedure, CT images provide data on
the general pathology of the body, which can generate detailed information concerning
trauma injuries. MRIs are focused on specific areas of the body, providing data related
to tissue, muscles, and organs. Metabolite counts in the brain are measured using MRI
spectroscopy to determine time of death.
Despite the general reluctance to perform autopsies, recent reports indicate that up
to 46% of relatives may agree to an autopsy if approached sensitively by the physi-
cian (Osborn and Thompson, 2001). In a Norwegian hospital, only 9% of deceased
patients’ next of kin who were approached refused permission for the performance of
an autopsy (Ebbesen et al., 2001). Lugli and colleagues (1999), in Switzerland, found
that simple measures such as including attending physicians in autopsy discussions
with relatives and providing physicians with communication training on how to deal
with relatives led to an increase in autopsy rates from 16% in 1997 to 36% in 1998.
Another study, however, reported that 80% of families refused permission for autopsies
(Combes et al., 2004). It is interesting to note that studies seem to suggest that hospitals
in which physicians are more hesitant to seek consent are the ones with the lowest
autopsy rates (Burton and Underwood 2003; Burton et al., 2004).
Physicians’ hesitancy to seek autopsies may stem from several factors. Doctors may
face legal repercussions if clinical diagnostic errors are exposed during the autopsy
(Hasson and Gross, 1974). Moreover, many health care professionals believe, despite
the high frequency of missed diagnoses, that advances in diagnostic procedures, such
as MRIs and CT scans, have reduced the value of autopsies. Chariot et al. (2000) cited
perceived long response time, low satisfaction with the content of the autopsy report,
and difficulties in obtaining consent from relatives as the three major barriers facing
clinicians who request autopsies.
9.6 Conclusion
Shojania et al. (2003) suggested that, though the possibility of clinically discordant
diagnoses at autopsy has decreased over time, it still remains high enough to warrant
the ongoing use of autopsies. The gold standard of diagnosis should not be allowed to
disappear. Medical and accreditation institutions that drop minimum autopsy rates from
regulatory guidelines are depriving the health care system not only of a valuable quality
assessment method but also of information that can help lead to a healthier community.
Despite this, in 1986, Medicare removed the minimum autopsy requirement for par-
ticipating hospitals (O’Leary, 1996); even before that, the Board of Commissioners of
the Joint Commission on Accreditation of Hospitals had discontinued its mandate that
20% of hospital deaths undergo autopsies (Reichart and Kelly, 1985). These systemwide
changes have reduced the likelihood that autopsies will be conducted.
References
Anderson RE. The autopsy as an instrument of quality assessment. Classification of premortem
and postmortem diagnostic discrepancies. Arch Pathol Lab Med 1984;108:490–493.
100 9 The value of autopsy in detecting medical error and improving quality
Balaguer MJV, Gabriel BF, Braso AJV, Nunez SC, Catala BT, and Labios GM. El papel de la
autopsia clinica en el control de calidad de los diagnosticos clinicos en una unidad de
urgencias. Anales de Medicina Interna 1998;15:179–182. [in Spanish]
Battle RM, Pathak D, Humble CG, et al. Factors influencing discrepancies between premor-
tem and postmortem diagnoses. JAMA 1987;258:339–344.
Bove KE, Iery C. The role of autopsy in medical malpractice cases. Arch Pathol Lab Med
2002;126:1023–1031.
Burton JL. The autopsy in modern undergraduate medical education: a qualitative study of
uses and curriculum considerations. Med Educ 2003;37:1073–1081.
Burton JL, Clinical, educational, and epidemiological value of autopsy. Lancet 2007;369:
1471–1480.
Burton JL, Underwood JC. Necropsy practice after the “organ retention scandal”: requests,
performance, and tissue retention. J Clin Pathol 2003;56:537–541.
Burton EC, Phillips RS, Covinsky KE, et al. The relation of autopsy rate to physician’s beliefs
and recommendations regarding autopsy. Am J Med 2004;117:255–261.
Chariot P, Witt K, Patot V, et al. Declining autopsy rate in a French hospital. Arch Pathol Lab
Med 2000;124:739–735.
Chin JE, editor. Creutzfeld-Jakob disease. In: Control of Communicable Diseases Manual.
Washington, DC: American Public Health Association; 2000:183–186.
Combes A, Mokhtari M, Couvelard A, et al. Clinical and autopsy diagnoses in the intensive
care unit: a prospective study. Arch Intern Med 2004;164:389–392.
Coradazzi AL, Morganti AL, Montenegro MR. Discrepancies between clinical diagnoses and
autopsy findings. Braz J Med Biol Res 2003;36:385–391.
Doyle YG, Harrison M, O’Malley F. A study of selected death certificates from three Dublin
teaching hospitals. J Public Health Med 1990;12:118–123.
Ebbesen J, Buajordet I, Erikssen J, et al. Drug-related deaths in a department of internal medi-
cine. Arch Intern Med 2001;161:2317–2323.
Ermenc B. Comparison of the clinical and postmortem diagnoses of the causes of death.
Forensic Sci Int 2000;114:117–119.
Esiri MM. Why do research on human brains? In: Gunning J, Holm S, eds. Ethics Law and
Society. Aldershot: Ashgate; 2005:33–39.
Fruhbeck G. Death of the teaching autopsy: advances in technology have not reduced the
value of the autopsy. BMJ 2004;328:165–166.
Garcia-Martin M, Lardelli-Claret P, Bueno-Cavanillas A, Luna-del-Castillo JD, Espigares-
García M, Gálvez-Vargas R. Proportion of hospital deaths associated with adverse events.
Jour of Clin Epidem 1997;50:1319–1326.
Gibson TN, Shirley SE, Escoffery CT, Reid M. Disrecpancies between clinical and postmortem
diagnoses in Jamaica: a study from University Hospital of the West Indies. J Clin Pathol
2004;57:980–985.
Goldman L, Sayson R, Robbins S, Cohn LH, Bettmann M, Weisberg M. The value of the
autopsy in three medical eras. N Engl J Med 1983;308:1000–1005.
Grade MH, Zucoloto S, Kajiwara J, Fernandes M, Couto LG, Garcia SB. Trends of accuracy
of clinical diagnoses of the basic cause of death in a university hospital. J Clin Pathol
2004;57:369–373.
Gut AL, Ferreira AL, Montenegro MR. Autopsy: quality assurance in the ICU. Intensive Care
Med 1999;25:360–363.
Hasson J, Gross H. The autopsy and quality assessment of medical care. Am J Med.
1974;56:137–140.
Kalra J, Entwistle L, Suryavanshi S, Chadha R. Quality assessment using concordance and
discordance rates in medical findings. Clin Govern 2010;15:128–133.
References 101
Khong TY, Arbuckle SM. Perinatal pathology in Australia after Alder Hey. J Paediatr Child
Health, 2002; 38:409–411.
Lesar TS, Briceland LL, Delcoure K, Parmalee JC, Masta-Gornic V, Pohl H. Medication pre-
scribing errors in a teaching hospital. JAMA 1990; 263:2329–2334.
Lugli A. Anabitarte M, beer JH, Effect of simple interventions on necropsy rate when active
informed consent is required. Lancet 1999;354:1391.
Lundberg GD. Low-tech autopsies in an era of high-tech medicine: continued value for qual-
ity assurance and patient safety. JAMA 1998;280:1273–1274.
Medicare Payment Advisory Committee. Report to Congress: Selected Medicare Issues. Chap-
ter 3: addressing health care errors under Medicare. Washington, DC:1999:47–50.
Mort TC, Yeston NS. The relationship of premortem diagnoses and postmortem findings in a
surgical intensive care unit. Crit Care Med 1999;27:299–303.
O’Connor AE, Parry JT, Richardson DB, Jain S, Herdson PB. A comparison of the antemortem
clinical diagnosis and autopsy findings for patients who die in the emergency department.
Acad Emerg Med 2002;9:957–959.
O’Leary DS. Relating autopsy requirements to the contemporary accreditation process. Arch
Pathol Lab Med 1996;129:763–766.
Osborn M, Thompson EM. What have we learnt from the Alder Hey affair? Asking for consent
would halt decline in voluntary necropsies. BMJ 2001;322:1542–1543.
Perkins GD, McAuley DF, Davies S, Gao F. Discrepancies between clinical and postmortem
diagnoses in critically ill patients: an observational study. Crit Care 2003;7:129–132.
Reichart CM, Kelly VI. Prognosis for the autopsy. Health Aff 1985;4:82–92.
Roosen J, Franse E, Wilmer A. Comparison of premortem clinical diagnoses in critically ill
patients and subsequent autopsy findings. Mayo Clin Proc 2000;75:262–267.
Rosenbaum GE, Burns J, Johnson J, Mitchell C, Robinson M, Truog RD. Autopsy consent
practice at US teaching hospitals: result of a national survey. Arch Intern Med 2000;160:
374–380.
Shojania KG, Burton EC, McDonald KM, Goldman L. Changes in rates of autopsy-detected
diagnostic errors over time: a systematic review. JAMA 2003;289:2849–2456.
Tse GM, Lee JC. A 12-month review of autopsies performed at a university-affiliated teaching
hospital in Hong Kong. Hong Kong Med J 2000;6:190–194.
Venters GA. New variant of Creutzfeldt-Jakob disease – the epidemic that never was. BMJ
2002;323:858–861.
Will RG, Ironside JW, Zeidler M, et al. A new variant of Creutzfeldt-Jakob disease in the UK.
Lancet 1996;347:921–925.
Wilson DG, McArtney RG, Newcombe RG, et al. Medication errors in pediatric practice:
insights from a continuous quality improvement approach. Eur J Ped 1998;157:769–774.
10 Total quality management, six-sigma, and health care
10.1 Introduction
The quality of health care has been a primary concern of many governments world-
wide. The IOM conducted a National Roundtable on Healthcare Quality as early as
1996 to deliberate quality-of-health and health care issues in the United States. The
roundtable comprised 20 representatives of the private and public sectors, medical
and nursing practitioners, academicians, business professionals, patient advocates,
media persons, and health administrators. The role of the National Health Service
(NHS) in the United Kingdom in improving the quality of patient care was outlined
clearly in the government white paper The NHS: Modern and Dependent (Crook,
2002). In this paper, the concept of clinical governance has been vividly set out.
Clinical governance is defined as a “system through which [NHS] organizations are
accountable for continuously improving the quality of their services and safeguarding
high standards of care by creating an environment in which excellence in clinical care
will flourish” (Scally et al., 1998). Chassin and Galvin (1998), making a statement on
the IOM National Round Table, wrote: “Problems in healthcare quality are serious and
extensive; they occur in all delivery systems and financing mechanisms.” The authors
further noted that “Americans bear a great burden of harm because of these problems,
a burden that is measured in lost lives, reduced functioning and wasted resources,”
and called for urgent action.
104 10 Total quality management, six-sigma, and health care
Tab. 10.1: Levels of sigma performance and corresponding defects per million.
6 3.4
5 233
4 6,210
3 66,807
2 308,537
1 690,000
±2 standard deviations (SD) of the mean. In a Gaussian distribution, this would result
in only a 4.5% defect rate, but considering the potential for health care usage, this
would translate into an appalling 45,400 DPM opportunities. These figures would be of
little solace to an already ill patient. Clinical diagnostic laboratories are content if their
results enclose ±2SD or ±3SD limits. In other words, they find defect rates of 45,400
DPM opportunities and 2,700 DPM opportunities (see fTab. 10.2) to be acceptable
performance (Harry et al., 2000).
Many claim that little is gained from improving a process performance beyond the
five-sigma, or 233 DPM, level. It is felt that six-sigma method applications can actu-
ally tolerate small shifts in the process mean without increasing the defect rate that
significantly. With a six-sigma process, we are assured that the process is still producing
results within the desired specifications and with low defect rates. The six-sigma process
provides an added advantage in that it is easily monitored with any QC procedure,
unlike a process at five-sigma or lower sigma levels, where the choice of QC procedure
is more important.
In any process, variation is inherent. It is variation in the process that creates the
opportunities for errors to happen, and therefore it should be seen as the “enemy” (Lan-
ham, 2003). Walter Shewhart (1980) described two types of variation: common cause
and special cause. Common-cause variations are intrinsic to a process and require ac-
tion on the process itself to decrease the variation, whereas special-cause variation
occurs due to factors extrinsic to the process, which require identification and action.
The key lies in minimizing these variations and producing a stable process. These stable
processes exhibit common-cause variations, which are best reduced by correcting the
underlying process (Mohammed et al, 2001). It is the variation in a process that has to
be minimized and controlled to achieve high-quality results. Reduction in variation is
also a core concern in clinical governance. Shewhart (1980) also devised control charts,
a graphical methodology for differentiating between the two types of variation. The
defects occurring through common-cause variation fall within the upper and lower lines
of the graph (control limits), and special-cause variation are represented by the data
points that fall outside the control limits. He suggested using limits of three-sigma from
the mean. If beyond these points, it was suggested that the process required correction.
If one were to apply the three-sigma limit for accepting a process, it would translate into
66,807 DPM opportunities.
106 10 Total quality management, six-sigma, and health care
>2SD 45,400
>3SD 2,700
>4SD 63
>5SD 0.6
>6SD 0.002
>7SD 3 10-6
their regular jobs and accompanying normal duties. The Green Belts do, however, play
a crucial role in bringing the concept and analytical tools of six-sigma training directly
to everyday activities of the workplace and the process. It is principally for this reason
that organizations desiring greater success in the six-sigma projects think about training
a large segment of their workforce to be Green Belts. Six-sigma implementation involves
every level of an organizational hierarchy, with the top level providing the leadership
and the bottom level driving the whole process.
The performance improvement methodology or model used in six-sigma is most often
“DMAIC” (Define, Measure, Analyze, Improve, Control). Each letter stands for one of
the different stages involved in the implementation of a six-sigma strategy. Though other
methodologies for implementing six-sigma strategies also exist, they are practiced by
only a few companies and hence are discussed briefly later. When six-sigma is referred
to, it is invariably the DMAIC methodology that is being mentioned (see fFig. 10.1).
This is a stepwise graded approach for enhancing the quality of a process to produce the
desired goals. The various stages are:
– Define – the problem in terms of the process, goals the project intends to achieve,
custom deliverables, and any other components essential to quality.
– Measure – the process quantitatively, which is best achieved by data gathering,
which helps in assessing the current performance levels and comparison with the
best practices.
– Analyze – the process by using a root-cause analysis approach to determine where
the problem is originating and which problem is most responsible for the deteriorating
quality of the process.
– Improve – the process by eliminating the defects through identification of causes.
– Control – the process so that the improvements are sustained and defects do not
re-emerge later.
SUSTAIN THE
IMPROVEMENTS
It should be re-emphasized here that using the DMAIC approach to the six-sigma
quality improvement strategies aids in achieving a very low number of DPM opportuni-
ties, ideally equal to 3.4 DPM opportunities, which betters 99% performance levels
by being 99.9997% perfect all the time. Such a high level of perfection eventually
means higher efficiency, leading to reduction in costs, efforts, and time expended and
to improved overall client satisfaction.
Apart from the DMAIC methodology, which is essentially used for a preexisting pro-
cess that is defective, the DMADV (Define, Measure, Analyze, Design, Verify) method-
ology is used when a new process is being developed or when the DMAIC process has
failed to correct a defect. In the DMADV approach, the goals and methods are similar
to those in the DMAIC approach – that is, to reduce defects to levels below 3.4 DPM
opportunities and data driven, respectively. The difference lies only in the last two steps.
In DMADV, they are Design (create a detailed design of the process to meet customer
needs) and Verify (check the design performance and its ability in meeting customer
needs).
There are other less commonly used methodologies, such as DMADOV (Define,
Measure, Analyze, Design, Optimize, Verify), a variant of DMADV; DCCDI (Define,
Customer Concept, Design, Implement); IDOV (Identify, Design, Optimize, Validate);
and DMEDI (Define, Measure, Explore, Develop, Implement).
Data
Collection Current and Setting of High
Novel Quality Professional Education
and Total Quality
Improvement Quality +
Evidence Training Improvement
Based Techniques Standards
Practices
the preanalytical phase. As reviewed earlier, a majority of errors occur in this stage.
Because a laboratory analyzes what is delivered to it, preanalytical factors have a direct
bearing on the analytical and postanalytical stages of the testing process. This problem
is not restricted to the centralized laboratory model but applies also to point-of-care
testing (POCT). In a well-reviewed paper on the various aspects of POCT, St-Louis has
stressed the demands that the POCT presents in terms of quality and the importance of
QA to address all phases of a test performance (St-Louis, 2000).
Six-sigma can be applied widely in all three stages of a clinical diagnostic laboratory
testing process. In the preanalytical stage, it can be used to enhance quality of information
on requisitions, patient identification, and specimen collection and transportation. In the
analytical stage, it can find applications in reducing laboratory testing errors and avoiding
misinterpretation, misreading, and misjudging of the results. In the postanalytical stage, it
can be used successfully to reduce the turnaround time for obtaining the results.
Previous research in this area has failed to provide us with clear directions for im-
proving quality in our laboratories. It has, rather inadvertently, focused on descriptive
statistics and fallen short of exposing the real underlying issues of quality failures. These
failures and barriers offer abundant opportunities for further research and development
of processes that are efficient and of high quality. In summary, quality in clinical labora-
tories is driven by application of data-driven approaches and evidence-based practices
(see fFig. 10.2). This approach helps in setting up professional, high-quality standards
and, coupled with education and training, helps transform a laboratory culture into a
quality-conscious setting.
10.5 Conclusion
It is imperative for the health care sector in general and for clinical diagnostic labora-
tories in particular to promote and develop a culture of safety with the aid of modern
quality management techniques and tools. Today’s quality assurance and improvement
activities in clinical laboratories are governed by the Clinical Laboratory Improvement
Amendments of 1988 (CLIA ’88) and Joint Commission on Accreditation of Healthcare
Organizations (JCAHO) guidelines. However, it must be mentioned that the influence of
CLIA ’88 and JCAHO guidelines is largely confined to clinical laboratories in the United
States and may not apply to clinical laboratories in Canada and elsewhere. The criteria
set out by CLIA ’88 and JCAHO, though highly effective, are not very demanding for
analytical performance and are based on two-sigma to three-sigma process goals only.
110 10 Total quality management, six-sigma, and health care
The goals of six-sigma quality are impressive and set demanding standards that appear
to be more compatible with patient safety. In addition, the present philosophy of quality
assurance being defined as “find a problem, fix a problem” is not feasible; significant
improvements in laboratory performance call for more systematic approaches. The six-
sigma concept provides an opportunity for major improvements and helps achieve the
vision of ultimate quality to deliver error-free and timely clinical diagnostic laboratory
services.
References
Chassin MR, Galvin RW. The urgent need to improve health care quality. Institute of Medicine
National Roundtable on Health Care Quality. JAMA 1998;280:1000–1005.
Crook M. Clinical governance and pathology. J Clin Pathol 2002;55:177–179.
Eichhorn JH. Prevention of intraoperative anesthesia accidents and related severe injury
through safety monitoring. Anesthesiology 1989;70:572–527.
Harry MJ. “Six sigma: a breakthrough strategy for profitability.” Quality Progress
1998;31:60–64.
Harry M, Schroeder R. The Breakthrough Management Strategy Revolutionizing the World’s
Top Corporations. New York: Doubleday; 2000.
Johnstone PA, Hendrickson JA, Dernbach AJ, et al. Ancillary services in the health care indus-
try: is Six Sigma reasonable? Qual Manag Health Care 2003;12:53–63.
Lanham B. Beth Lanham on Six Sigma in healthcare. Interview by Luc R. Pelletier. J Healthc
Qual 2003;25:26–27, 37.
Mohammed MA, Cheng KK, Rouse A, Marshall T. Bristol, Shipman, and clinical governance:
Shewhart’s forgotten lessons. Lancet 2001;357:463–467.
Scally G, Donaldson LJ. The NHS’s 50 anniversary. Clinical governance and the drive for
quality improvement in the new NHS in England. BMJ 1998;317:61–65.
Shewhart WA. Economic control of quality of manufactured product. New York: D Van Nos-
trand; 1931. (Reprinted by ASQC Quality Press, 1980).
St-Louis P. Status of point-of-care testing: promise, realities, and possibilities. Clin Biochem
2000;33:427–440.
Index
accidents, 10, 20, 26, 57, 62, 63, 82, critical thinking, 27
97, 110 culture of safety, 8, 11, 16, 20, 45, 51,
accountable, 2, 23, 103 53, 61, 62, 109
accreditation, 29, 36, 52, 59, 69, 70, 73, cut-off values, 70
88, 92, 99, 101
active error, 24 defects, 23, 66, 104, 105, 106, 107, 108
adverse events, 31, 1, 2, 4, 15, 5, 6, 7, definitions of medical error, 12, 13
8, 9, 10, 51, 31, 38, 41, 44, 48, 49, diagnostic, 5t, 32, 67t, 95, 109
66, 56, 57, 59, 62, 63, 77, 87, 77, disclosure, 63, 77, 78, 83, 85, 91, 92, 93
78, 82, 84, 85, 87, 96, 88, 90, 92,
93, 100 ED. See emergency departments
American Accreditation Council for education, 14, 17, 18, 37, 38, 52, 56,
Graduate Medical Education 60, 62, 63, 69, 73, 83, 96, 100,
(ACGME), 18, 19 109
American Medical Association educational programs, 36, 60
(AMA), 1 emergency departments (ED), 3, 32, 38,
analytical errors, 33, 68, 69 39, 41, 85
attitude, 25, 53, 54, 57, 108 emergency medicine, 38
autopsies, 95, 96, 97, 98, 99, 100, 101 epidemiology, 4
error rates, 4, 23, 32, 33, 35, 38, 43, 51,
bias, 7, 42, 71 66, 68, 104
bioethics, 23, 98 external quality assessment, 74
blame, 18, 21, 23, 25, 53, 57, 65, 78,
79, 83, 84, 87 fatigue, 18, 19, 25, 26, 27, 37, 69
blunder, 31, 32, 33, 46, 66, 74
harm, 1, 2, 8, 9, 12, 13, 16, 17, 43, 57,
Canadian Patient Safety Institute, 39, 46, 78, 81, 90, 91, 92
60, 62, 73 Harvard Medical Practice (HMP), 4, 31
clinical chemistry, 53, 58, 75 human error, 25, 28, 43, 44, 69, 73, 79
clinical diagnostic laboratories, 31, 32, human performance, 27
65, 66, 68, 69, 108, 109
cognitive decision-making errors, 28 industrial-grade quality performance,
cognitive influence, 1, 27 103
communication, 7, 13, 14, 34, 38, 40, industrial sectors, 103, 104
43, 48, 56, 61, 68, 73, 80, 81, 84, Institute of Medicine (IOM), 1, 2, 4, 7, 8,
88, 89, 99 11, 13, 15, 17, 19, 20, 21, 23, 31,
compensation, 77, 82, 83, 84, 89 32, 38, 47, 49, 51, 52, 57, 59, 60,
consequences, 35 63, 83, 85, 87, 88, 93, 103, 110
critical events, 43, 57 insurance, 9, 84, 89, 91
112 Index
intensive care units, 31, 38, 41, 45, 47 pressures, 16, 18, 29, 38
international laws, 87 proactive, 16
IOM. See Institute of Medicine professionalism, 56, 62, 64
training, 18, 19, 24, 26, 27, 29, 38, 42, voluntary reporting, 37, 44, 57, 58
43, 46, 48, 55, 56, 63, 70, 83, 98,
99, 106, 109 work demand, 17
turnaround time (TAT), 40, 70, 109 workload, 19, 24, 26, 37, 42, 43, 49, 57, 70