Article Text


Clinical decision-making: coping with uncertainty
  1. A F West1,
  2. R R West2
  1. 1Child and Adolescent Mental Health Service, Wokingham Hospital, Wokingham, Berkshire RG41 2RE, UK
  2. 2Epidemiology, University of Wales College of Medicine, Cardiff, UK
  1. Correspondence to:
 Dr Andrew West

Statistics from

Despite increasing medical knowledge uncertainty will always remain


Reliable research and audit information are essential to increasing medical knowledge and improving health service delivery. However there are limits to available information in terms of quality, reliability, and applicability. Furthermore, however much information is gathered, there will always be a degree of uncertainty at the point of making clinical decisions with individual patients. Unrealistic lay and professional expectations of the efficacy of information and that certainty is achievable, may be altering the traditional clinician-patient relationship. One therapeutic role of a clinician is containing the anxieties aroused in the context of uncertainty, and this role may be becoming more difficult. Reliance on protocols and fear of reprimand may lead to clinicians, in some areas of medical care, abandoning their patients at a time of need.


Many acknowledge that there has been an information explosion in the health services over the last few decades. Vastly increased volumes of information are gathered in pursuit of knowledge and in the name of audit. The former provides the backbone of “evidence based medicine” and the latter facilitates clinical governance.1–3 Almost certainly, more information is being gathered than will ever be put to use. The authors of this paper—one a clinician, the other an epidemiologist—view these developments from different perspectives and share a concern. They recognise the extent to which properly collected and handled research information and properly collected and handled audit information could and should improve health service delivery,4 yet are also aware of the limits to the information, both in its comprehensiveness and in its usefulness. What is more, they join to describe how the current emphasis on apparently factual information lets down both doctor and patient because it denies the importance of a crucial aspect of the relationship, namely the management of uncertainty. However much information or “evidence” there is to hand (and there will now always be more than can be satisfactorily held in mind) a decision still has to be made and, at the point of making each decision, there will always be some uncertainty.


Few would question that a rationale exists for the “evidence based medicine” movement. Reliable information is essential to both scientific advancement and process management. This need for information has led to such quantities of evidence that clinicians need assistance in choosing which evidence should influence their practice.5


There are limits to the information available through research and review.6 While some areas may be over-researched, for example the surgeon general of the USA reported thousands of papers into the health effects of smoking,7 there are other areas, for example in the psychotherapies, where randomised trials and meta-analytic evidence are hard to come by and some would argue may even be inappropriate. There are also problems concerning the quality of information. Doctors are now trained to the point where they can spot a paper's more obvious shortcomings such as sampling errors, data loss, and incorrect conclusions. This is necessary because, despite a general level of awareness, important errors still slip through peer review and appear in print. Perhaps the main function of teaching medical students “critical appraisal”8 is not to enable them to judge a paper, but to give them a healthy scepticism and suspicion of the published word. What, then, does the non-statistician trust as evidence? For some the meta-analysis was, for a time, perceived as the head of the “hierarchy of evidence” but it is now more widely recognised that meta-analysis is itself a sampling procedure and is subject to random effects9–11 and we are now seeing meta-analyses of meta-analyses.12 Even good systematic reviews present the clinician with questions of external validity or applicability of research findings in clinical practice.13 The certainty sought by research forces researchers to simplify the measurable situation to the point where it is often unrepresentative of the typical clinical scenario: it will almost always be possible to show differences between the research samples and clinical populations, never mind the individual patient.

“Complexity science”14 may go some way towards explaining why complex systems, like people, do not behave in predictable ways, yet those authors appear to remain seduced by the search for certainty, for they say, “Ultimately the only way to know exactly what a complex system will do is to observe it”. Such reductio ad absurdum illustrates a crucial point. Prediction of the future on past observations is probabilistic and is not the same as knowing exactly what will happen. Although all previous swans were white, we cannot be certain that the next will not be black. Prediction is particularly difficult when n = 1.


The Yerkes-Dodson law, first propounded in 1908, describes how arousal and task-difficulty impact on the time taken by individuals to perform at a given task; too little arousal and people cannot be bothered; too much and they becomes distracted.15 Perhaps there is a critical amount of information that a clinician can hold and which enables optimal functioning: too little and they feel uninformed; too much and they are confused with information overload. Erring in either direction could make decision-making harder. Certainly the curvilinear relationship between performance and levels of reward or punishment imply that, if the stakes are too high, a clinician's ability to perform at a discriminant task, such as decision-making, will fall off.


The existence of an information mountain provides a myth of certainty for the patient, the public and perhaps for health care policy-makers. Certainty is an illusion. In most professional spheres the expert is more critical of the available evidence, than is the lay person. Hopefully, clinicians are taught to be more critical than patients but they are faced with a dilemma. They appear to be very well informed yet are acutely aware of the fallibility of the information that is expected to guide their practice. How do clinicians respond to this dilemma? Their day consists of a succession of points at which decisions have to be made on the basis of incomplete and inadequate information and in a context in which risk, never mind error, is outlawed.

Quick and easy solutions exist, which can be divided broadly into the clean and the dirty: either the responsibility is passed on to protocols and guidelines, or on to the patients themselves. Guidelines still leave the clinician the onus of making a decision and so really come into the category of “evidence”. Protocols absolve the clinician who follows the protocol correctly, but facilitate criticism of the clinician who does not. They can restrict research, ironically limiting the acquisition of new evidence. They appear to restrict what action is ethical but actually remove ethical considerations from the immediate clinical decision. Finally, they threaten to depersonalise the clinical relationship to the point where the patient is effectively consulting an automaton instead of a human being.

The dirty option is to pass the responsibility on to the patient. This extreme form of user involvement has the superficial appeal of “political correctness” but is in reality an abrogation of responsibility. At least, while this form of “shared care” remains fashionable, clinicians who take this route are unlikely to be criticised, but something has gone wrong with the clinical relationship (or for that matter any client-professional relationship) when clinicians quote figures, survival rates, and probabilities of side effects at their patients instead of offering them educated professional guidance and support in the difficult process of reaching a decision.


Each of the above options may have a role but both carry a risk, particularly when applied in high doses, of undermining the relationship between the patient and the clinician. One aspect of this relationship is the creation of some belief in the process of healing. The placebo effect is mostly encountered in research and has become regarded primarily as an obstacle to be overcome. Attempts have been made to dissect its effects and explain them, for example, in behavioural terms16 and the literature acknowledges that, as well as being a bane of research, the placebo effect can be considered a boon to therapy.17 The placebo effect might be enhanced by bolstering its credibility with arcane terminology but undermining the special relationship between the clinician and the patient is likely to have the opposite effect. Another crucial function of the clinician-patient relationship is that of containing the patient's anxiety, much of which arises out of uncertainties of various kinds. To be able to do this, clinicians must be able to shoulder a substantial part of that anxiety themselves and this is only possible if they are not assailed by too many personal anxieties arising out of the clinical situation—such as summary suspension lurking round the corner. While profession-led audit may be fully justified to raise standards and reduce the incidence of errors, the current emphasis on personal error places too great a burden of personal anxiety on individual clinicians. The results are likely to include disaffection, burn-out, and a profession that seeks early retirement. The clinician requires a modicum of immunity from the consequences of mistakes, which should not be overly eroded by scrutiny, however necessary that scrutiny for malpractice or incompetence might be. Scrutiny should be focused more on systems than on individuals.


In the 1950s Isabel Menzies Lyth was asked to advise a hospital in relation to its nursing organisation.18 She and her team did so, using her very considerable psychoanalytic experience and insight. She exposed a number of mechanisms developed, largely unintentionally, as defences against the anxieties of intrinsically anxiety-provoking work. These mechanisms included denial of the significance of the individual (both patient and nurse), “the attempt to eliminate decisions by ritual task-performance”, “reducing the weight of responsibility in decision making by checks and counterchecks”, and a training system that concentrated on communication of fact and technique, as though responsibility and personal maturity were innate qualities that could be taught. The result was an incomplete avoidance of the primary anxiety, a failure to support and encourage nurses in working through and mastering those anxieties, and the creation of a good deal of secondary anxiety. What is striking, on looking back over this analysis, is the extent to which the uses of information discussed in the present paper can be described in similar terms. The writing of Lyth is just one potent example of qualitative research, not easily amenable to meta-analysis. The lesson has only been partly learned.


Physicists have long recognised the uncertainty principal.19 Uncertainty, like anxiety, cannot be “killed”. It can only be lived with, contained, or uncontained. The current attempt to “kill” uncertainty in our society risks removing from the medical profession some important “tools of the trade” and placing obstacles in their way. Too much emphasis in medical training on removing or reducing uncertainty will crowd out what little attention is being paid to educating doctors into the maturity and wisdom that they require to be able to accompany people in times of need, contain their own and their patients' anxieties and facilitate healing and recovery in an uncertain world. The public is being told, by selective publication, what to expect or demand from their doctors: factual diagnosis, when the wise doctor knows that many diagnoses are not circumscribed entities and that no tests have 100% sensitivity and specificity; factual prognosis, when the closest that can be achieved is an estimate, with a 95% confidence interval, for an average person in the population to which the patient might belong; and most effective or best available treatment, when the question “best at what?” is constantly begged and seldom satisfactorily answered.

Improving scientific knowledge is clearly a laudable objective, although it may reduce clinical uncertainty less than expected. Pretending that the clinical predicament can be reduced to a series of certainties by the recruitment of “evidence” will not work and unrealistic expectations of that stratagem may make the situation worse. Some degree of uncertainty was always here to stay and evidence, even of the highest quality, is only evidence. There will always be judgments to be made by responsible, informed, and compassionate people. They may not be able to perform these broader roles, in communication, holding anxieties and managing uncertainty, unless trained for and supported in them.

Despite increasing medical knowledge uncertainty will always remain


View Abstract

Request permissions

If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.