In recent years I’ve been doing a lot more prehospital care. To be honest, I’ve found it very challenging, although it is also rewarding. In the prehospital setting, we don’t have a lot of diagnostic tools, so we are pretty reliant on the history and clinical examination of our patients to decide what to suspect and what to do about it. This is because we simply don’t have the same capacity or kit in the prehospital service that we do in hospital. The primary survey really relies on the clinical skills of the clinician, and although we do now carry ultrasound to the scene, there is relatively little additional testing information available to us.
Similarly, in the Emergency Department I am often making quite difficult probabalistic decisions based on relatively simple clinical information and I know that I don’t get it right all the time. For example, I’ve given blood to patients who subsequently probably did not need it, and I’ve been a bit late with the blood for patients who probably did! I suspect I am not alone in reflecting that it’s a tricky and uncertain world we practice in.
Is it just me, though, or is it just tricky? Well, this week, we have a paper that might help us answer the question. Special thanks to Cliff Reid for spotting this one and posting it on Twitter. The abstract is below, but please read the full paper yourself and make your mind up.
The Abstract – Diagnostic accuracy of clinical examination​1​
Background – Timely and accurate identification of life- and limb-threatening injuries (LLTIS) is a fundamental objective of trauma care that directly informs triage and treatment decisions. However, the diagnostic accuracy of clinical examination to detect LLTIS is largely unknown, due to the risk of contamination from in-hospital diagnostics in existing studies. Our aim was to assess the diagnostic accuracy of initial clinical examination for detecting life- and limb- threatening injuries (LLTIS). Secondary aims were to identify factors associated with missed injury and overdiagnosis, and determine the impact of clinician uncertainty on diagnostic accuracy.
Wohlgemut JM, Marsden MER, Stoner RS, et al. Diagnostic accuracy of clinical examination to identify life- and limb-threatening injuries in trauma patients. Scand J Trauma Resusc Emerg Med. Published online April 7, 2023. doi:10.1186/s13049-023-01083-z
Methods – Retrospective diagnostic accuracy study of consecutive adult (≥ 16 years) patients examined at the scene of injury by experienced trauma clinicians, and admitted to a Major Trauma Center between 01/01/2019 and 31/12/2020. Diagnoses of LLTIs made on contemporaneous clinical records were compared to hospital coded diagnoses. Diagnostic performance measures were calculated overall, and based on clinician uncertainty. Multivariate logistic regression analyses identified factors affecting missed injury and overdiagnosis.
Results – Among 947 trauma patients, 821 were male (86.7%), median age was 31 years (range 16-89), 569 suffered blunt mechanisms (60.1%), and 522 (55.1%) sustained LLTIS. Overall, clinical examination had a moderate ability to detect LLTIs, which varied by body region: head (sensitivity 69.7%, positive predictive value (PPV) 59.1%), chest (sensitivity 58.7%, PPV 53.3%), abdomen (sensitivity 51.9%, PPV 30.7%), pelvis (sensitivity 23.5%, PPV 50.0%), and long bone fracture (sensitivity 69.9%, PPV 74.3%). Clinical examination poorly detected life- threatening thoracic (sensitivity 48.1%, PPV 13.0%) and abdominal (sensitivity 43.6%, PPV 20.0%) bleeding. Missed injury was more common in patients with polytrauma (OR 1.83, 95% CI 1.62-2.07) or shock (systolic blood pressure OR 0.993, 95% CI 0.988-0.998). Overdiagnosis was more common in shock (OR 0.991, 95% CI 0.986-0.995) or when clinicians were uncertain (OR 6.42, 95% CI 4.63-8.99). Uncertainty improved sensitivity but reduced PPV, impeding diagnostic precision.
Conclusions – Clinical examination performed by experienced trauma clinicians has only a moderate ability to detect LLTIs. Clinicians must appreciate the limitations of clinical examination, and the impact of uncertainty, when making clinical decisions in trauma. This study provides impetus for diagnostic adjuncts and decision support systems in trauma.
What kind of paper is this?
This is a retrospective cohort study. Retrospective cohort studies are often useful in exploring hypotheses and looking for associations, but they do rely on data that was not collected for the intended purpose of the study. That does mean that they are subject to a number of biases that can affect the results. In this study, the authors were reliant on whether clinicians recorded their predictions about which life-threatening injuries were likely to be present.
Tell me about the patients.
The authors identified a cohort of trauma patients who had been admitted to a major trauma centre. The data was consecutive and only involved adult patients over roughly a two-year period. The patients were all treated by a single air ambulance service (London Air Ambulance). There is a little confusion here as this was a single MTC, but not all major trauma patients arrive by that particular service, and indeed, London HEMS transports patients to other trauma centres and trauma units. So the patient cohort is a subset of patients in this particular MTC and a subset of patients treated by this particular air ambulance service. These sort of restrictions are commonly found in retrospective cohort studies and they may introduce bias into the results.
Thermal injury patients and children were excluded.
What were they testing?
Essentially the study compares what the prehospital team thought the injuries were with the eventual diagnosis from hospital records. In this ambulance service, clinicians are required to list what life-threatening injuries are suspected so the data is at least collected contemporaneously and systematically. However, there were differences in how clinicians expressed their certainty with their assessment. The authors have tried to apply a scoring system based on adjectives used by clinicians e.g. likely vs. suspicious vs. possibly have different weightings. Whilst this seems sensible, it is based on a US model and may not, therefore, work as well in a multinational cohort of clinicians who may use different language preferences.
The final diagnoses were gleaned from the hospital records. In the UK this is a pretty robust system as it feeds into hospital records and into the UK TARN database that is a well established system for determining the injury burden of patients.
Tell me about the analysis.
This is somewhat exploratory. The authors first sought to define what they thought was a range of ideal-unacceptable outcomes for sensitivity, specificity, NPV and PPV. They did this using a consensus methodology of expert clinicians, which is reasonable as I don’t think there is any externally validated measure in practice. However, it is does lack external validity and is also seemingly defined from within the service being examined in the study. It is likely that those developing the measures would have a-priori knowledge of clinician performance from past governance processes. That might introduce bias in where the bar is set (for good or bad).
The authors then performed univariate and multivariate analyses to look for factors that might be associated with missed diagnoses. This is a reasonable approach to explore the data further, although can only define association and not causation.
What did they find?
The authors looked at 947 patients. As with most major trauma studies, the majority of patients were young and male.
Three main groups were looked at
- Life-threatening injury
- Limb threatening injury
- Life-threatening bleeding
The paper has a nice graphic shown below that shows that, across a range of conditions, the clinicians’ predictions were pretty poor, to be honest. In terms of identifying life-threatening injury (look at the sensitivity and PPV graphs), performance is especially poor on this data.
The multivariate analyses showed that polytrauma, SBP, and expressed diagnostic uncertainty were linked to overprediction of life-threatening injury.
The multivariate analyses showed that polytrauma, SBP, and heart rate were linked to overprediction of life-threatening injury.
So should we stop clinically examining the patients?
Clearly not, and that would not be the right message to take away from this paper. Clinical examination will always be a part of the initial assessment of patients, but this paper reinforces the fact that we cannot entirely rely on clinical examination. It must be supplemented by further and more definitive testing, which may, of course, include repeated clinical examination.
That said, I think this paper is hypothesis-generating rather than definitive. Whilst I probably agree with the findings I think the detailed results are subject to a number of biases. The retrospective nature of the study, the definition of the standard being created from within the team analysed, and the fact that this is a subset of service patients in a subset of the wider trauma centre will almost certainly influence the accuracy of the data, but it is unlikely to make such a difference that clinical examination would turn out to be highly accurate.
The authors’ conclusion that further diagnostic adjuncts and decision-support systems are needed seems reasonable, although these, too, would need to be tested for their effectiveness in future studies.
The bottom line is that I won’t be throwing my stethoscope away just yet, nor will I not bother to examine patients, but I will continue to be sceptical about my own and others’ clinical certainty about the presence of life-threatening injury.
References
- 1.Wohlgemut JM, Marsden MER, Stoner RS, et al. Diagnostic accuracy of clinical examination to identify life- and limb-threatening injuries in trauma patients. Scand J Trauma Resusc Emerg Med. Published online April 7, 2023. doi:10.1186/s13049-023-01083-z