|Home | About | Journals | Submit | Contact Us | Français|
The algorithmic approach to guidelines has been introduced and promoted on a large scale since the 1970s. This study aims at comparing the performance of three algorithms for the management of chronic cough in patients with HIV infection, and at reassessing the current position of algorithmic guidelines in clinical decision making through an analysis of accuracy, harm and complexity.
Data were collected at the University Hospital of Kigali (CHUK) in a total of 201 HIV-positive hospitalised patients with chronic cough. We simulated management of each patient following the three algorithms. The first was locally tailored by clinicians from CHUK, the second and third were drawn from publications by Médecins sans Frontières (MSF) and the World Health Organisation (WHO). Semantic analysis techniques known as Clinical Algorithm Nosology were used to compare them in terms of complexity and similarity. For each of them, we assessed the sensitivity, delay to diagnosis and hypothetical harm of false positives and false negatives.
The principal diagnoses were tuberculosis (21%) and pneumocystosis (19%). Sensitivity, representing the proportion of correct diagnoses made by each algorithm, was 95.7%, 88% and 70% for CHUK, MSF and WHO, respectively. Mean time to appropriate management was 1.86 days for CHUK and 3.46 for the MSF algorithm. The CHUK algorithm was the most complex, followed by MSF and WHO. Total harm was by far the highest for the WHO algorithm, followed by MSF and CHUK.
This study confirms our hypothesis that sensitivity and patient safety (i.e. less expected harm) are proportional to the complexity of algorithms, though increased complexity may make them difficult to use in practice.
The algorithmic approach to guidelines has been introduced and promoted on a large scale since the 1970s. This flowchart representation of step-by-step clinical logic guides the management of a patient with symptoms, clinical signs, or results of technical examinations. The transition from one step to the next is mostly dichotomous, which means that only one out of two choices can be made at each step. Moreover, the logic is serial: only one pathway can be followed by a single patient.
The original purpose of algorithmic guideline implementation was twofold. First, with continuing concern over the rising costs of health care, health policy makers have been impressed by the high levels of variation and inappropriateness in medical care. Second, in resource-poor settings, nurses provide frequently the first-line medical care. To deal with the relative lack of knowledge in this medical professional group, algorithms have been used to help rationalise and standardise clinical decision making, an approach considered so far to be effective and efficient . Algorithms have been pivotal in the task-shifting or task-sharing by nurses that has been promoted over the last 10 years to increase access to antiretroviral therapy.
Clinical algorithms are often controversial: on the one hand, they have often been considered with suspicion because of their restrictive, "cookbook"-medicine approach. On the other hand, they have been viewed as a valuable educational tool, providing neophytes with a method for making diagnostic and therapeutic decisions in clinical practice apprenticeship . Although both of these views may make sense, none of them has been validated.
Algorithms usually lead to a single diagnosis, potentially neglecting another serious one, leading to delayed investigations and treatment. In some situations, the burden of harm might grow considerably: when we follow an algorithmic guideline step-by-step, it is obvious to stop at the first diagnosis corresponding with the patient's findings. It induces a tunnel-vision and limits the search of certainty in diagnosis.
Epidemiology shows the relevance of the dynamics of the disease spectrum in time as well as in space. The utility of a given clinical algorithm can vary substantially across different epidemiologic and socioeconomic contexts. As a result, every algorithmic guideline must be adapted to every local context and regularly revised . Moreover, the emphasis on development and deployment of medical guidelines has led to the introduction of multiple versions addressing the same problem: uncertainty is increased with the fact that clinicians have to select among several guidelines. Barak et al. tried to solve this problem by creating a method of text-to-algorithm conversion .
Several studies have evaluated or compared guidelines but none of them has compared them in terms of complexity or harm [4-7]. Wabitsch, in his efforts to achieve reproducible, clinically flexible and applicable guidelines, developed a systematic procedure to tailor guidelines to specific settings. Doing this, he compared two modified guidelines with a global one, but he did not consider harm . Because of the impact guidelines may have on clinical practice, Pearson et al. proposed a method to develop, use and compare them . This methodology-the Clinical Algorithm Nosology (CAN)-measures overall design complexity independent of algorithm content, quantitatively delineates algorithm differences, and qualitatively estimates the potential impact of such differences on patient care, by scoring the degree of similarity. This simple method may help clinicians to make the best choice between several algorithms about the same subject or complaint in their setting.
At the end of 2007 the global estimate of the number of people living with HIV was about 33,2 million, with more than 96% in low- and middle-income countries . In Rwanda, HIV prevalence turns around 2,9 to 3,2% .Except for some countries, the majority of people living with HIV still have no access to highly active antiretroviral therapy (HAART). Chronic cough is a key problem of medical decision making in the management of immune-compromised HIV patients, since a large cluster of dangerous but treatable diseases present this way.
The purpose of this research was to compare virtually three algorithmic guidelines intended for use at a tertiary level in terms of sensitivity, delay to diagnosis, harm, complexity and similarity. We intended to evaluate how complexity relates to accuracy and expected harm. We hypothesized that increasing complexity, thereby approaching the clinician's intuitive logic, improves accuracy and diminishes expected harm, but that, on the other hand, the degree of complexity can become too high and make the algorithm less useful in practice.
Data were collected at the University Hospital of Kigali (CHUK), a national reference centre for Rwanda, during a 4 months period in 2002. The study included all adults (> 15 years of age) hospitalised in the internal medicine department with cough of more than three weeks' duration as their chief complaint, known or newly confirmed HIV-positive and suspected of World Health Organisation (WHO) stage 3 or 4 . Patients who had a known non-infectious cause of chronic cough (chronic obstructive lung disease, bronchial asthma, pleural effusion due to cardiac or renal insufficiency) were excluded from the study. The reference standard for the actual disease of each patient was the final diagnosis made by clinicians at the end of the follow-up with the available diagnostic tools (e.g., chest X-ray, Gram and Ziehl-Neelsen stain, culture; bronchoscopy was done for selected cases).
Clinical care depended only on the clinician's judgment. Moreover, clinicians were blinded to hypothetical results of the application of the algorithms.
For each patient, we collected data through a questionnaire in Epi-Info software 2000: age, sex, history, physical examination, tests done, treatment, outcome and final diagnosis. Data were made anonymous for analysis. Sensitivity was computed with and without considering patients for whom no final diagnosis was made and those with a rare diagnosis not included in the algorithms. For the remainder of the analysis they were excluded.
Using the graphic format described by Margolis, a serial and dichotomous flowchart representation adopted by the Harvard Community Health Plan, a locally tailored algorithm was developed based on clinicians' logic and availability of complementary exams (Figure (Figure1).1). Two residents, under the supervision of their professor of internal medicine, created the CHUK algorithm. They were not amongst the clinicians who made the final diagnosis considered as the gold standard. Clinicians only controlled the construction of the algorithm for inconsistencies.
The Médecins sans Frontières (MSF) algorithm was published in 2001 by L. Lynen for management of opportunistic infections in HIV-positive patients, and based on an extensive literature review (Figure (Figure2)2) [12,13]. The WHO algorithm was published earlier in the 1990s (Figure (Figure3)3) . Both publications proposed guidelines for care at three levels. In our assessment of the three algorithms we only used the chronic cough algorithm for the tertiary care level.
We virtually applied the three algorithms to each patient, using data from the database, simulating a step-by-step management as would have been done following the algorithm. We calculated the number of patients who found a way through the algorithm ("fit") and we compared their management in the ward with the pathway they would have followed with the algorithm. Sensitivity represents the proportion of correct diagnosis made by an algorithm. We estimated the sensitivity first for each diagnosis, and further for the algorithm as a whole, with the final clinical diagnosis as the "true condition". If the algorithm missed a diagnosis, we considered this as a false negative for the true condition. If another diagnosis was reached instead of the true condition, it was considered as a false negative for the true condition and as a false positive for the wrong condition. In other words, the false positive is a wrong diagnosis whereas the false negative is a missed diagnosis.
Starting from the clinical state box to the action box, we defined for each diagnosis a minimal hypothetical (ideal) delay, depending on the execution of the different investigations. For every patient we compared the time clinicians needed in the ward to reach a therapy, with the delay predicted by the different algorithms. Delays were considered on theoretical grounds, supposing that all steps had to be taken in a sequential way (which clinicians often don't do: e.g., if the X-ray shows a miliary pattern, they will not wait for all three sputa examinations).
We compared the algorithms following the Clinical Algorithm Nosology. It is made up of two components . The Clinical Algorithm Structural Analysis (CASA) measures overall design and flow pattern complexity independent of algorithm content. A formula expresses the relative complexity of a clinical algorithm in a score:
where n1Dx = number of diagnostic boxes, n2D0 = number of all other boxes, and ∑ (Lp) i = the summation of "loop parameters" (i.e. nodes that send the clinician back to an earlier point). The Clinical Algorithm Patient Abstraction (CAPA) then accomplishes a clinical comparison of two algorithms and explains the differences between them. The CAPA has three steps: clinical rule analysis, patient abstraction and patient comparison. "Identical" (score = 10) means same order of clinical exams and decision, "similar" (score = 8) means the same steps and treatment but in different order and "different" (score = 0) means a management without the same diagnosis or therapeutic decision.
With a questionnaire featuring scoring scales from 0 to 10, nine internal medicine clinicians of CHUK estimated intuitively the weight of harm by commission (false positives) and harm by omission (false negatives) in therapeutic decisions compared to natural death . We relied only on intuitive estimations, after a training session explaining the different key factors determining harm: disease and treatment morbidity and mortality, treatment efficacy and cost, stigma, contagiousness. For each algorithm we computed the harm per diagnosis, multiplying false positives and false negatives with their respective harm, and summing these for the overall harm.
Data were recorded for 201 patients. Nine patients had no diagnosis, 9 had a rare diagnosis not included in the algorithms, leaving 183 for analysis. One patient had two probable diagnoses. Sex ratio was 82/119 M/F (0,69), mean age was 35,9 years (range 15-70 years). Baseline characteristics and main diagnostic frequencies are summarized in table table1.1. (Table (Table1)1) All forms of tuberculosis were by far the most frequent diagnosis (46,4%). Mortality was high (39%).
Of the 184 cases (one patient with 2 diagnoses) we made follow the CHUK algorithm, 183 fit a pathway (99,5%). Comparing the algorithm pathways with the one followed by the clinicians, for 61% the management was "identical", for 38% it was "different" and for 1% it was "similar".
The same procedure was also applied for each of the 184 cases to the other two algorithms under study. A fit was found in 183 (99.5%), 183 (99.5%), and 171 (92.9%) cases for the CHUK, MSF and WHO algorithms, respectively. Sensitivity was 95.7%, 88% and 70.1% for the CHUK, MSF and WHO algorithms, respectively (87.5%, 80.5% and 64.2 while including patients with unknown and rare diagnosis). (Table (Table22)
Table Table33 shows that for most diagnoses a certain delay to appropriate management was present. (Table (Table3)3) For example, for empyema, seven days were lost by the CHUK clinicians before a therapeutic action. When we compare the average delay of CHUK and MSF guidelines with the hypothetical minimal delay according to the local situation regarding availability and speed of execution of tests and treatments, the CHUK algorithm was the fastest, followed by that of MSF. For the WHO algorithm, delays were difficult to calculate because of missing practical specifications.
The CHUK algorithm was the most complex with a CASA score of 80, followed by MSF with a score of 73 and WHO with a score of 40. The CAPA score was 0 meaning that there is (almost) no similarity between the three algorithms.
Table Table44 shows the average intuitive harm for false positives and negatives for each disease, weighted by the nine internal medicine clinicians compared to natural death. On average, the weight of omission error was much higher than the weight of commission error (8 versus 4.5). The total harm was the highest for the WHO algorithm with a score of 487.5. On the other hand, the CHUK algorithm was by far the least harmful followed by MSF algorithm with scores of 91 and 240, respectively. If we split total harm into "false-positive harm" and "false-negative harm", we obtain, respectively, for CHUK scores of 27 and 63, for MSF 64 and 176 and for WHO 60 and 427.
The purpose of this study was to assess the hypothetical utility of three algorithms in terms of sensitivity, delay to action and harm at a tertiary care level. The CHUK guidelines proved to be the most complex, but the most sensitive and the least harmful. The results confirm that increasing complexity improves sensitivity and diminishes expected harm. But the degree of complexity becomes so high that it might be one of the reasons why clinicians are reluctant to use them. As a matter of fact, none of these algorithms has ever been used in daily practice in CHUK.
Due to limited diagnostic means, the gold standard in this study is the clinical diagnosis made by the clinicians at CHUK. This might be considered as a weakness. However, the prevalence of final diagnoses is close to the spectrum of diseases that cause chronic cough in people living with HIV in Africa : tuberculosis remains the main diagnosis, found in 46% of cases in our study. It is now well described that tuberculosis is the leading opportunistic infection in most low-resource settings, unlike industrialised countries in which a preponderance of Pneumocystis jirovecii (formerly carinii) is seen. Our study reveals a prevalence of 19% of pneumocystosis. This is fourfold the prevalence found by Batungwanayo in the same hospital ten years earlier, but this result was based on microbiological evidence after bronchoscopy and was restricted to unsolved cases . In Uganda bronchoscopy in smear negative patients showed pneumocystosis in 38% . Since in our study the diagnosis of pneumocystosis was made mainly radiologically, and since we consider all patients with persistent cough, these data are difficult to compare. A review in Thailand of causes of all respiratory symptoms in HIV infected patients found 23% cases of pneumocystosis, a rate more similar to our data .
In our study, mortality rate was high (39%). At that moment in time, ART was not subsidized; patients feared hospitalisation in our wards because they knew beforehand the high mortality, leading to late referral or consultation. Beyond this, some deaths could be attributed to missed diagnosis of uncommon diseases.
The average delay to appropriate management would be shorter for the CHUK algorithm (1.8 days) than for the MSF (3.4 days). The CHUK delay is therefore closer to that seen in clinical practice at CHUK, as well as to the ideal one (Table (Table3).3). The theoretical differences between CHUK and MSF algorithms can be explained by the sequence of investigations: CHUK algorithm focused first on an X-ray to diagnose urgent pathologies, while MSF first focuses on sputum smears to diagnose the most frequent pathology (TB). In practice the X-ray results are read on day 1 while the sputa are being analyzed.
The delay of the WHO algorithm was difficult to evaluate because this guideline is mostly theoretical and gives no practical details: reading it seems simple, but to put it into practice is more complex and becomes impossible for several diagnoses, leaving a lot of freedom to the clinician. To illustrate this, there is no way to distinguish lobar and atypical pneumonia; in pleurisy, they recommend doing a tap, but there is no explanation about the interpretation of the results or about steps to take.
The analysis of the algorithms brings to the fore that the CHUK algorithm is the one that leads to the highest sensitivity and induces the lowest harm in terms of morbidity and mortality. This is probably because we tried to come as close as possible to the intuitive reasoning of the experienced CHUK clinicians. The algorithm was developed by trainees of the CHUK clinicians, and CHUK clinicians are used in this analysis as the gold standard. On the other hand, it is by far the most complex with a high CASA score of 80. Part of the high complexity in the CHUK algorithm is explained by the presence of many loops. We see a clear inverse relation between the increasing sensitivity and the decreasing harm for all three algorithms. In addition, with regards to the sensitivity we note that despite such a complexity, 8 patients (4,3%) would not reach a correct final diagnosis.
Firstly we noticed that in the set of patients and also in the different diagnoses available in the CHUK algorithm, uncommon but treatable diseases such as aspergillosis, nocardiosis, cryptococcosis, and some others are not taken into account. These diseases, however, are well described as opportunistic infections in HIV-positive patients and deserve priority before considering frequent diseases because of their seriousness and treatability. We omitted them from the analysis since algorithms cannot cover all rare diseases.
Secondly, in current practice in low resource settings, diagnosis of pulmonary infection is often a presumption, based on clinical signs and radiological results because of the limited availability of serology, culture or PCR. These presumptive diagnoses were considered as gold standard-not always with formal proof. It casts doubt on the reliability of the comparison of the diagnosis obtained by following the algorithm with the gold standard diagnosis, as the latter may have been false. The ideal situation should offer availability of complementary exams to confirm these presumptive diagnoses, but it would imply submitting patients to invasive procedures and spending much more money. Moreover, in advanced AIDS, more than one pathogen is frequently identified [18,19].
Finally, a large part of the difference in sensitivity of the CHUK and MSF algorithms is due to one single step differentiating between lobar and atypical pneumonia.
The three algorithms might not be comparable: the WHO algorithm leaves much to the discretion of the clinician, without specifications.
The CAN methodology can be applied easily. By itself, however, it does not judge which of the algorithms is likely to provide better patient care. This sole method of comparison is not really used in practice and remains an elusive concept. It is only a comparative tool to quantify the degree of similarity and complexity with no attention for sensitivity and harm which is the most important quality of a diagnostic aid. The authors mention that the ideal objective in the comparison of algorithms would involve decision analyses of the management pathways and prospective clinical trials comparing patient outcome in guided care . Decision analysis would include sensitivity analysis of sensitivity and specificity of every finding for every concerned disease, of different sequences of nodes, and expressing global harm in DALY's or QUALI's for every disease. An almost impossible task, at least for complex algorithms as the ones discussed here. On the other hand, prospective clinical trials would involve huge sample sizes, as for every disease a sufficient number of patients should be included. Moreover, ethical requirements should be met.
So far, a practical way of comparing clinical guidelines proving their reliability before their implementation is still needed to validate any guideline program . In our study, we did not apply algorithms in real patients and evaluate the outcome. We consider that our methodology is anyhow a necessary first step in validation: a correct scientific output is reached without the risk of increasing harm applying algorithms in real clinical decision making.
Intuitive weighting of harm is a gross intuitive estimation given by clinicians on a simple scale scoring from 0 to 10 about commission error (false positives) and omission error (false negatives) in therapeutic decisions. Calculated weighed harm is more accurate than intuitive weighed harm, because it is based on observed probabilities related with the outcome of treated and non treated diseased and non diseased patients. In a recent study, we proved that the ratio of the intuitively estimated harm was almost 32 times lower than the calculated based on (by the same clinicians) intuitively estimated influencing factors, and two times lower than the calculated weighed harm based on literature data . For our study and purpose, we relied only on the intuitive one to draw conclusions, conscious of the lack of accuracy.
Our study has shown that some published algorithms have limited accuracy. This can be overcome by increasing the complexity, at the possible expense of user-friendliness. We hope that this study encourages policymakers to "dry-run" new draft-algorithms with an existing database of a cohort of patients, before organizing a formal validation in a real situation. The latter is not always possible, given the lack of a gold standard in most situations where the algorithm is to be applied.
A second impact of this study might be that writing algorithms for problems with a large differential diagnosis at the tertiary care level should be abandoned for many reasons, for instance: the burden of work necessary to adapt the algorithm to time and place, the neglect of rare but serious and treatable diseases, the possible coexistence of several pathologies, the reduction of a clinical presentation to only a few signs, and the difficulty in reaching a reasonable accuracy. We might state that at this level of care, medical decision tools based on a serial dichotomous approach may be replaced by the more human parallel and weighed logical approach recently proposed [20-23].
Little research has been done to evaluate the real-world use of algorithms by health professionals [1,3]. Anecdotal evidence shows that most nurse practitioners do not follow their algorithms. As far as we are concerned, doctors never follow them for diagnosis; simple therapeutic flow-charts on the other hand, do well.
The concept of threshold approach in clinical decision-making has existed since the 1970s, and this concept has been substantiated to the point that ignoring it would be unethical [24-26]. All the more, a threshold approach might solve many conflicts provoked by the algorithmic logic: future studies will have to demonstrate it.
Considering that application of an algorithmic approach involves a short course training, especially for nurses and inexperienced personnel, future studies should assess which logical approach is the most effective for a given amount of training .
Inexperienced medical staff could go on using the algorithmic approach in order to solve simple problems at the first level of care. The condition is that they should have a short basic training and that the flowchart should be validated, adapted to the local context and regularly revised. But at the tertiary care level, approaching the competence of specialised clinicians by improving sensitivity and diminishing expected harm will increase complexity at that point that it turns out almost impossible to apply.
The authors declare that they have no competing interests.
CM contributed to the concept, data collection and analysis, JN contributed to analysis and writing, BB contributed to concept and analysis, LL contributed to analysis and supervision of writing, YC contributed to data collection and analysis, AA contributed to concept, data collection and analysis, JVDE contributed to concept, analysis, writing and overall supervision. All authors read and approved the final manuscript.
The pre-publication history for this paper can be accessed here:
The authors wish to thank all clinicians involved in the care of the patients in the internal medicine ward of CHUK.