Skip to main content
  • Research article
  • Open access
  • Published:

Development and validation of an algorithm to recalibrate mental models and reduce diagnostic errors associated with catheter-associated bacteriuria



Overtreatment of catheter-associated bacteriuria is a quality and safety problem, despite the availability of evidence-based guidelines. Little is known about how guidelines-based knowledge is integrated into clinicians’ mental models for diagnosing catheter-associated urinary tract infection (CA-UTI). The objectives of this research were to better understand clinicians’ mental models for CA-UTI, and to develop and validate an algorithm to improve diagnostic accuracy for CA-UTI.


We conducted two phases of this research project. In phase one, 10 clinicians assessed and diagnosed four patient cases of catheter associated bacteriuria (n= 40 total cases). We assessed the clinical cues used when diagnosing these cases to determine if the mental models were IDSA guideline compliant. In phase two, we developed a diagnostic algorithm derived from the IDSA guidelines. IDSA guideline authors and non-expert clinicians evaluated the algorithm for content and face validity. In order to determine if diagnostic accuracy improved using the algorithm, we had experts and non-experts diagnose 71 cases of bacteriuria.


Only 21 (53%) diagnoses made by clinicians without the algorithm were guidelines-concordant with fair inter-rater reliability between clinicians (Fleiss’ kappa = 0.35, 95% Confidence Intervals (CIs) = 0.21 and 0.50). Evidence suggests that clinicians’ mental models are inappropriately constructed in that clinicians endorsed guidelines-discordant cues as influential in their decision-making: pyuria, systemic leukocytosis, organism type and number, weakness, and elderly or frail patient. Using the algorithm, inter-rater reliability between the expert and each non-expert was substantial (Cohen’s kappa = 0.72, 95% CIs = 0.52 and 0.93 between the expert and non-expert #1 and 0.80, 95% CIs = 0.61 and 0.99 between the expert and non-expert #2).


Diagnostic errors occur when clinicians’ mental models for catheter-associated bacteriuria include cues that are guidelines-discordant for CA-UTI. The understanding we gained of clinicians’ mental models, especially diagnostic errors, and the algorithm developed to address these errors will inform interventions to improve the accuracy and reliability of CA-UTI diagnoses.

Peer Review reports


Catheter-associated urinary tract infection (CA-UTI) and catheter-associated asymptomatic bacteriuria (CA-ABU) are very common yet distinct forms of catheter-associated bacteriuria [1, 2]. In CA-UTI, the patient has specific urinary symptoms, and the condition merits treatment with antibiotics [1]. In contrast, CA-ABU is marked by the absence of urinary-specific symptoms, and treatment with antibiotics does not reduce mortality, bacteremia, or subsequent risk of UTI [2]. Thus, both the Infectious Diseases Society of America (IDSA) and US Preventive Services Task Force discourage screening for, and treatment of, CA-ABU in most clinical settings [3]. Recent guidelines by IDSA provide excellent summaries of the evidence supporting these recommendations [1, 2].

Despite the IDSA guidelines, inappropriate treatment of CA-ABU with antibiotics is widespread, and guidelines adoption remains modest [4]. Recent studies of CA-ABU in hospital settings show as many as 52% of patients with CA-ABU being treated unnecessarily with antibiotics [59]. The cause of CA-ABU overtreatment is multifaceted and grounded in the clinical norms and inappropriately constructed mental models clinicians use to make diagnostic decisions for patients with catheter-associated bacteriuria. Conventional teaching is that the bladder and the urine within it are sterile, but this “norm” does not apply to catheterized patients in contemporary medical settings. Making the diagnosis of CA-ABU requires the clinician to discount clinical cues, such as bacteriuria and pyuria, because neither of these can be used to distinguish between CA-ABU and CA-UTI [1, 2, 10]. Another clinical norm that runs counter to evidence is the erroneous belief that vague, non-urinary symptoms can be attributed to bacteriuria [1114]. Additionally, clinicians often overweigh the risk of withholding antibiotics while underweighting the risk of antibiotic exposure in an individual patient [15]. These evidence-discordant norms and biases produce decision-making processes that differ in distinct and clinically important ways from evidence-based guidelines for diagnosing catheter-associated bacteriuria [16, 17].

The IDSA guidelines governing catheter-associated bacteriuria are based on high-quality reviews of the available evidence [1, 2]. However, the complexity and sheer length of the guidelines (51 pages) may impede their uptake [18, 19]. Classically, diagnostic reasoning is thought to involve complex, analytical evaluations of clinical and laboratory cues to frame prior probabilities of differential diagnoses to arrive at the accurate diagnosis [20]. Empiric evidence suggests that clinicians store disease models reflecting common symptoms associated with diseases within their memory [21]; these models are called ‘mental models’. These mental models are normally constructed during training when clinicians learn the symptoms associated with diseases, and are enhanced as they gain experience throughout their career. When the mental models are not accurate, diagnostic errors may occur. Use of inaccurate (guidelines-discordant) mental models associated with CA-ABU can result in misdiagnosing CA-ABU as CA-UTI. Mental models for CA-UTI that are properly constructed (guidelines-concordant) are reflected in Table 1 left column. A guidelines-discordant mental model for CA-UTI (commonly used when misdiagnosing CA-ABU as CA-UTI) is shown in the right column of Table 1.

Table 1 Components of Clinicians’ Mental Models Diagnosis of Catheter-Associated Urinary Tract Infection (CAUTI)

In essence, prior research demonstrates that inappropriate treatment of CA-ABU with antibiotics is widespread and guidelines adoption remains modest. To address this problem, the first objective of this research was to confirm our suspicion that clinicians’ mental models are inaccurately constructed and to find the points of difference from evidence-based guidelines. The second objective was to develop a means of re-directing clinicians’ mental models by creating a valid and reliable algorithm grounded in clinical evidence, with the ultimate objective of informing a guidelines implementation intervention.


We framed the problem and our approach to the problems using a two phase study approach based on the following hypotheses. First, when clinicians attempt to differentiate catheter-associated bacteriuria as either CA-UTI or CA-ABU, their mental models include both guidelines-discordant and guidelines-concordant cues resulting in (a) poor diagnostic accuracy differentiating between CA-UTI and CA-ABU (reliability with clinical guidelines and/or clinical experts) and (b) low rates of diagnostic agreement between each other (low inter-rater reliability among non-experts). Phase 1 evaluated the accuracy and inter-rater reliability of clinicians’ mental models for catheter-associated bacteriuria. Second, we distilled the IDSA guidelines into an algorithm to attempt to improve diagnostic accuracy and inter-rater reliability by substituting guidelines-concordant cues in place of guidelines-discordant cues. In essence, the algorithm serves to recalibrate clinicians’ mental models for differentiating CA-UTI from CA-ABU. Phase 2 describes the development, preliminary validation, and evaluation of inter-rater reliability of this algorithm for recalibrating clinicians’ mental models. This research was conducted with the approval of the Baylor College of Medicine Internal Review Board (protocol H #24180).

Phase 1 – Clinicians’ decision-making when diagnosing CA-UTI

Study design

We conducted a comprehensive assessment of the diagnostic cues clinicians use when distinguishing between CA-UTI and CA-ABU, through a case-based diagnosis exercise followed by in-depth, cognitive interviews [22].


Participants consisted of six physicians and four allied health professionals recruited from a convenience sample of experienced clinicians working in local acute and extended care facilities of a single health system. The sample of clinicians included three males and seven females with a range of 11-15 years of experience treating older patients in long-term care and inpatient settings (see Table 2). All 10 participants reviewed each of the four cases for a total of 40 cases on which analyses are based.

Table 2 Characteristics of study participants


Clinicians were asked to review the electronic medical records of four patients with positive urine cultures associated with an indwelling urinary catheter. Henceforth, for simplicity, we will refer to these positive urine cultures, both bacteriuria and funguria, as “bacteriuria”. All participants independently reviewed the same four urine cultures, representing four distinct cases. The selected cases were actual patient cases representing a spectrum of clinical cues and treatments representative of CA-UTI and CA-ABU. Each case presentation had at least one clinical cue shown in prior studies to influence physicians’ decision-making regarding antimicrobial treatment (e.g., older age, pyuria, and type of organism) [6, 23]. Table 3 describes the patient cases. For each of the four cases, clinicians answered two written questions: (1) Do you feel this is a CA-UTI or CA-ABU? and (2)What helped you decide if this case was a CA-UTI or CA-ABU? Subjects then underwent a cognitive interviewing exercise [22] to elicit their reasoning processes where they answered the following question: “For the case of Patient X, was your decision of CA-UTI versus CA-ABU influenced by any of the following?” Choices included pyuria (white blood cells in urine), systemic leukocytosis (white blood cells in bloodstream), type of organism in the urine (Gram negative, Gram positive or fungal), elderly or frail patient, weakness, cloudy urine, foul-smelling urine, and specific urinary symptoms (e.g. dysuria—painful urination or frequency). The researcher stated each probe one-at-a-time, after which the clinician responded with a ‘yes’ or ‘no’. To assess use of the guidelines, clinicians were then asked “For the case of Patient X, do you feel that you applied the IDSA guidelines to arrive at your decision?” With participant consent, all interviews were tape recorded and subsequently transcribed and analyzed.

Table 3 Description of bacteriuria cases diagnosed by long-term care providers and signs/symptoms endorsed

Analyses for phase 1

First, we categorized each of the signs or symptoms participants identified as influencing their decision as (1) a guidelines-concordant clinical cue for distinguishing between CA-UTI and CA-ABU, or (2) a guidelines-discordant clinical cue that should not be used to distinguish between CA-UTI and CA-ABU. We then calculated Fleiss’ kappa to examine the overall inter-rater reliability of diagnoses for the four cases across all 10 clinicians, as well as the inter-rater reliability of clinicians that reported using the guidelines in their decision-making and those that reported not doing so. The MAGREE.SAS macro in SAS Version 9.2 was used to calculate the generalized kappa of Fleiss.

Phase 2 – guideline-based algorithm development and validation

Algorithm development

We prepared a diagnostic algorithm for catheter-associated bacteriuria based on the IDSA guidelines. This evidence-based, diagnostic algorithm was designed to improve clinicians’ diagnostic ability to distinguish between CA-UTI and CA-ABU. The first version of the algorithm, developed according to the 2005 IDSA guidelines on CA-ABU and the 2009 IDSA guidelines on CA-UTI [1, 2], was formatted as a flowchart to fit onto a pocket-sized card for high portability. The algorithm was evaluated for content and face validity, and revisions were made accordingly after each evaluation (see below for details). The final version of the algorithm was then used by three trained research personnel to classify cases of bacteriuria as either CA-UTI or CA-ABU (See Figure 1).

Figure 1
figure 1

Final form of the comprehensive algorithm. This figure represents two sides, front and back, of a pocket card used in our guidelines implementation project, “Kicking CAUTI Campaign.”

Algorithm validation: content validity

To examine content validity of the algorithm, the initial version was distributed via email to all 11 expert panel members of the IDSA CA-UTI and CA-ABU guidelines committee. In a cover letter to our email message, we asked three survey questions (see Table 4 for specific wording); each question maps to one of the following standards: criterion, diagnostic, and feasibility. The response to each question was scored on a 10-point scale, with higher numbers indicating stronger agreement. We also provided a space for respondents to make comments or suggestions regarding the algorithm. Seven (64%) panel members responded with a numeric score, while one additional panel member provided comments only. The mean score for each question and each respondent was computed, and comments were tabulated for review by our research team.

Table 4 Ratings of the diagnostic algorithm by expert members of the Infectious Diseases Society of America (IDSA) guidelines panel for diagnosing and managing catheter associated bacteriuria

Algorithm validation: face validity

Design and participants

We conducted cognitive interviews with six non-expert clinicians recruited from a purposive sample of clinicians working in local acute and extended care facilities to evaluate face validity of the algorithm. These participants, four internal medicine resident physicians, one nurse practitioner, and one physician assistant were chosen because all routinely provide care for catheterized adults and thus would be potential users of the algorithm.


Participants were asked the following question regarding each step of the algorithm: “As you look at this diamond (decision point) or box (procedure step), what are you thinking it refers to?” Cognitive interview responses were categorized based on whether there was a misunderstanding, wrong interpretation, over-interpretation, correct interpretation, or off-topic response to each step of the algorithm [22]. The number of responses in each category was tallied, and percentage of total responses falling into each category was calculated. Based on the answers given, the algorithm was further revised to its ultimate form. This final version (see Figure 1) was sent back to the lead authors of the relevant IDSA guidelines [1, 2] for their final input; neither suggested any substantial changes.

Algorithm reliability: inter-rater reliability between non-experts and clinical expert

In order to determine if use of a guideline concordant algorithm has the potential to recalibrate inaccurately built mental models resulting in improved diagnostic accuracy of CA-UTI and CA-ABU, one expert and two non-expert providers used the algorithm to classify 71 distinct cases of catheter-associated bacteriuria arising in our local acute and extended care facilities as either CA-UTI or CA-ABU. Reliability ratings between non-expert and expert raters were used to confirm the ability of the algorithm to improve diagnostic accuracy. We also calculated inter-rater reliability of the algorithm between ratings of the two non-experts. Cases were chosen consecutively at 3-4 month intervals over a 10 month period. Each case was classified independently by the expert and at least one non-expert, resulting in 110 paired comparisons (as 20 cases were rated by the expert and both non-experts). Non-expert raters were trained to use the algorithm in introductory exercises prior to performing Cohen’s simple kappa the case classifications. Following procedures identical to those we used in Phase 1, the three raters were each given full access to the patients’ medical records including dates of the relevant urine cultures. Raters classified each case independently, and each rater was blinded to the other raters’ classifications. Because we were interested in the inter-rater reliability between specific pairs of raters, Cohen’s simple kappa was used to examine inter-rater reliability of accurate diagnoses aided by the algorithm between the expert and each non-expert and explore inter-rater reliability of the algorithm between non-experts [24].


Phase 1– clinicians’ decision-making when diagnosing CA-UTI

Twenty-one (53%) of the 40 bacteriuria diagnoses by study participants were guideline-concordant (Table 3). Furthermore, only seven of the ten clinicians said that they applied the IDSA guidelines to arrive at their decisions; the other three said they did not use the guidelines by intention or they had not read the guidelines.

In terms of guideline-concordant clinical cues, six clinicians consistently identified fever as a guideline-concordant cue for CA-UTI (Table 3), and two clinicians commented on the presence or absence of urinary symptoms as being influential in their decision-making. On any given case, guideline-discordant cues (Table 1, right column) were endorsed by five or more clinicians when attempting to distinguish CA-ABU from CA-UTI.

All ten clinicians correctly identified case 4 as CA-ABU (Table 3), but some for the wrong reasons, as six reported that the organism type (Candida) influenced their decision, and three reported that the lack of leukocytosis influenced their decision. Interestingly, three clinicians also cited the presence of a chronic catheter in an elderly patient as a significant factor in their decision. This sign lead them to think that the patient was more likely to be colonized with a fungal agent. In Table 3 case 3, only one clinician correctly identified this case as a CA-ABU as the others were misled by guideline-discordant cues, e.g., leukocytosis and weakness in the patient. In Table 3 case 2, six clinicians correctly identified fever as a clinical cue for CA-UTI, but only four clinicians concluded that the patient had CA-UTI. Additionally, three clinicians each endorsed organism type and low number of organisms as leading them to conclude the patient had CA-ABU. In essence, clinicians could not distinguish which of these cues were guideline-concordant for CA-UTI. For Table 3 case 1, six clinicians arrived at the diagnosis of CA-ABU, but seven reported the guideline-discordant cues of leukocytosis as influential in their decision making. In this case, the patient was on oral steroids and therefore had an alternative explanation for his leukocytosis. The IDSA guidelines explicitly state that leukocytosis is not a reliable clinical cue for CA-UTI [1, 2].

Inter-rater reliability among all 10 clinicians was fair (Fleiss’ kappa = 0.35, 95% CIs = 0.21 and 0.50) [24]. Inter-rater reliability among the seven clinicians reporting that they used the guidelines was also fair (Fleiss’ kappa = 0.28, 95% CIs = 0.07 and 0.50). Inter-rater reliability among the three clinicians reporting that they did not use the guidelines was substantial (Fleiss’ kappa = 0.63, 95% CIs = 0.06 and 1.00), i.e., they arrived at the same diagnosis, but these diagnoses were not always guideline-concordant [24]. Therefore, despite the higher reliability rating, these clinicians’ mental modes resulted in poorer diagnostic accuracy.

Phase 2 – guideline-based algorithm development and validation

Content validity

Eight (73%) of the 11 members of the IDSA guidelines committee responded to our request for comments on the original algorithm. We received 27 specific comments addressing about half of the processes (boxes) or decision points (diamonds) in the algorithm. Ten (37%) of these overall comments concerned changing the recommended duration of treatment to reflect the patient’s response to therapy; we had misinterpreted this point in the guidelines. Seven of 11 members of the IDSA guidelines committee scored the algorithm along three standards (criterion, diagnostic, and feasibility) for measuring the quality of the algorithm content. Table 4 provides the mean ratings provided by IDSA guidelines committee members for each standard with each of the standards having an acceptable mean rating between 7.1 and 8.1. We modified the algorithm format to fit standard flow-charting in response to specific suggestions.

Face validity

Non-expert clinicians (see second portion of Table 2) reviewed the algorithm. Cognitive interviews with six clinicians produced comments for a total of 164 distinct comments about processes (boxes) or decision points (diamonds) contained within the algorithm. For 123 (75%) of the comments, respondents correctly interpreted the meaning of the process or decision point of the algorithm. Eighteen (11%) of the responses were incorrect, 13 (8%) were over-interpreted, 7 (4%) were misunderstood and 3 (2%) responses were off-topic. An example of over-interpretation was a box that stated “work up another cause,” which was interpreted as “do a chest x-ray and obtain an abdominal film.” The diamond that received the greatest number of incorrect responses was originally worded “Bacteriuria ≥103 CFU/ml?”. We subsequently modified the text to read “Were there more than 1000 organisms/ml?” to reduce the need for real-time mathematical transformations and make interpretation unambiguous. The algorithm was modified to address the issues raised in the cognitive interviews, leading to its final form (see Figure 1), as approved by the two lead guidelines authors.

Reliability of the algorithm for diagnostic accuracy and inter-rater reliability

Using the final version of the algorithm, three providers classified 71 cases of catheter-associated bacteriuria. Of these cases, 28 were CA-UTI and 42 were CA-ABU as per the IDSA definitions for these conditions. Forty-nine cases were rated by both the expert and non-expert #1, forty-one were rated by both the expert and non-expert #2, and twenty cases were rated by both non-expert #1 and non-expert #2. Inter-rater reliability between the expert and non-expert #1 was substantial (Cohen’s kappa = 0.72, 95% CIs = 0.52 and 0.93). Inter-rater reliability between the expert and non-expert #2 was almost perfect (Cohen’s kappa = 0.80, 95% CIs = 0.61 and 0.99). The average inter-rater reliability among the expert and both non-experts was substantial (average Cohen’s kappa = 0.76), suggesting improved diagnostic accuracy among non-experts with the clinical expert (criterion standard). Inter-rater reliability between the two non-experts was also substantial (Cohen’s kappa = 0.88, 95% CIs = 0.64 and 1.00) [24].


Our data show that clinicians who routinely care for patients with urinary catheters use mental models that are often guidelines-discordant when classifying cases of catheter-associated bacteriuria as either CA-UTI or CA-ABU. Their decision cues consist of a heterogeneous group of signs and symptoms, many of which are not supported by evidence or run counter to evidence, as per IDSA guidelines [1, 2]. The low level of accuracy and reliability of these clinicians’ diagnoses underscores the need for recalibrating their mental models to be compatible with evidence as documented in the IDSA guidelines for catheter-associated bacteriuria.

To address this need, we developed and validated an algorithm to enhance adoption of IDSA guidelines into diagnostic decisions for catheter-associated bacteriuria. A comprehensive version of the algorithm was created by mapping key decision points outlined in the CA-UTI and CA-ABU guidelines. Expert members of the IDSA guidelines panels provided content validation of the comprehensive algorithm with ratings along a 10-point scale for criterion, diagnostic, and feasibility standards. Cognitive interviews further established the face validity and usability of the comprehensive algorithm. From these results, we revised the algorithm. Finally, we established the reliability of the algorithm for accurately diagnosing cases as CA-UTI versus CA-ABU between expert and non-expert users and the inter-rater reliability of the algorithm between two non-expert users. High reliability between the clinical expert and each non-expert suggests improvement in diagnostic accuracy aided by the algorithm.

The results of the current study build on previous work that demonstrates physicians are more likely to treat bacteriuria with antibiotics (and therefore assume that the patient has a UTI) when patients have clinical cues that are consistent with prior diagnostic norms and practice (e.g., bacterial as opposed to fungal infection, higher white blood cell counts in the urine, positive urine nitrites, or a change in vague behaviors from baseline) [6, 12, 23]. The use of such guidelines-discordant cues leads to the inappropriate antimicrobial treatment of CA-ABU, and, as seen in our study, inconsistent (i.e., poor inter-rater reliability) and inaccurate diagnostic decision-making among clinicians. These guideline-discordant signs and symptoms are present within mental models that clinicians use to make diagnostic and treatment decisions [25].

Valid mental models built on prototypical cues (guideline-concordant signs or symptoms in Table 1) for CA-UTI can help to differentiate CA-UTI from CA-ABU among patients with catheter-associated bacteriuria. However, when these mental models are incorrectly constructed using cues that do not have high predictive validity (e.g., pyuria, and other guideline-discordant symptoms) or cannot help to differentiate the two subgroups (e.g., bacteriuria is present in both CA-UTI and CA-ABU), poor diagnostic accuracy and reliability will be the result.

The diagnosis and management of catheter-associated bacteriuria can be improved through the recalibration of clinicians’ mental models so that they are concordant with IDSA guidelines for differentiating CA-UTI from CA-ABU. This recalibration requires mindfulness of the guideline-discordant cues clinicians use when making diagnostic errors and substitution of guideline-concordant cues. Our study is consistent with prior evidence suggesting that simple methods, such as the use of checklists, algorithms, or protocols, combined with interventions such as audit and feedback, can enhance guideline adoption [2628].

The current study has several limitations. Participants in our sample are not representative of clinicians from all fields of medicine, but they do reflect a group of clinicians (physicians and allied health providers) who regularly manage catheter-associated bacteriuria in a typical inpatient setting. Another limitation is that the two non-experts were trained by an infectious disease expert who was very familiar with the guidelines content and how to use the algorithm. However, this training process is reflective of how algorithms are often implemented in real-world setting. Indeed, we are currently studying case-based audit and feedback as a focused training method for using our algorithm to decrease inappropriate use of antibiotics for CA-ABU [28]. Another limitation is the modest sample size in Phase 1; however, the number of case pairs in Phase 1 was sufficient for reliability testing. Finally, the 10 cases were chosen to be “difficult” in that each challenged clinical norms, thus likely magnifying the disagreement between the clinician’s diagnosis and the guidelines-concordant diagnoses. Case classification, albeit time-consuming, is a fundamental first step for any quality improvement project related to CA-UTI. The algorithm developed in this study may improve the efficiency and reliability of case classification.


During the diagnostic process, clinicians commonly compare patient’s symptoms to previously constructed mental models associating signs and symptoms to diseases. We have shown that use of improperly constructed (guideline discordant) mental models may result in diagnostic errors. Guidelines serve many of the same functions as mental models, in that they help identify data that are relevant to a particular diagnosis and exclude irrelevant data. The length and complexity of many guidelines limit their feasible dissemination and adoption in busy clinical settings. We have also shown that algorithms that simplify guidelines to better support decision-making in medical settings may help physicians identify and recalibrate inaccurate mental models, move toward more evidence-concordant diagnostic decisions, and reduce diagnostic errors.


  1. Hooton TM, Bradley SF, Cardenas DD: Diagnosis, prevention, and treatment of catheter-associated urinary tract infection in adults: 2009 international practice guidelines from the Infectious Disease Society of America. Clinical Infectious Disease. 2010, 50 (5): 625-663. 10.1086/650482.

    Article  Google Scholar 

  2. Nicolle LE, Bradley S, Colgan R, Rice JC, Schaeffer AJ, Hooton TM: Infectious Diseases Society of America guidelines for the diagnosis and treatment of asymptomatic bacteriuria in adults. Clin Infect Dis. 2005, 40 (5): 643-654. 10.1086/427507.

    Article  PubMed  Google Scholar 

  3. Lin K, Fajardo K: Screening for asymptomatic bacteriuria in adults: evidence for the U.S. Preventative Services Task Force reaffirmation recommendation statement. Ann Intern Med. 2008, 149: W20-4. 10.7326/0003-4819-149-1-200807010-00006.

    Article  PubMed  Google Scholar 

  4. Gross PA, Patel B: Reducing antibiotic overuse: A call for national performance measure for not treating asymptomatic bacteriuria. Clin Infect Dis. 2007, 45 (10): 1335-1337. 10.1086/522183.

    Article  PubMed  Google Scholar 

  5. Enterococcal bacteriuria is often overtreated and rarely leads to infectious complications. 2011, Boston: the 49th Annual Meeting of the Infectious Diseases Society of America

  6. Cope M, Cevallos ME, Cadle RM, Darouiche RO, Musher DM, Trautner BW: Inappropriate treatment of catheter-associated asymptomatic bacteriuria in a tertiary care hospital. Clin Infect Dis. 2009, 48: 1182-8. 10.1086/597403.

    Article  PubMed  Google Scholar 

  7. Dalen DM, Zvonar RK, Jessamine PG: An evaluation of the management of asymptomatic catheter-associated bacteriuria and candiduria at The Ottawa Hospital. Can J Infect Dis Med Microbiol. 2005, 16: 166-170.

    PubMed  PubMed Central  Google Scholar 

  8. Gandhi T, Flanders SA, Markovitz E, Saint S, Kaul DR: Importance of urinary tract infection to antibiotic use among hospitalized patients. Infect Control Hosp Epidemiol. 2009, 30: 193-5. 10.1086/593951.

    Article  PubMed  Google Scholar 

  9. Khawcharoenporn T, Vasoo S, Ward E, Singh K: Abnormal urinalysis finding triggered antibiotic prescription for asymptomatic bacteriuria in the ED. Am J Emerg Med. 2011, 29 (7): 828-830. 10.1016/j.ajem.2011.04.007.

    Article  PubMed  Google Scholar 

  10. Tambyah PA, Maki DG: The relationship between pyuria and infection in patients with indwelling urinary catheters: a prospective study of 761 patients. Arch Intern Med. 2000, 160: 673-7. 10.1001/archinte.160.5.673.

    CAS  PubMed  Google Scholar 

  11. Silver SA, Baille L, Simor AE: Positive urine cultures: A major cause of inappropriate antimicrobial use in hospitals?. Canadian Journal of Infectious Diseases and Medical Microbiology. 2009, 20 (4): 107-111.

    Article  PubMed  PubMed Central  Google Scholar 

  12. Walker S, McGreer A, Simor AE, Armstrong E, Loeb M: Why are antibiotics prescribed for asymptomatic bacteriuria in institutionalized elderly people? A qualitative study of physicians’ and nurses’ perceptions. CMAJ. 2000, 163 (3): 273-277.

    CAS  PubMed  PubMed Central  Google Scholar 

  13. Drinka PJ, Crnich CJ: Diagnostic accuracy of criteria for urinary tract infection in a cohort of nursing home residents. J Am Geriatr Soc. 2008, 56: 376-7.

    Article  PubMed  Google Scholar 

  14. Sundvall PD, Ulleryd P, Gunnarsson RK: urine culture doubtful in determining etiology of diffuse symptoms among elderly individuals: a cross-sectional study of 32 nursing homes. BMC Fam Pract. 2011, 12: 36-10.1186/1471-2296-12-36.

    Article  PubMed  PubMed Central  Google Scholar 

  15. Powers JH: Risk perception and inappropriate antimicrobial use: yes, it can hurt. Clin Infect Dis. 2009, 48: 1350-3. 10.1086/598184.

    Article  PubMed  Google Scholar 

  16. Kahneman D, Fredrick S: Representativeness revisited: Attribute substitution in intuitive judgment. Heuristics and biases: The psychology of intuitive judgment. Edited by: Gilovich T, Griffin D, Kahneman D. 2002, Cambridge, UK: Cambridge University Press, 49-81.

    Chapter  Google Scholar 

  17. Cavazos JM, Naik AD, Woofter A, Abraham NS: Barriers to physician adherence to nonsteroidal anti-inflammatory drug guidelines: A qualitative study. Aliment Pharmacol Ther. 2008, 28: 789-798. 10.1111/j.1365-2036.2008.03791.x.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  18. Grol R, Grimshaw J: From best evidence to best practice: effective implementation of change in patients’ care. Lancet. 2003, 362: 1225-1230. 10.1016/S0140-6736(03)14546-1.

    Article  PubMed  Google Scholar 

  19. Foy R, Eccles MP, Jamtvedt G, Young J, Grimshaw JM, Baker R: What do we know about how to do audit and feedback?. Pitfalls in applying evidence from a systematic review. BMC Health Services Research. 2005, 5: 50-

    Article  CAS  PubMed  Google Scholar 

  20. Sox HC, Blatt MA, Higgins MC, Marton KI: Medical Decision Making. 2007, Philadelphia: American College of Physicians Press, 2

    Google Scholar 

  21. Gigerenzer G, Gaissmaier W: Heuristic Decision Making. Annu Rev Psychol. 2011, 62: 451-482. 10.1146/annurev-psych-120709-145346.

    Article  PubMed  Google Scholar 

  22. Cognitive Interviewing: A How-to Guide. Short Course presented at the 1999 Meeting of the American Statistical Association. 1999, Durham, NC: Research Triangle Institute

    Google Scholar 

  23. Chant C, Dos Santos CC, Saccucci P, Smith OM, Marshall JC, Friedrich JO: Discordance between perception and treatment practices associated with intensive care unit-acquired bacteriuria and funguria: A Canadian physician survey. Crit Care Med. 2008, 36 (4): 1158-1166. 10.1097/CCM.0b013e3181692af9.

    Article  PubMed  Google Scholar 

  24. Landis JR, Koch GG: The measurement of observer agreement for categorical data. Biometrics. 1977, 33 (1): 159-174. 10.2307/2529310.

    Article  CAS  PubMed  Google Scholar 

  25. Wegwarth O, Gaissmaier W, Gigerenzer G: Smart heuristics for doctors and doctors-in-training: heuristics in medicine. Medical Education. 2009, 43: 721-728. 10.1111/j.1365-2923.2009.03359.x.

    Article  PubMed  Google Scholar 

  26. Cahill NE, Suurdt J, Ouellette-Kuntz H, Heyland DK: Understanding adherence to guidelines in the intensive care unit. Development of a comprehensive framework. J Parenter Enteral Nutr. 2010, 34 (6): 616-624. 10.1177/0148607110361904.

    Article  Google Scholar 

  27. Loeb M, Brazil K, Lohfield L, McGreer A, Simor A: Effect of a multifaceted intervention on number of antimicrobial prescriptions for suspected urinary tract infections in residents of nursing homes: Cluster randomized control trial. BMJ. 2005, 331 (7518): 669-10.1136/bmj.38602.586343.55.

    Article  PubMed  PubMed Central  Google Scholar 

  28. Trautner BW, Kelly PA, Petersen N, Hysong S, Kell H: A hospital-site controlled intervention using audit and feedback to implement guidelines concerning inappropriate treatment of catheter-associated asymptomatic bacteriuria. Implement Sci. 2011, 6 (1): 41-10.1186/1748-5908-6-41.

    Article  PubMed  PubMed Central  Google Scholar 

Pre-publication history

Download references


This work was supported by a grant from the Veterans Administration, Health Services Research and Development Program (IIR 09-104, Trautner PI) and with resources and use of facilities at the Houston VA HSR&D Center of Excellence (HFP90-020) at the Michael E. DeBakey VA. Dr. Trautner received support from a VA Rehabilitation Research and Development career development award (B4623). Dr. Payne receives support from the Office of Academic Affiliations, post-doctoral fellowship program in Health Services Research. Dr. Naik received additional support from a Doris Duke Charitable Foundation Clinical Scientist Development Award.

Author information

Authors and Affiliations


Corresponding author

Correspondence to Aanand D Naik.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

BT conceived of the study, participated in its design, drafted the manuscript and had final approval of the version submitted to the journal. RB participated in study design, obtained data and contributed to interpretation, and helped draft the manuscript. AA conducted statistical analysis and data interpretation, and helped draft the manuscript. SH participated in study design, contributed to interpretation of data, and revised manuscript critically. AG obtained data and contributed to interpretation. PK participated in study design, contributed to interpretation of data, and revised manuscript critically. VP contributed to interpretation of data and manuscript revision. AN participated in study conception and design, helped draft the manuscript, and had final approval of the version submitted to the journal. All authors read and approved the final manuscript.

Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

Authors’ original file for figure 1

Rights and permissions

Open Access This article is published under license to BioMed Central Ltd. This is an Open Access article is distributed under the terms of the Creative Commons Attribution License ( ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Trautner, B.W., Bhimani, R.D., Amspoker, A.B. et al. Development and validation of an algorithm to recalibrate mental models and reduce diagnostic errors associated with catheter-associated bacteriuria. BMC Med Inform Decis Mak 13, 48 (2013).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: