- Research article
- Open Access
- Published:

# A comparative analysis of multi-level computer-assisted decision making systems for traumatic injuries

*BMC Medical Informatics and Decision Making*
**volume 9**, Article number: 2 (2009)

## Abstract

### Background

This paper focuses on the creation of a predictive computer-assisted decision making system for traumatic injury using machine learning algorithms. Trauma experts must make several difficult decisions based on a large number of patient attributes, usually in a short period of time. The aim is to compare the existing machine learning methods available for medical informatics, and develop reliable, rule-based computer-assisted decision-making systems that provide recommendations for the course of treatment for new patients, based on previously seen cases in trauma databases. Datasets of traumatic brain injury (TBI) patients are used to train and test the decision making algorithm. The work is also applicable to patients with traumatic pelvic injuries.

### Methods

Decision-making rules are created by processing patterns discovered in the datasets, using machine learning techniques. More specifically, CART and C4.5 are used, as they provide grammatical expressions of knowledge extracted by applying logical operations to the available features. The resulting rule sets are tested against other machine learning methods, including AdaBoost and SVM. The rule creation algorithm is applied to multiple datasets, both with and without prior filtering to discover significant variables. This filtering is performed via logistic regression prior to the rule discovery process.

### Results

For survival prediction using all variables, CART outperformed the other machine learning methods. When using only significant variables, neural networks performed best. A reliable rule-base was generated using combined C4.5/CART. The average predictive rule performance was 82% when using all variables, and approximately 84% when using significant variables only. The average performance of the combined C4.5 and CART system using significant variables was 89.7% in predicting the exact outcome (home or rehabilitation), and 93.1% in predicting the ICU length of stay for airlifted TBI patients.

### Conclusion

This study creates an efficient computer-aided rule-based system that can be employed in decision making in TBI cases. The rule-bases apply methods that combine CART and C4.5 with logistic regression to improve rule performance and quality. For final outcome prediction for TBI cases, the resulting rule-bases outperform systems that utilize all available variables.

## Background

According to a 2001 National Vital Statistics Report [1], nearly 115,200 deaths occur each year due to traumatic injury, and many patients who survive suffer life-long disabilities. Among all causes of death and permanent disability, traumatic brain injury (TBI) is the most prevalent. Of the 29,000 children who are hospitalized each year with TBI, a significant percentage will suffer from neurological impairment [2]. It has also been reported that the traumatic brain injuries are the most expensive affliction in the United States, with an estimated cost of $224 billion [3].

Computer-aided systems can significantly improve trauma decision making and resource allocation. Since trauma injuries have specific causes, all with established methods of treatment, fatal complications and long-term disabilities can be reduced by making less subjective and more accurate decisions in trauma units [4]. In addition, it has been suggested that an inclusive trauma system with an emphasis on computer-aided resource utilization and decision making may significantly reduce the cost of trauma care [1].

Since the treatment of traumatic brain injuries is extremely time-sensitive, optimal and prompt decisions during the course of treatment can increase the likelihood of patient survival [5, 6]. It is also believed that the predicted length of stay in the ICU is an important factor when deciding on the patient transport method (i.e. ambulance or helicopter), as more critical patients are expected to spend more time in the ICU, and these stand to benefit the most from helicopter transport. Studies have emphasized the critical impact of helicopter transport on trauma mortality rates, since the speed of ambulance transport is limited by road and weather conditions, and may also be constrained by traffic congestion. However, it is difficult to compare ground and helicopter transportation and the corresponding care provided to the patients [7]. Cunningham [8] attempts a comparison based on the outcome of the treatment given to trauma patients. Based on his study, patients in critical condition are more likely to survive if transported via helicopter. However, the high cost of helicopter transport remains a major problem [9, 10]. In recent studies, Gearhart evaluated the cost-effectiveness of helicopter for trauma patients and suggested that on average the helicopter transport cost is about $2,214 per patient, and $15,883 for each additional survivor [11]. Eventually, the cost is almost $61,000 per surviving trauma patient. Eckstein [12] states that 33% of patients who are transported by helicopter are discharged home from the emergency department [12], rather than being sent to ICU. This indicates that a significant number of trauma patients transported by helicopter actually have relatively minor injuries. This emphasizes the necessity of a comprehensive transport policy based on patient condition and predicted outcome.

Several computer-assisted systems already exist for decision-making in trauma medicine. The majority of these systems [13, 14] are designed to perform a statistical survey of similar cases in trauma databases, based only on patient demographics. As such, they may not be sufficiently accurate and/or specific for practical implementation. Other medical decision making systems employ the predictive capabilities of artificial neural networks [15–17]; however, due to the 'black box' nature of these systems, the reasoning behind the predictions and recommended decisions is obscured. Currently, none of these existing systems are in widespread use in trauma centers. There are three main reasons: the use of non-transparent methods, such as neural networks; the lack of a comprehensive database integrating all relevant available patient information for specific prediction processes; and poor performance due to the exclusion of relevant attributes and the inclusion of those irrelevant to the current task, resulting in rules that are too complicated to be clinically meaningful.

Several machine learning algorithms are commonly applied to medical applications. These include support vector machines (SVM), and decision tree algorithms such as Classification and Regression Trees (CART) and C4.5. Boosting is also employed for improving classification accuracy. However, despite the relatively successful performance of these algorithms in medical applications, they have limited success in separating and identifying important variables in applications where there are a large number of available attributes. This suggests that combining machine learning with a method to identify the most uncorrelated set of attributes can increase our understanding of the patterns in medical data and thus create more reliable rules. The literature of biomedical informatics reinforces the benefits of this approach. Andrews et al. [18] use decision tree (DT) and logistic regression (LR) methods to identify the commonalities and differences in medical database variables. Kuhnert [19] emphasises that non-parametric methods, such as CART and multivariate adaptive regression splines, can provide more informative models. Signorini et al. [20] design a simple model containing variables such as age and GCS, but the small number of attributes may limit the reliability of the generated rules. Guo [21] finds that CART is more effective when combined with the logistic model, and Hasford [22] compares CART and logistic regression, and finds that CART is more successful in outcome prediction than logistic regression alone.

Therefore, a possible approach to create accurate and reliable rules for decision making is to combine machine learning and statistical techniques [23, 24]. This paper analyzes the performance of several combinations of machine learning algorithms and logistic regression, specifically in the extraction of significant variables and the generation of reliable predictions. Though a transparent rule-based system is preferable, other methods (such as neural networks) are also tested in the interest of comparision. A computational model is developed to predict final outcome (home or rehab and alive or dead) and ICU length of stay. In addition, we identify the factors and attributes that most affect decision making in the treatment of traumatic injury.

Our hypotheses are as follows:

1. We hypothesize that a rule-based system, attractive to physicians as the reasoning behind the rules is transparent and easy to understand, can be as accurate as "black-box" methods such as neural networks and SVM.

2. We hypothesize that when trained correctly, a computer-aided decision making system can provide clinically useful rules with a high degree of accuracy.

3. Studies mentioned earlier have examined which variables are most significant in the recommendation/prediction making process. We hypothesize that airway status, age, and pre-existing conditions such as myocardial infarction and coagulopathy are significant variables.

## Methods

Rules are created by processing patterns discovered in the traumatic brain injury (TBI) datasets. More specifically, they are generated by analyzing the logical and grammatical relationships among the input features and the resulting outcomes. Rules are formally defined as grammatical expressions of knowledge extracted using specific logical operations on the available features [6].

CART and C4.5 are among the most popular algorithms for creating reliable rules, but they are limited in their ability to identify the most significant variables. We therefore perform statistical analysis using logistic regression, which is typically effective in discovering statistically significant regression coefficients [24]. Although stepwise regression is designed to find significant variables, it may not perform well with CART when dealing with small scale datasets [25]. Therefore, in this paper, logistic regression with direct maximum likelihood estimation (Direct MLE) is used.

### Dataset

Three different datasets are used in the study: on-site, off-site, and helicopter. The on-site dataset contains data captured at the site of the accident; the off-site dataset is formed at the hospital after patients are admitted; and the helicopter dataset consists of the records for patients who are transported to hospital by helicopter. The on and off-site datasets are used to predict patient survival (dead/alive) and final outcome (home/rehab), and the helicopter dataset is used to predict ICU length of stay, which is a measure used in estimating the need for helicopter transportation. The datasets are provided to us by the Carolinas Healthcare System (CHS) and the National Trauma Data Bank (NTDB).

#### On-site dataset

When making decisions based on the variables available at the accident scene, one has to consider the unavailability of important factors such as pre-existing conditions (comorbidities). Decisions must therefore be made without knowledge of these factors. Some physiological measurements are also excluded because they are only collected after arrival at the hospital. Table 1 presents the variables collected for this dataset, which consist of four categorical and six numerical attributes.

#### Off-site dataset

The off-site dataset contains information on comorbidities and complications, and includes all variables. A total of 1589 cases are included in the database: 588 fatal and 1001 non-fatal. The inputs include both categorical and numerical attributes. The predicted outcomes are defined as the patients' survival, i.e. alive or dead, and the exact outcome for surviving patients, i.e. rehab or home.

Table 2 presents the variables for our dataset. Among the categorical variables, "prexcomor" represents any comorbidities that may negatively impact a patient's ability to recover from the injury and any complications. Other terms are defined in the table description.

#### Helicopter dataset

This dataset is formed based on the records of patients who were transported to hospital by helicopter. The variables are age, gender, blood pressure, cheifcomp (the type of injury), airway (the type of device used to assist patients with breathing), prefluids (the amount of blood provided to the patients), GCS, heart rate, respiration rate, ISS-Head&Neck, and ISS. Age, blood pressure, GCS, heart rate, ISS-Head&Neck, ISS, and respiration rate are classified as numerical variables. The final outcome is the number of days spent in ICU, as this is considered the most informative measure when deciding the means of transport to hospital. In our dataset, ICU stay ranges between 0 and 49 days. The use of a relatively small dataset with so many outcomes may result in a complex model that is hard to explain and understand. Inspired by Pfahringer [26], the dataset is classified into two groups. The non-severe group contains patients who stayed in the ICU less than 2 days (ICU stay ≤ 2 days). The severe group consists of patients who stayed in the ICU more than 2 days (ICU stay ≥ 3 days). This threshold was chosen based on discussion with trauma experts. In total, the dataset contains 497 cases: 196 severe and 301 non-severe [10]. Table 3 describes the helicopter dataset in more detail.

### Learning algorithms

It is known that the patterns observed in trauma cases are often extremely complicated; that is, the treatment outcomes for two apparently similar patients may turn out to be significantly different. Linear methods have proven insufficient even in the analysis of patterns as simple as the "exclusive-or" function. Because these limitations are inherited by linear regression methods, the use of non-linear techniques for computer-aided trauma systems has been broadly encouraged [27]. Neural networks are a common choice; however, they are not transparent, since the knowledge learned from the training examples is hidden within the structure and weights of the networks [28]. While there are existing methods that can extract approximate rules to represent this hidden knowledge, they cannot truly represent the trained networks [6]. Support Vector Machines (SVM's) and AdaBoost share the same problem: the knowledge used in the decision making process is not visible to humans, a requirement that is extremely important in medical applications. Rule-based methods such as CART and C4.5 provide completely transparent computational decision making systems while still utilizing some nonlinear capabilities. Considering the importance of decision transparency in medical informatics, we use CART and C4.5 as the main algorithms for rule extraction.

#### Classification and Regression Tree (CART)

CART, designed by L. Breiman [29], applies information-theoretic concepts to create a decision tree. This allows for the capture of rather complex patterns in data, and their expression in the form of transparent grammatical rules [30]. CART's nonlinear extensions are still widely used in data mining and machine learning, due to the algorithm's efficiency in dealing with multiple data types [31] and missing data. In the latter case, CART simply uses a substitution value, defined as a pattern similar to the best split value in the node [29]. In addition, CART supports an exhaustive search of all variables and split values to find the optimal splitting rules for each node. The splitting stops at the pure node containing fewest examples.

#### C4.5

C4.5 [15, 32, 33] extends Quinlan's basic ID3 decision tree algorithm [34]. It is more successful in avoiding overfitting, is able to handle continuous variables, and is more computationally efficient. To generate rules, C4.5 uses a divide-and-conquer algorithm to split training data into disjoint regions of the variable space, according to pre-assigned target labels [9]. At each step, C4.5 splits on the best attribute according to the gain criterion. This criterion is based on entropy, i.e. the randomness of the class distribution in the dataset. The criterion is the greatest difference in entropy of the class probability distribution of the current subset S and the subsets generated by the split.

where *p*(*k*
_{
i
}, *S*) is the relative frequency of examples in S that belong to class *k*
_{
i
}. The best split is the one that most reduces this value. The output of the algorithm is a decision tree, which can be easily represented as a set of symbolic IF-THEN rules.

#### Adaptive Boost (AdaBoost)

AdaBoost, introduced by Freund and Schapire [35], is an algorithm that constructs a robust classifier as a linear combination of weak classifiers. Adaboost repeatedly calls a given weak learning algorithm in a set of rounds *t* = 1, ..., *T*. A distribution of weights is maintained over the training set, such that *D*
_{
t
}(*k*) is the distribution's weight for training example k on round t. The aim of the weak learner is to find a good weak hypothesis *h*
_{
t
}: *X* → {-1, +1} for the distribution *D*
_{
t
}, where goodness is measured by the error of the hypothesis with respect to *D*
_{
t
}. Then *D*
_{
t
}is updated such that incorrectly classified examples have their weights increased, forcing the weak classifier to concentrate on the more difficult training examples. Correspondingly, correctly classified examples are given less weight. Adaboost selects some parameter *α*
_{
t
}to denote the importance of *h*
_{
t
}, and after all rounds are complete, the final hypothesis *H* is a weighted majority vote of all *T* weak hypotheses. It has been shown that as with other boosting algorithms, if each weak hypothesis is at least slightly better than random, then the training error falls at an exponential rate. However, Adaboost is also able to adapt to the error rates of individual weak hypotheses, so each subsequent classifier is adjusted in favor of examples mislabelled by previous classifiers [36].

#### Support Vector Machine (SVM)

SVMs [37] are supervised learning methods used primarily for classification. An SVM treats its input data as two sets of vectors in n-dimensional space: positive and negative examples. In this space, it constructs an optimal hyperplane that preserves the maximum distance between the two sets [38]. Since SVM is able to handle large feature spaces, it has been successfully used in solving many real world problems such as text categorization, image classification, protein analysis, cancer data classification, and hand-writing recognition [39]. Consider a set of N labelled training examples *D* = (*x*
_{1}, *y*
_{1}),..., (*x*
_{
n
}, *y*
_{
n
}) with *y*
_{
i
}∈ {+1, -1} and *x* ∈ *R*
^{d}, where *d* is the dimensionality of the input. Let *φ*: *R*
^{d}→ *F* be the mapping function from the input space to the feature space. If the two classes are linearly separable, the SVM algorithm finds a hyperplane (*w*, *b*) that maximizes the margin

where *b* is a real number (bias term) and *w* and *F* have the same dimensionality. For an unknown input vector *x*
_{
j
}, classification means finding:

*f*(*x*
_{
j
}) = *sgn*(*y*
_{
i
}<*w*, *φ*(*x*
_{
i
}) > -*b*)

It can be shown that this minimum occurs when *w* = Σ_{
i
}
*α*
_{
i
}
*γ*
_{
i
}
*φ*(*x*
_{
i
}), where *α*
_{
i
}is a positive real number that represents the strength of training point *x*
_{
i
}in the final classification decision. The subset of points where ai is non-zero consists of the points closest to the hyperplane, and these are the support vectors. Since SVM is able to handle large feature spaces, it is frequently used in many real world problems even though it is computationally expensive [39].

#### Neural networks

A neural network processes training examples individually, and learns by comparing its classification of the input (which is initially largely arbitrary) with the given correct classification. In particular, Radial Basis Function (RBF) networks are well suited to solving pattern classification problems due to their simple topological structure and their capability for faster learning. A standard RBF network is a supervised feed-forward back propagation neural network, consisting of an input layer, a hidden layer and an output layer. One of the most common basis functions for the hidden layer is the family of Gaussian functions whose outputs are inversely proportional to the distance from the center of the neuron. Given a finite set of training data {(*x*
_{
j
}, *y*
_{
j
})|*j* = 1, . . ., *m*}, and the center vector of basis function *c*
_{
i
}, the equation for a simple output is:

where *N* is the number of neurons in the hidden layer, and *α*
_{
i
}are the weights minimizing least square between real output and approximate output. Typically a Gaussian activation function producing a radial function of the distance between each hidden unit weight vector and each pattern vector is used as a basis function:

where *σ* indicates the neuron radius [40, 41]. RBFs utilize the distance in feature space to calculate the weight for each neuron.

### Pre-processing

The datasets contain nominal categorical variables, such as gender and complication type. Gender is replaced by a binary variable (0 for male, 1 for female). Every nominal value is dummy-coded (Yes/No to 1/0) and treated as an individual attribute. Ten fold cross-validation is used to measure the generalization quality and scalability of the rules. Each dataset is divided into ten mutually exclusive subsets [42], and in each stage nine are used for training and one is used for testing. Ten different trees are therefore formed for each dataset.

### Rule performance metrics

Once a variety of rules are generated, the performance of each rule is measured as the probability of correct prediction. Assume that *D* is a dataset including the instance (*x*
_{
i
}, *y*
_{
i
}), where *y*
_{
i
}is the real survival outcome. Let *D*
_{
r
}be the training set, and a subset *D*
_{
t
}∈ (*D*\*D*
_{
r
}) be used for testing. The performance of the rule is calculated as:

*acc*
_{
R
}= *prob*(*y*
_{
i
}= *y*
^{R}|(*x*
_{
i
}, *y*
_{
i
}) ∈ *D*
_{
t
})

where *y*
^{R}is the outcome produced by induction, i.e. the expected classification. The number of positive matches in the testing set is used as a measure of rule accuracy. Rule accuracy can also be estimated as follows:

where TP is the number of true positives, TN is the number of true negatives, FP is the number of false positives, and FN is the number of false negatives. Sensitivity and specificity are then used to assess the quality of the rules. These measures are useful, as they calculate the probabilities of false positives and false negatives separately; one may be significantly higher than the other, and this can be obscured in a single average error measure. The formulae for these measures are shown below.

where TP, TN, FP, and FN are defined as before. In this application, high sensitivity is more important than high specificity. When patient lives are at stake – for example, in the choice of transportation – false positives are preferable to false negatives, even if they incur greater financial cost.

### Improving rule quality

Once the most accurate rules have been extracted, direct maximum likelihood estimation with logistic regression is used to improve rule quality. The logistic function calculates the expected probability of a dichotomy as:

where *X*
_{
i
}are variables with numeric values, *Y* is the outcome (dichotomous; 0 or 1, e.g. Alive/Dead), and the *β*'s are the regression coefficients that quantify the contributions of the numeric variables to the overall probability [22].

Logistic regression provides knowledge of the relationships and strengths among the multiple independent variables and the response variable. It does not assume any distribution on the independent variables; they do not have to be normally distributed, linearly related or of equal variance within each group. The most important interpretation from logistic regression is the odds ratio, which measures the strength of the partial relationship between an individual predictor and the outcome event [43].

The advantage of using the logit scale for interpretation is that the relationship between the logit and the predictors is linear. To check this linearity assumption, we used scatter plot and residual analysis. The results showed linear relationships for all variables, though some were weaker than others. In the interests of brevity, in this paper we present the results for only two variables: Head AIS and Age. First, we present the scatter plot between the logit and its predictor, and then the residual plot between them using regression analysis. If the linearity assumption is satisfied, we would expect the residuals to vary randomly – i.e. they would not demonstrate any pattern. If the residual plot appears to form a curve, there may be a nonlinear relationship in the variable. This analysis was performed using statistical analysis software (SAS). Figure 1 and Figure 2 present the scatter plots and residual plots using Age and Head AIS as the predictors for patient survival.

If the plots of the residuals versus the predictors do show curvature, a quadratic term should be tested for statistical significance for suggesting better model. If the coefficient for this quadratic term is significant, the quadratic term should be included. Even though our model does not show any strong curvature, we test the Head AIS variable using a quadratic term, to validate our results. The model is as follows:

*logit* = *α* + *βx* + *γx*
^{2}

where *α* is an intercept term, *β* is a parameter of the predictor, and *γ* is a parameter of squared predictor. For the Head AIS variable, the estimate of *β* is -0.1820 (p value = 0.0015), and the estimate of *γ* is -0.0124 (p value = 0.2058). These p values indicate that Head AIS does not require a quadratic term; therefore, there is a linear relationship between the logit and its predictor.

To test the significance of the individual variables, we compare a reduced model that drops one of the independent variables with a full model using log likelihood test. The likelihood ratio test itself does not tell us if any particular independent variables are more important than others. However, by estimating the maximum likelihood, we can analyze the difference between results for the full model and results for a nested reduced model which drops one of the independent variables. A non-significant difference indicates no effect on performance of the model, hence we can justify dropping the given variable. We call this directed MLE.

The test takes the ratio of the maximized value of the likelihood function for the full model (*L*
_{1}) over the maximized value of the likelihood function for the simpler model (*L*
_{0}). The resulting likelihood ratio is given by:

If the chi-square value for this test is significant, the variable is considered to be a significant predictor. Following these tests, only the significant variables (p value <= .05) are selected.

Note that forward and stepwise model selections are also available to discover the significance of individual attributes [19, 25]. In the literature of statistical regression, the stepwise method is commonly used to find the best subset of variables for outcome prediction, considering all possible combinations of variables. However, the stepwise approach may not guarantee that the most significant variables are selected due to the repetition of insertion and deletion. For example, age may not be selected as important variable; however, physicians may believe that patient age is important in deciding treatment. Therefore, we prefer to use directed MLE for our medical application. Our other justification for using MLE is empirical; in our previous study [10], we found that the direct MLE method has slightly higher accuracy in finding significant variables than stepwise and forward model selection. A statistical analysis tool, in this case SAS, is used to calculate the significance of individual attributes.

### Constructing reliable rules

As mentioned previously, SVM and neural networks do not directly produce grammatical rules; therefore, only CART and C4.5 are considered for rule extraction. Those variables identified as significant are used as input variables to CART and C4.5. Also, if a rule is created only to accommodate one or two examples, it may be too specific to be applied to the entire population. Consequently, only the rules with both high accuracy and a sufficiently large number of supporting examples are used to form the rule base. Note that SVM, Neural Networks and AdaBoost are still tested in the interests of performance comparision, even though they do not generate rules. These algorithms are in widespread use, and comparing them to the rule based CART and C4.5 algorithms tests and validates the accuracy and stability of the rule-based system.

## Results

The average accuracy of survival prediction without any knowledge of pre-existing conditions is 73.9%, rising to 75.8% when this knowledge is included. The off-site dataset is therefore used for further prediction tests, as it contains records of pre-existing conditions. We discovered that knowledge of these conditions appears at the highest level of the tree when using CART and C4.5, indicating their potential importance in the decision-making process. In particular, coagulopathy (bleeding disorder), which can result in severe haemorrhage, may be among the most important factors to consider in patients with TBI.

Due to the transparent nature of the rule-based system used in this study, the generated rules can not only help trauma experts predict the likelihood of survival, but also provide the reasoning behind these predictions in order to help physicians better allocate their resources.

Since the total number of examples used for training is rather small, initially only rules with at least 85% prediction accuracy on the testing sets are included in the rule base. This threshold was chosen following discussion with trauma experts. However, we also incorporate rules with accuracy between 75% and 85%. There are two reasons for this. Firstly, the accuracy of a rule may be low due to the lack of of a truly complete database, rather than a flaw in the rule itself. Secondly, even though a rule may have low accuracy, it might include knowledge of hidden relationships between variables. For example, most trauma experts consulted believed that a patient with an ISS score over 25 would have little chance of survival. However, the survival probability might be higher for a patient with a high ISS score, but lower head and thorax AIS score, provided appropriate and prompt treatment is provided. Therefore, we will use those rules with accuracy between 75% and 85% as additional "supporting rules" in suggesting possible treatment. This issue is addressed further in the discussion section.

### Significant variable selection

In order to improve the rule quality and accuracy, it is essential that we identify the key variables in the dataset. In addition, shorter rules that are based on fewer, more significant variables are more clinically useful for physicians. Direct MLE with logistic regression is used to accurately extract these key variables from our helicopter and off-site datasets; the results for the off-site dataset are presented in Table 4. It can be seen that nine important variables are identified. Using standard deviations, Wald chi-squares are computed on each variable and the odd ratios are interpreted as showing a strong relationship between the outcome and the independent variables. Table 5 presents the significant variables extracted from the helicopter dataset. Only five of the eleven original variables are identified as significant.

In this study the scale of the data is small and several variables are unknown, so participating physicians assisted in identifying significant variables. These physicians selected age, GCS, blood pressure, pulse rate, respiration rate, and airway as important factors.

### Measuring performance

The prediction results of five different machine learning methods are compared in Table 6. The performance for all algorithms is clearly superior when only significant variables are used. In addition, using only the most significant variables is shown to result in a more balanced testing-training performance. Discussion with physicians revealed that generated recommendations and predictions must be transparent in their reasoning; our system therefore uses CART and C4.5 to predict patient survival. If physicians understand the reasoning behind decisions and it follows their own, their confidence in the system may be increased. If the system's reasoning is clinically meaningless, they can choose to disregard the recommendation; however, if the reasoning has some clinical merit, this may alert them to previously hidden factors affecting patient outcome.

Table 7 presents the performance accuracy in outcome prediction (rehabilitation or home) for the off-site dataset, and prediction of ICU days for the helicopter dataset. In both cases, only the significant variables are used. No attempt is made to use all available variables, since the survival prediction test has already confirmed the improved performance when using only significant variables.

We also generate Receiver Operating Characteristic (ROC) curves – plots of the true positive rate (sensitivity) versus the false positive rate (1-specificity) – in order to evaluate the model performance. First, we perform ROC analysis on the patient surivival prediction results. Table 8 compares the area under the curve (AUC) for the ROC curves generated using all available variables and significant-only variables. The table shows that results are improved when using only significant variables in the model. Therefore, when dealing with the helicopter dataset, we only perform ROC analysis on the significant-variable-only model. The results are presented in Table 9. Notice that there is no large difference in ROC analysis results among the various machine learning methods. However, when the dataset is small – such as our data used for ICU days prediction – logistic regression outperforms the other methods. Figure 3 and Figure 4 present sample ROC plots for logistic regression using only significant variables for survival and ICU days prediction respectively.

### Constructed database using CART and C4.5

Numerous rules were generated with the CART and C4.5 rule extraction algorithm. Following discussion with trauma experts, we identified the robust rules as those with over 85% accuracy. For survival prediction, the average rule accuracy using all available variables is 82%, and 83.9% when using only the most significant variables.

Table 10 presents the most reliable generated rules for survival prediction (> 85% accuracy); Table 11 contains survival rules with accuracy between 75% and 85%. Similarly, Table 12 presents the most reliable generated rules for outcome prediction (> 85% accuracy), and Table 13 contains outcome rules with accuracy between 75% and 85%. Finally, Table 14 presents the most reliable generated rules for ICU days prediction (> 85% accuracy), and Table 15 contains ICU days rules with accuracy between 75% and 85%. Note that the rules with accuracy between 75% and 85% may not be sufficiently reliable, yet may contain useful pattern information, as described in the discussion section.

## Discussion

We developed a computer-aided rule-base using significant variables selected via logistic regression, and showed that the approximations of the variables help increase rule quality. Our intent is to extract and formulate medical diagnostic knowledge into an appropriate set of transparent decision rules that can be used in a computer-assisted decision making system. The proposed method extracts the most significant variables using logistic regression with direct maximization likelihood estimation. By comparing the performances using five machine learning algorithms – AdaBoost, C4.5, CART, RBF neural network, and SVM – using all available variables and significant variables only, we found that using only the most significant variables provides a considerable improvement in performance. All five methods show improvement across all-available and significant-variables-only, indicating that our proposed selection method is robust and efficient.

The performance of individual rules was measured; reliable rules were identified as those with accuracy above 85%. In addition, all rules we selected were considered reliable if the number of cases in the dataset matching the rule was higher than a specified threshold. Rule sensitivity and specificity were also measured, and the average sensitivity and specificity for the three outcome pairs (alive/dead, home/read, severe/non-severe) are 87.4% and 88.4% respectively. This indicates that our method performs well. Some additional improvements may be needed to improve rule quality. In particular, large and well balanced datasets across all outcome classes could improve overall quality, as well as sensitivity and specificity. Full sensitivity and specificity results for the datasets are presented in Table 16.

One important issue in rule selection is how to deal with rules with accuracy below 85%. When using only the over-85% rules, some medical knowledge in the database might have been ignored. The accuracy of a rule may be low due to the lack of "database completeness", rather than a flaw in the rule itself. Therefore, rules with less than 85% accuracy cannot be completely removed from the rule based system. We will instead use those rules as additional "supporting rules" in suggesting possible treatment. For example, according to trauma experts, patients with a high ISS score (> 25) are least likely to survive. However, we found some rules with surprising implications. For instance, one of these "counterintuitive" rules pointed to the fact that there are 52 alive cases (3.3%) with ISS high scores (38). Of these 52 patients, 33 (63.5%) have high AIS head scores (≥ 4), and 38 patients (73%) are male. Considering the above conditions, surviving patients have lower thorax (average score = 2.61) and lower abdomen AIS scores (average score = 1.03) than fatal cases. These fatal cases typically have a higher head AIS score (average score = 5.08) than surviving patients (average head score = 3.90). In addition, we found that none of the surviving patients have complications such as coagulopathy, and only a few had a pre-existing disease (in particular, Insulin Dependency and Myocardial Infarction).

While only Acute Respiratory Distress Syndrome (ARDS) is usually considered an impact factor in predicted survival, according to the created rules, pre-existing conditions, Acute Respiratory Distress Syndrome (ARDS), Insulin Dependency, Myocardial Infarction, and Coagulopathy all have significant impact. Also, airway status (need/not need) was identified as a primary factor in predicting the number of ICU days for patients transported via helicopter.

Note that for ICU length of stay prediction, 74.6% of patients stayed at in ICU less than 2 days. Only 25.4% of patients stayed more than 2 days, and only 2.9% of those were in ICU for more than 20 days. This reinforces Eckhart's point that many patients are transported via helicopter unnecessarily. Therefore, the use of accurate ICU days prediction rules may help improve the efficiency of helicopter transport, considering cost effectiveness as well as the treatment of patients in critical condition.

## Conclusion

The results in this paper provide a framework to improve the physicians' diagnostic accuracy with the aid of machine learning algorithm. The resulting system is effective in predicting patient survival, and rehab/home outcome. A method has been introduced that creates a variety of reliable rules that make sense to physicians by combining CART and C4.5 and using only significant variables extracted via logistic regression. The resulting computer-aided decision-making system has significant benefits, both in providing rule-based recommendations and in enabling optimal resource utilization. This may ultimately assist physicians in providing the best possible care to their patients. The diagnosis of future patients may also be improved by analyzing all possible rules associated with their symptoms.

The system will be tested at all 17 hospitals of the Carolinas Healthcare System (CHS). Software that provides the computer-aided decision making system will be optimized and made available to the academic community as a web-based application, as well as a software tool on portable personal computing devices. Feedback from every hospital will then be considered and used to validate and improve the system.

## References

Anderson RN, Minino AM, Fingerhut LA, Warner M, Heinen MA, Eds: Deaths: Injuries, 2001. National Vital Statistics Reports. 2001, 52 (21): 1-87.

Centres for Disease Control and Prevention: Facts about Concussion and Brain Injury and Where to Get Help. Atlanta, GA. 2004

The Coalition for American Trauma. [http://www.traumacoalition.org]

Expert Working Group: Traumatic Brain Injury in the United States: Assessing Outcomes in Children. Atlanta, GA. 2000

Fabian TC, Patton JH, Croce MA, Minard G, Kudsk KA, Pritchard FE: Blunt carotid injury: Importance of early diagnosis and anticoagulant therapy. Ann Surg. 1996, 223 (5): 513-525.

Jagielska I: Linguistic rule extraction from neural networks for descriptive data mining. Proc. 2nd Int'l Conf. Knowledge-Based Intelligent Electronic Systems: 21–23 Apr 1998, Adelaide. 1998, 89-92.

Arfken CL, Shapiro MJ, Bessey PQ, Littenberg B: Effectiveness of Helicopter versus Ground Ambulance Services for Interfacility Transport. Journal of Trauma-Injury Infection & Critical Care. 1998, 45 (4): 785-790.

Cunningham P, Rutledge R, Baker CC, Clancy TV: A comparison of the association of helicopter and ground ambulance transport with the outcome of injury in trauma patients transported from the scene. J Trauma. 1997, 43 (26): 940-946.

Ruggieri S: Efficient C4.5. IEEE Trans on Knowl and Data Eng. 2002, 14 (2): 438-444.

Ji SY, Huynh T, Najarian K: An intelligent method for computer-aided trauma decision making system. ACM-SE 45: Proceedings of the 45th annual southeast regional conference. 2007, 198-202.

Gearhart PA, Wuerz R, Localio AR: Cost-effectiveness analysis of helicopter EMS for trauma patients. Ann Emerg Med. 1997, 30 (4): 500-506.

Eckstein M, Jantos T, Kelly N, Cardillo A: Helicopter Transport of Pediatric Trauma Patients in an Urban Emergency Medical Services System: A Critical Analysis. J Trauma. 2002, 53 (2): 340-344.

Haug PJ, Gardner RM, Tate KE: Decision support in medicine: examples from the HELP system. Comput Biomed Res. 1994, 27 (5): 396-418.

Fitzmaurice JM, Adams K, Eisenberg JM: Three Decades of Research on Computer Applications in Health Care: Medical Informatics Support at the Agency for Healthcare Research and Quality. Journal of the American Medical Informatics Association. 2002, 9 (2): 144-160.

Quinlan J: Improved use of continuous attributes in C4.5. Journal of Artificial Intelligence Research. 1996, 4: 77-90.

Clarke JR, Hayward CZ, Santora TA, Wagner DK, Webber BL: Computer-generated trauma management plans: comparison with actual care. World J Surg. 2002, 26 (5): 536-538.

Najarian K, Darvish A: Neural Networks and Their Applications in Biomedical Engineering. Wiley Encyclopaedia of Biomedical Engineering. Edited by: Akay M. 2006, Hoboken, NJ: John Wiley & Sons

Andrews PJ, Sleeman DH, Statham PF, McQuatt A, Corruble V, Jones PA, Howells TP, Macmillan CS: Predicting recovery in patients suffering from traumatic brain injury by using admission variables and physiological data: a comparison between decision tree analysis and logistic regression. J Neurosurg. 2003, 97 (2): 440-442.

Kuhnert PM, Do K, McClure R: Combining non-parametric models with logistic regression: an application to motor vehicle injury data. Comput Stat Data Anal. 2000, 34 (3): 371-386.

Signorini DF, Andrews PJD, Jones PA, Wardlaw JM, Miller JD: Predicting survival using simple clinical variables: a case study in traumatic brain injury. J Neurol Neurosurg Psychiatry. 1999, 66: 20-25.

Guo HM, Shyu YI, Chang HK: Combining logistic regression with classification and regression tree to predict quality of care in a home health nursing data set. Stud Health Technology Inform. 2006, 122: 891-

Hasford J, Ansari H, Lehmann K: CART and logistic regression analyses of risk factors for first dose hypotension by an ACE-inhibitor. Therapie. 1993, 48 (5): 479-482.

Kelemen A, Liang Y, Franklin S: A Comparative Study of Different Machine Learning Approaches for Decision Making. Advances in Simulation, Computational Methods and and Soft Computing. Edited by: Mastorakis E. 2002, Piraeus, Greece: WSEAS Press

Snedecor GW, Cochran WG: Statistical Methods. 1989, Iowa State University Press, 8

Long WJ, Griffth JL, Selker HP, D'Agostino RB: A comparison of logistic regression to decision-tree induction in a medical domain. Comput Biomed Res. 1993, 26: 74-97.

Pfahringer B: Compression-Based Discretization of Continuous Attributes. Proc. 12th Int'l Conf. Machine Learning: Tahoe City. 1995, 456-463.

Mac Namee B, Cunningham P, Byrne S, Corrigan OI: The problem of bias in training data in regression problems in medical decision support. AI Med. 2002, 24: 51-70.

Tu JV: Advantages and disadvantages of using artificial Neural networks versus Logistic regression for predicting medical outcome. J Clin Epidemiol. 1996, 49 (11): 1225-1231.

Breiman L: Classification and Regression Trees. 1993, Boca Raton: Chapman & Hall

Loh WY, Vanichsetakul N: Tree-Structured Classification via Generalized Discriminant Analysis. Journal of the American Statistical Association. 1988, 83 (403): 715-725.

Fu CY: Combining loglinear model with classification and regression tree (CART): an application to birth data. Computational Statistics & Data Analysis. 2004, 45 (4): 865-874.

Quinlan JR: Generating production rules from decision trees. Proc. 10th Int'l Joint Conf. Artificial Intelligence (IJCAI-87): 1987, Milan. 1987, 997-1003.

Quinlan JR: C4.5: Programs for Machine Learning. Morgan Kaufmann. 1993

Quinlan JR: Learning efficient classification procedures and their application to chess end games. Machine Learning, An Artificial Intelligence Approach. Edited by: Michalski RS, Carbonell JG, Mitchell TM. 1983, Palo Alto, CA: Tioga Publishing Company, 463-482.

Freund Y, Schapire R: A short introduction to boosting. J Japan Soc for Artif Intel. 1999, 14 (5): 771-780.

Fan W, Stolfo SJ, Zhang J: The application of AdaBoost for distributed, scalable and on-line learning. KDD '99: Proceedings of the fifth ACM SIGKDD international conference on Knowledge discovery and data mining. 1999, 362-366.

Vapnik V: . Statistical Learning Theory. 1998, New York: Wiley

Joachims T: Text categorization with support vector machines: learning with many relevant features. Proceedings of ECML-98, 10th European Conference on Machine Learning, 1398, Chemnitz, DE. 1998, 137-142.

Furey TS, Cristianini N, Duffy N, Bednarski DW, Schummer M, Haussler D: Support vector machine classification and validation of cancer tissue samples using microarray expression data. Bioinformatics. 2000, 16 (10): 906-914.

Lee C, Chung P, Tsai J, Chang C: Robust radial basis function neural networks. Systems, Man, and Cybernetics, Part B, IEEE Transactions on. 1999, 29 (6): 674-685.

Lu Y, Sundararajan N, Saratchandran P: Performance evaluation of a sequential minimal radial basisfunction (RBF) neural network learning algorithm. IEEE Trans Neural Networks. 1998, 9 (2): 308-318.

Stone M: Cross-Validatory Choice and Assessment of Statistical Predictions. J Royal Statistic Soc Series B (Methodological). 1974, 36 (2): 111-147.

Hosmer D, Lemeshow S: Applied Logistic Regression (Wiley Series In Probability and Statistics). 1989, New York: Wiley, Chapter 1:

### Pre-publication history

The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1472-6947/9/2/prepub

## Acknowledgements

This research was partially funded by research grants from Health Services Foundation, Carolinas HealthCare System, and Virginia Commonwealth University. This material is based upon work supported by the National Science Foundation under Grant No.IIS0758410. The authors would like to thank these institutions for their support.

## Author information

### Authors and Affiliations

### Corresponding author

## Additional information

### Competing interests

The authors declare that they have no competing interests.

### Authors' contributions

TH is responsible for obtaining the data and providing feedback on the medical impact of the results. SJ, RS and KN have designed the algorithms, analyzed the data, and drafted the manuscript. All authors have equal participation in the study as well as preparation of the final paper.

## Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

## Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

## About this article

### Cite this article

Ji, SY., Smith, R., Huynh, T. *et al.* A comparative analysis of multi-level computer-assisted decision making systems for traumatic injuries.
*BMC Med Inform Decis Mak* **9**, 2 (2009). https://doi.org/10.1186/1472-6947-9-2

Received:

Accepted:

Published:

DOI: https://doi.org/10.1186/1472-6947-9-2

### Keywords

- Support Vector Machine
- Traumatic Brain Injury
- Multivariate Adaptive Regression Spline
- Predict Patient Survival
- National Trauma Data Bank