- Research article
- Open Access
- Open Peer Review

This article has Open Peer Review reports available.

# A simulation study comparing aberration detection algorithms for syndromic surveillance

- Michael L Jackson
^{1, 2}Email author, - Atar Baer
^{1}, - Ian Painter
^{3}and - Jeff Duchin
^{1, 2}

**7**:6

https://doi.org/10.1186/1472-6947-7-6

© Jackson et al; licensee BioMed Central Ltd. 2007

**Received: **06 December 2006

**Accepted: **01 March 2007

**Published: **01 March 2007

## Abstract

### Background

The usefulness of syndromic surveillance for early outbreak detection depends in part on effective statistical aberration detection. However, few published studies have compared different detection algorithms on identical data. In the largest simulation study conducted to date, we compared the performance of six aberration detection algorithms on simulated outbreaks superimposed on authentic syndromic surveillance data.

### Methods

We compared three control-chart-based statistics, two exponential weighted moving averages, and a generalized linear model. We simulated 310 unique outbreak signals, and added these to actual daily counts of four syndromes monitored by Public Health – Seattle and King County's syndromic surveillance system. We compared the sensitivity of the six algorithms at detecting these simulated outbreaks at a fixed alert rate of 0.01.

### Results

Stratified by baseline or by outbreak distribution, duration, or size, the generalized linear model was more sensitive than the other algorithms and detected 54% (95% CI = 52%–56%) of the simulated epidemics when run at an alert rate of 0.01. However, all of the algorithms had poor sensitivity, particularly for outbreaks that did not begin with a surge of cases.

### Conclusion

When tested on county-level data aggregated across age groups, these algorithms often did not perform well in detecting signals other than large, rapid increases in case counts relative to baseline levels.

## Keywords

## Background

In the short time since syndromic surveillance[1] was introduced as an early warning system for detecting outbreaks, considerable effort and expense have gone into developing syndromic surveillance systems. Although there have been substantial developments in the methods and tools used for this practice, the public health value of the various approaches to syndromic surveillance has rarely been evaluated. In particular, a critical component needing further study is the relative accuracy and timeliness of the aberration detection methods of these systems.

In aberration detection, statistical models determine whether the counts in a given syndrome and day are unusually high and thus worth investigating. Many statistical algorithms are available, including control-chart-based models[2, 3], scan statistics[4, 5], autoregressive moving averages[6], and regression models[7, 8]. To optimize outbreak detection, surveillance system designers and users need to understand which methods perform well or poorly in different settings.

Many studies have described the performance of individual aberration detection methods [3–10]. However, multiple algorithms have seldom been compared on the same data [11–14], which is problematic because algorithms that work well for one data source may not do as well for another. Further, studies comparing different algorithms have either tested the algorithms on only a single type of outbreak[11, 14], or have used simulated baseline data rather than syndrome counts from real systems[12, 13]. Using real system data is important, because it is unknown how well simulated data approximate the relevant features of real syndrome counts. Testing performance on outbreaks of different temporal distributions and sizes is important because of the uncertainty about the types of outbreaks likely to be encountered in practice. To date, no published studies have systematically compared detection methods using real syndromic surveillance data. This lack of comparisons on actual syndromic surveillance data makes it difficult to select aberration detection methods objectively.

To address these limitations, we compared the utility of six commonly-used aberration detection algorithms using data from our syndromic surveillance system at Public Health – Seattle & King County. We simulated syndrome counts that might result from a variety of outbreaks, and added these to actual daily counts from syndromes monitored by our system. We then evaluated the performance of the six algorithms on the resulting data.

## Methods

Our syndromic surveillance system receives data from 18 of 19 emergency departments (EDs) in King County. Each morning, EDs send data on all visits that occurred the previous day, including the date and time of the visit; the patient's age, sex, and home zip code; a free-text chief complaint; diagnosis, if available; and disposition. Chief complaints are classified into syndromes based on the presence or absence of key words using a modified version of the chief complaint coder developed by the New York City Department of Health and Mental Hygiene[15]. For each syndrome, daily case counts are determined both by age group and aggregated across age groups. In this study, we used the aggregated counts. Because historical data were not available from all EDs when we began this study, we restricted our analysis to the nine EDs with complete reporting from 2001–2004.

Because we could not use all of the syndromes we monitor in this analysis, we chose a representative set by grouping syndromes into four categories based on their mean daily counts, and then selecting one syndrome from each group. The final syndromes we used as baselines had mean (standard deviation) daily counts of 60 (16.0), 35 (9.9), 10 (4.0), and 2 (1.6) visits per day. These counts corresponded to ED visits for respiratory illness, influenza-like illness, and asthma syndromes, and pneumonia hospitalizations, respectively. From these four baselines we used data from 2001 through 2003 to provide background counts for the algorithms, and used data from 2004 for testing the algorithms. For each of these syndromes, we calculated the variability in daily counts by weekday and month using Poisson regression [see Additional file 1]. Notably, all four baselines had significant month effects, indicating the presence of seasonal trends. The four baselines also showed significant day-of-week effects.

We compared the performance of six aberration detection methods. We evaluated the three control-chart-based algorithms commonly referred to as C1, C2, and C3[13]. For C1 and C2, the test statistic on day t was calculated as

S_{t} = max(0, (X_{t} - (μ_{t} + k*σ_{t}))/σ_{t})

where X_{t} is the count on day t, k is the shift from the mean to be detected, and μ_{t} and σ_{t} are the mean and standard deviation of the counts during the baseline period. For C1, the baseline period is (t-7,...,t-1); for C2 the baseline is (t-9,...,t-3). The test statistic for C3 is the sum of S_{t} + S_{t-1} + S_{t-2} from the C2 algorithm.

We evaluated a generalized linear model (GLM), using a three-year baseline and Poisson errors, with terms for day of the week, month, linear time trend, and holidays. The full model for the expected count on day t was

E(X_{t}) = β0 + β1(Sunday) + ... + β6(Friday) + β7(January) + ... + β17(November) + β18(Holiday) + β19(time trend)

and the test statistic was the probability from a Poisson distribution of observing at least X_{t} cases given E(X_{t}).

We also included two Exponential Weighted Moving Average (EWMA) models[16], using a 28-day baseline and smoothing constants of 0.4 (EWMA4) and 0.9 (EWMA9). The smoothed daily counts were calculated as

Y_{1} = X_{1}; Y_{t} = ω*X_{t} + (1 - ω)*Y_{t}-1

and the test statistic was calculated as

T_{t} = (Y_{t} - μ_{t})/[σ_{t}*(ω/(2 - ω))^{1/2}]

where ω is the smoothing constant and other notation as defined above. The baseline period for μ_{t} and σ_{t} was set to (t-30,...,t-3).

### Outbreak simulation

Next, we simulated daily outbreak signal counts by creating every unique combination of the temporal distributions, durations, and sizes. Since all outbreaks of one day duration have the same temporal distribution, this gave us 310 different outbreak signals. For each signal, we converted the epidemic curve of the appropriate historical outbreak into an empirical probability distribution. We divided this distribution into the appropriate number of days for the simulated outbreak. We then calculated the cumulative probability of a case occurring on or before day d for each day (1,...,d) of the outbreak signal. Finally, we assigned each case a random number between 0 and 1, chosen from a uniform random distribution. The total number of cases on day d was the sum of all cases whose random number was greater than the cumulative probability for day d-1, and less than or equal to the cumulative probability for day d. This random assignment was repeated for each of the 310 outbreak signals.

### Algorithm testing

We created evaluation datasets by adding the simulated outbreak signals to each of the four baselines, with the outbreak counts starting on January 2nd, 2004. To avoid bias due to day-of-the-week effects and seasonality, we repeated this process starting the outbreak on every other day of 2004 for each of the four baselines. This gave us 183 datasets per outbreak per baseline, for a total of 226,920 datasets for analysis.

We applied the six algorithms to the evaluation datasets, and calculated two outcome measures for each algorithm on each dataset. The first was whether the algorithm ever detected the outbreak signal. The second was the earliest day of an alert, among signals that were detected. The earliest day of alert was counted from the day on which the first simulated case was added. Due to the stochastic nature of these simulations, this was not always the first day of the epidemic. For example, consider a simulated outbreak signal of five cases 32 days duration started on January 2^{nd}, 2004. The cases might appear on days 3, 5, 8, 17, and 30 of the signal. In this situation, we would begin counting the days until detection starting from day 3 (January 4^{th}).

We calculated these outcomes while running the algorithms at an alert rate of 0.01 (an average of one alert every hundred days). We set this alert rate empirically by applying each algorithm to each baseline without any added outbreak signals, and determining the threshold that would yield an average of one alert per 100 days. Note that in this study we defined an alert rate, rather than a false positive rate (which is 1 – specificity). Calculating a false positive rate (i.e. a specificity) requires assuming that the baseline syndromes did not contain any signals from true outbreaks in the community. This is an unreasonable assumption, given the known yearly influenza epidemics and the probable presence of other unknown outbreaks. By using an alert rate instead of a false positive rate, we allowed for the existence of these signals in our baseline data.

### Comparing algorithm performance

For each of the 310 outbreak signals, in each of the four baseline syndromes, we computed the sensitivity (that is, the probability of detecting the signal given the presence of an outbreak) by averaging the detection outcome across all 183 analysis runs. We also computed the median of the earliest day of detection. We used ANOVA to test for significant differences in the algorithms' sensitivities. Separate tests were conducted by baseline and by outbreak distribution, duration, and size. In each stratum of these four grouping variables we tested for significant differences between algorithms in the probability of detection. We also tested for differences within each algorithm across the strata. For all ANOVA comparisons that were significant at the 0.05 level, we compared the performance of all pairs of algorithms by t-test.

## Results

Mean percent of outbreak signals detected by six aberration detection algorithms, tested on four baseline syndromes, at an alert rate of 0.01

Syndrome: | C1 | C2 | C3 | GLM | EWMA 0.4 | EWMA 0.9 | F value | p-value | |
---|---|---|---|---|---|---|---|---|---|

| Mean | 62.0% | 69.3% | 72.0% | 85.3% | 80.5% | 79.6% | 24.94 | <0.0001 |

Lower CI* | 58.3% | 65.8% | 68.6% | 82.5% | 77.2% | 76.3% | |||

Upper CI | 65.8% | 72.9% | 75.5% | 88.0% | 83.9% | 82.9% | |||

| Mean | 35.4% | 52.2% | 48.8% | 70.0% | 52.3% | 57.7% | 36.63 | <0.0001 |

Lower CI | 31.7% | 48.5% | 45.1% | 66.5% | 48.6% | 53.9% | |||

Upper CI | 39.0% | 56.0% | 52.5% | 73.4% | 56.0% | 61.5% | |||

| Mean | 21.2% | 17.9% | 11.1% | 34.4% | 19.0% | 28.2% | 84.08 | <0.0001 |

Lower CI | 19.6% | 16.5% | 10.4% | 31.9% | 17.4% | 26.0% | |||

Upper CI | 22.9% | 19.2% | 11.9% | 36.9% | 20.5% | 30.4% | |||

| Mean | 16.4% | 17.1% | 10.8% | 27.8% | 13.3% | 15.7% | 84.90 | <0.0001 |

Lower CI | 15.1% | 16.0% | 10.3% | 25.7% | 12.5% | 14.5% | |||

Upper CI | 17.6% | 18.3% | 11.4% | 29.8% | 14.0% | 16.9% | |||

| 205.58 | 349.50 | 531.70 | 398.73 | 550.30 | 406.76 | |||

| <0.0001 | <0.0001 | <0.0001 | <0.0001 | <0.0001 | <0.0001 |

Mean percent of outbreak signals detected by six aberration detection algorithms, tested on five outbreak distributions, at an alert rate of 0.01

Distribution: | C1 | C2 | C3 | GLM | EWMA 0.4 | EWMA 0.9 | F value | p-value | |
---|---|---|---|---|---|---|---|---|---|

| Mean | 37.4% | 41.5% | 36.5% | 56.5% | 41.6% | 46.9% | 12.85 | <0.0001 |

Lower CI* | 33.5% | 37.5% | 32.4% | 52.4% | 37.5% | 42.7% | |||

Upper CI | 41.2% | 45.4% | 40.7% | 60.5% | 45.8% | 51.1% | |||

| Mean | 38.1% | 41.7% | 38.0% | 56.5% | 42.4% | 47.5% | 9.49 | <0.0001 |

Lower CI | 33.9% | 37.3% | 33.5% | 52.1% | 37.7% | 42.8% | |||

Upper CI | 42.3% | 46.2% | 42.6% | 60.9% | 47.0% | 52.1% | |||

| Mean | 32.2% | 37.2% | 34.6% | 53.0% | 40.1% | 43.7% | 12.11 | <0.0001 |

Lower CI | 28.3% | 33.1% | 30.2% | 48.7% | 35.6% | 39.3% | |||

Upper CI | 36.0% | 41.4% | 38.9% | 57.3% | 44.6% | 48.1% | |||

| Mean | 32.1% | 37.5% | 34.8% | 52.8% | 41.2% | 44.1% | 11.57 | <0.0001 |

Lower CI | 28.3% | 33.4% | 30.4% | 48.4% | 36.7% | 39.6% | |||

Upper CI | 36.0% | 41.7% | 39.2% | 57.2% | 45.8% | 48.6% | |||

| Mean | 28.4% | 37.4% | 34.4% | 52.6% | 40.9% | 43.9% | 15.05 | <0.0001 |

Lower CI | 25.1% | 33.2% | 30.1% | 48.2% | 36.3% | 39.4% | |||

Upper CI | 31.7% | 41.5% | 38.8% | 57.0% | 45.5% | 48.4% | |||

| 4.39 | 1.22 | 0.05 | 0.86 | 0.13 | 0.65 | |||

| 0.0016 | 0.30 | 0.74 | 0.4881 | 0.97 | 0.63 |

Mean percent of outbreak signals detected by six aberration detection algorithms, tested on six outbreak sizes, at an alert rate of 0.01

Size: | C1 | C2 | C3 | GLM | EWMA 0.4 | EWMA 0.9 | F value | p-value | |
---|---|---|---|---|---|---|---|---|---|

| Mean | 10.4% | 11.9% | 9.6% | 16.0% | 10.3% | 12.0% | 9.55 | <0.0001 |

Lower CI* | 8.9% | 10.5% | 8.3% | 14.1% | 9.0% | 10.6% | |||

Upper CI | 11.8% | 13.3% | 10.9% | 18.0% | 11.7% | 13.5% | |||

| Mean | 16.4% | 18.8% | 15.4% | 30.2% | 20.2% | 22.9% | 10.01 | <0.0001 |

Lower CI | 13.7% | 16.2% | 13.0% | 25.7% | 16.7% | 19.0% | |||

Upper CI | 19.2% | 21.5% | 17.9% | 34.6% | 23.8% | 26.7% | |||

| Mean | 26.7% | 32.2% | 29.4% | 48.4% | 34.8% | 38.4% | 8.17 | <0.0001 |

Lower CI | 22.1% | 27.3% | 24.3% | 42.5% | 29.0% | 32.7% | |||

Upper CI | 31.3% | 37.2% | 34.5% | 54.3% | 40.6% | 44.2% | |||

| Mean | 38.9% | 46.2% | 42.5% | 63.0% | 47.8% | 52.5% | 7.42 | <0.0001 |

Lower CI | 33.3% | 40.0% | 36.0% | 57.2% | 41.3% | 46.1% | |||

Upper CI | 44.5% | 52.4% | 49.0% | 68.7% | 54.3% | 58.9% | |||

| Mean | 47.1% | 53.7% | 50.0% | 71.6% | 56.4% | 61.6% | 8.11 | <0.0001 |

Lower CI | 41.2% | 47.4% | 43.2% | 66.6% | 49.8% | 55.4% | |||

Upper CI | 53.0% | 59.9% | 56.9% | 76.6% | 63.0% | 67.7% | |||

| Mean | 52.9% | 58.8% | 53.8% | 77.4% | 62.7% | 67.9% | 9.83 | <0.0001 |

Lower CI | 47.0% | 52.8% | 47.0% | 72.9% | 56.5% | 62.2% | |||

Upper CI | 58.7% | 64.8% | 60.6% | 81.9% | 68.9% | 73.7% | |||

| 35.81 | 38.67 | 31.48 | 64.53 | 37.68 | 46.16 | |||

| <0.0001 | <0.0001 | <0.0001 | <0.0001 | <0.0001 | <0.0001 |

Mean percent of outbreak signals detected by six aberration detection algorithms, tested on seven outbreak durations, at an alert rate of 0.01

Duration: | C1 | C2 | C3 | GLM | EWMA 0.4 | EWMA 0.9 | F value | p-value | |
---|---|---|---|---|---|---|---|---|---|

| Mean | 59.8% | 58.2% | 44.5% | 73.4% | 52.6% | 64.6% | 2.58 | 0.03 |

Lower CI* | 47.5% | 45.7% | 31.0% | 62.4% | 39.6% | 52.4% | |||

Upper CI | 72.2% | 70.7% | 58.1% | 84.4% | 65.5% | 76.8% | |||

| Mean | 47.3% | 49.1% | 42.4% | 64.2% | 47.9% | 55.2% | 7.81 | <0.0001 |

Lower CI | 42.0% | 43.7% | 36.7% | 59.2% | 42.3% | 49.8% | |||

Upper CI | 52.6% | 54.4% | 48.0% | 69.2% | 53.5% | 60.7% | |||

| Mean | 39.6% | 43.5% | 40.1% | 57.7% | 43.6% | 48.9% | 6.43 | <0.0001 |

Lower CI | 34.5% | 38.2% | 34.7% | 52.6% | 38.2% | 43.5% | |||

Upper CI | 44.7% | 48.8% | 45.5% | 62.7% | 49.1% | 54.4% | |||

| Mean | 33.2% | 39.8% | 37.6% | 54.0% | 41.7% | 45.5% | 7.65 | <0.0001 |

Lower CI | 28.5% | 34.7% | 32.4% | 48.9% | 36.4% | 40.2% | |||

Upper CI | 37.8% | 44.9% | 42.9% | 59.0% | 47.1% | 50.8% | |||

| Mean | 27.7% | 36.4% | 34.7% | 51.3% | 39.2% | 42.2% | 10.55 | <0.0001 |

Lower CI | 23.8% | 31.7% | 29.8% | 46.3% | 34.0% | 37.2% | |||

Upper CI | 31.6% | 41.2% | 39.7% | 56.2% | 44.3% | 47.3% | |||

| Mean | 23.0% | 31.0% | 30.2% | 47.7% | 36.5% | 38.3% | 17.22 | <0.0001 |

Lower CI | 20.5% | 27.4% | 26.1% | 43.2% | 32.0% | 33.9% | |||

Upper CI | 25.5% | 34.6% | 34.4% | 52.2% | 41.0% | 42.7% | |||

| Mean | 26.5% | 31.2% | 27.4% | 47.5% | 36.5% | 37.5% | 36.25 | <0.0001 |

Lower CI | 25.5% | 29.7% | 25.0% | 44.2% | 33.1% | 34.6% | |||

Upper CI | 27.5% | 32.8% | 29.8% | 50.8% | 39.8% | 40.3% | |||

| 35.81 | 38.67 | 31.48 | 64.53 | 37.68 | 46.16 | |||

| <0.0001 | <0.0001 | <0.0001 | <0.0001 | <0.0001 | <0.0001 |

This stratification suggests several qualitative trends. First, regardless of distribution, the GLM tended to have low sensitivity when the baseline counts for a given syndrome were 35 per day or higher. As shown in Figure 3, outbreak detection under these circumstances was generally below 80%, except in some situations where the outbreak size was large relative to the baseline (i.e., 35–40 excess counts per day) and the surge in cases occurred on a single day. For example, a single-day surge of influenza-like illness resulting in 35 cases above the average daily baseline count of 35 cases was detected in 84% of the tests. That is, a doubling of the average daily case counts on a single day could be reliably detected 84% of the time. This detection level dropped to 55–74% (depending on the distribution) if the excess counts were spread over a period of two days. Sensitivity was poorer with smaller outbreak sizes. For example, a single-day surge of 20 cases of influenza-like illness (a 57% increase over the mean baseline count of 35 cases per day) was detected in less than half (46%) of the tests.

Second, the GLM's performance was generally better for situations where the average daily baseline count was lower (i.e., 10 or fewer cases). In this scenario, the algorithm tended to detect the surge in cases within the first or second day with high probability (>80%). The probability of detection was directly related to outbreak size and inversely related to outbreak duration.

Third, there were some qualitative differences in detection according to outbreak distribution. Distributions where the bulk of the cases occurred early (such as the airborne bioweapon) tended to be detected earlier than slowly-building outbreaks (such as the multi-modal community transmission).

## Discussion

Although public health departments have been quick to adopt syndromic surveillance systems for outbreak detection, few studies have demonstrated system effectiveness. The practical utility of syndromic surveillance depends on several factors, including the quality and accuracy of the source data; the sensitivity, specificity, predictive value, and timeliness of the aberration detection methods; and the user's response to alerts given by the system. In this study, we compared the performance of six aberration detection methods under a broad set of circumstances by adding simulated outbreak signals to actual daily syndrome counts. To our knowledge, this is the largest simulation study comparing syndromic surveillance algorithms that has been published to date. Our evaluation was based on 310 unique outbreak signals, each of which was tested on 183 separate dates in each of four baselines.

Consistent with previous evaluations of syndromic surveillance algorithms[10, 12, 24], we found that the ability to detect outbreaks was better with larger outbreaks and lower baseline counts. We also observed that detection of a signal of a fixed size increases as the baseline counts decrease, which has been observed elsewhere[13]. This is unsurprising, as a fixed signal size causes a greater relative increase in the case count for a baseline with few daily cases compared to a baseline with many daily cases. Furthermore, our findings suggest that the temporal distribution of cases (i.e., the shape of the epidemic curve) generally does not affect sensitivity but may affect timeliness, which Wang et al. observed in evaluating an autoregressive periodic model (APM)[10].

Beyond these general observations, we had two more specific findings. First, across different baselines, and different characteristics of the outbreak signals, the generalized linear model detected more outbreak signals than the other five algorithms. This was surprising, as we had expected to observe more heterogeneity in the relative performance of the algorithms, particularly across outbreak distributions. Buckeridge et al[25] have suggested that EWMA9 (which approximates a Shewhart chart) should outperform EWMA4 on single-day spikes and scattered signals (such as the multi-modal community transmission), while EWMA4 should be better on lognormal curves that increase rapidly (such as the close-contact community transmission). Furthermore, the control-chart-based methods have been reported to have good sensitivity for detection in rare syndromes, with C3 better suited than C1 and C2 for detecting slowly building outbreaks[13]. Of note, our baseline syndromes all had strong day-of-week trends [see Appendix]; since the GLM included weekday parameters, this may have contributed to its superior performance relative to the other methods, which do not correct for such trends[7].

Secondly, although we found that the GLM tended to perform better than the other methods, all six algorithms performed poorly in many outbreak scenarios. The algorithms were generally able to detect large one- or two-day surges in case counts (where "large" means exceeding twice or three times the standard deviation of the baseline), or signals of longer duration that were very large relative to the baseline. However, these are the types of signals most likely to be detected by the astute clinician (in the case of outpatient or ED data) or by an epidemiologist visually looking for jumps in counts or proportions in a time series. Aberration detection is most needed for detecting low-to-moderate increases in cases and for slowly increasing outbreaks. Yet when run at a rate of one alert per 100 days, none of the algorithms we tested detected these types of signals reliably, suggesting that users run a high risk of missing outbreaks of interest across a wide range of scenarios.

One feature of Figure 3 deserves mention here. As the duration of the signal increases, sensitivity appears to follow either a U-shaped trend (when sensitivity is high in the early days of the outbreak) or to increase with increasing duration. At first glance, this is counter-intuitive, as the ability to detect a signal of a given size should decrease as the cases are distributed over more days. In this case, the superior sensitivity for longer outbreaks does not represent a real benefit, but rather reflects the fact that there is a set alert rate of 0.01. An alert unrelated to the presence of a simulated outbreak, due to variation in the baseline syndrome counts, will occur roughly once every hundred days. It is more likely that such an alert will happen by chance during a simulated signal when that signal occurs over the course of 32 days than when the signal occurs over the course of one or two days.

There are several factors which may limit the generalizability of our findings. First, the results of this study are modelled on the allocation of cases along epidemic curves of a fixed distribution, duration, and size. We did not model the full progression from exposure to disease to healthcare use, as has been done previously for some limited outbreak types[17]. Rather, we attempted to cover a wide array of the outbreak signals that could occur in practice. Thus, our results may not be generalizable to all outbreak settings, because the scenarios we produced may not be entirely representative of the scenarios for which the algorithms were designed to operate[26]. In addition, we set the alert rates for each algorithm empirically, by applying the algorithms to the baseline syndromes and finding the threshold that would yield approximately one alert every 100 days. The sensitivity of the algorithms used in this analysis will likely differ if those same algorithms are applied in systems that use different thresholds. Furthermore, the thresholds that gave an alert rate of 0.01 in our data may yield different alert rates in other data, and may also differ when applied to stratifications of the data, such as by age categories or geographic groupings.

It is a limitation of the syndromic surveillance literature that each evaluation has used different baseline data. The baselines likely differ in terms of random and systematic variation. Furthermore, published studies have rarely reported detailed information about their baseline time series; often, studies have not even reported mean counts or variances. This limitation makes comparisons between studies difficult and extrapolation to other datasets uncertain. We included the appendix, with its detail about the syndromes we used, primarily so that our baselines can be compared with other time series, at least in terms of mean counts, day-of-week effects, and seasonal trends. This gives other users a better basis for comparing our baseline counts with their own syndromic data. However, there may be other features of syndromic time series that affect algorithm performance. The field of syndromic surveillance would greatly benefit from an analysis of the features of syndromic time series that impact detection, and the relative importance of these features.

The poor overall performance of the algorithms we examined raises another question: Are there other algorithms that may perform better? Because this study was computationally intensive, we did not evaluate other algorithms that have been proposed for syndromic surveillance, such as autoregressive models[6, 24] or the Pulsar method[11]. We were also unable to evaluate scan statistics[4, 5] or other methods that use spatial data, because our simulation method aggregated cases to the county level. Comparing our results to prior studies is difficult, not only because of the different baselines as mentioned above, but because studies have varied in the alert rates at which they have tested the algorithms. Prior studies have tended to use rates between 0.05 (one alert every 20 days) and 0.03 (roughly one alert per month). We feel that these rates are too high for routine surveillance and could desensitize users, leading to a reduced likelihood of following up on any given alert[27]. Because of the problems in comparing evaluations of algorithms across data sets, it is difficult to determine whether other algorithms might have performed better on our data. This remains an area for active research.

## Conclusion

The results of this study suggest that commonly-used aberration detection methods for syndromic surveillance often do not perform well in detecting signals other than large, rapid increases in case counts relative to baseline levels. To the degree that our results are generalizable to other settings, this poor performance may be a feature of other systems as well. These results suggest that users should exercise some caution in reviewing algorithm output. Although the GLM method tended to have better sensitivity overall, there was variability in algorithm performance across outbreak feature sets, illustrating that a one-size-fits-all method is unlikely. Additional work is still needed to develop and evaluate syndromic surveillance algorithms across outbreak signals and to determine the value of these systems in public health practice.

## Declarations

### Acknowledgements

This research is funded by the Centers for Disease Control and Prevention through a grant for Public Health Emergency Preparedness and Response, Focus Area B, Federal Catalogue Number 93.283. The funding agency did not contribute to the design or conduct of the study, the drafting of the manuscript, or the decision to publish.

## Authors’ Affiliations

## References

- Henning KJ: Appendix B: Syndromic Surveillance. Institute of Medicine, 2003. Microbial Threats to Health: Emergence, Detection, and Response. Edited by: Hamburg MA, Lederberg J. 2003, Washington, DC, The National Academies Press, 309-350.Google Scholar
- Hutwagner L, Thompson W, Seeman GM, Treadwell T: The bioterrorism preparedness and response Early Aberration Reporting System (EARS). J Urban Health. 2003, 80: i89-96.PubMedPubMed CentralGoogle Scholar
- Rossi G, Lampugnani L, Marchi M: An approximate CUSUM procedure for surveillance of health events. Stat Med. 1999, 18: 2111-2122. 10.1002/(SICI)1097-0258(19990830)18:16<2111::AID-SIM171>3.0.CO;2-Q.View ArticlePubMedGoogle Scholar
- Kleinman KP, Abrams AM, Kulldorff M, Platt R: A model-adjusted space-time scan statistic with an application to syndromic surveillance. Epidemiol Infect. 2005, 133: 409-419. 10.1017/S0950268804003528.View ArticlePubMedPubMed CentralGoogle Scholar
- Kulldorff M, Heffernan R, Hartman J, Assuncao R, Mostashari F: A space-time permutation scan statistic for disease outbreak detection. PLoS Med. 2005, 2: e59-10.1371/journal.pmed.0020059.View ArticlePubMedPubMed CentralGoogle Scholar
- Miller B, Kassenborg H, Dunsmuir W, Griffith J, Hadidi M, Nordin JD, Danila R: Syndromic surveillance for influenzalike illness in ambulatory care network. Emerg Infect Dis. 2004, 10: 1806-1811.View ArticlePubMedPubMed CentralGoogle Scholar
- Kleinman K, Lazarus R, Platt R: A generalized linear mixed models approach for detecting incident clusters of disease in small areas, with an application to biological terrorism. Am J Epidemiol. 2004, 159: 217-224. 10.1093/aje/kwh029.View ArticlePubMedGoogle Scholar
- Lazarus R, Kleinman K, Dashevsky I, Adams C, Kludt P, DeMaria A, Platt R: Use of automated ambulatory-care encounter records for detection of acute illness clusters, including potential bioterrorism events. Emerg Infect Dis. 2002, 8: 753-760.View ArticlePubMedPubMed CentralGoogle Scholar
- Olson KL, Bonetti M, Pagano M, Mandl KD: Real time spatial cluster detection using interpoint distances among precise patient locations. BMC Med Inform Decis Mak. 2005, 5: 19-10.1186/1472-6947-5-19.View ArticlePubMedPubMed CentralGoogle Scholar
- Wang L, Ramoni MF, Mandl KD, Sebastiani P: Factors affecting automated syndromic surveillance. Artif Intell Med. 2005, 34: 269-278. 10.1016/j.artmed.2004.11.002.View ArticlePubMedGoogle Scholar
- Dafni UG, Tsiodras S, Panagiotakos D, Gkolfinopoulou K, Kouvatseas G, Tsourti Z, Saroglou G: Algorithm for statistical detection of peaks–syndromic surveillance system for the Athens 2004 Olympic Games. MMWR Morb Mortal Wkly Rep. 2004, 53 (Suppl): 86-94.Google Scholar
- Hutwagner L, Browne T, Seeman GM, Fleischauer AT: Comparing aberration detection methods with simulated data. Emerg Infect Dis. 2005, 11: 314-316.View ArticlePubMedPubMed CentralGoogle Scholar
- Hutwagner LC, Thompson WW, Seeman GM, Treadwell T: A simulation model for assessing aberration detection methods used in public health surveillance for systems with limited baselines. Stat Med. 2005, 24: 543-550. 10.1002/sim.2034.View ArticlePubMedGoogle Scholar
- Zhang J, Tsui FC, Wagner MM, Hogan WR: Detection of outbreaks from time series data using wavelet transform. AMIA Annu Symp Proc. 2003, 748-752.Google Scholar
- Heffernan R, Mostashari F, Das D, Karpati A, Kuldorff M, Weiss D: Syndromic surveillance in public health practice, New York City. Emerg Infect Dis. 2004, 10: 858-864.View ArticlePubMedGoogle Scholar
- Burkom H: Development, Adaptation, and Assessment of Alerting Algorithms for Biosurveillance. Johns Hopkins APL Technical Digest. 2003, 24: 335-342.Google Scholar
- Buckeridge DL, Switzer P, Owens D, Siegrist D, Pavlin J, Musen M: An evaluation model for syndromic surveillance: assessing the performance of a temporal algorithm. MMWR Morb Mortal Wkly Rep. 2005, 54 (Suppl): 109-115.Google Scholar
- Mandl KD, Reis BY, Cassa C: Measuring Outbreak-Detection Performance by Using Controlled Feature Set Simulations. MMWR. 2004, 53 (Suppl): 130-136.Google Scholar
- Meselson M, Guillemin J, Hugh-Jones M, Langmuir A, Popova I, Shelokov A, Yampolskaya O: The Sverdlovsk anthrax outbreak of 1979. Science. 1994, 266: 1202-1208. 10.1126/science.7973702.View ArticlePubMedGoogle Scholar
- Outbreaks of Norwalk-like viral gastroenteritis–Alaska and Wisconsin, 1999. MMWR Morb Mortal Wkly Rep. 2000, 49: 207-211.Google Scholar
- Outbreak of Ebola hemorrhagic fever Uganda, August 2000-January 2001. MMWR Morb Mortal Wkly Rep. 2001, 50: 73-77.Google Scholar
- Measles outbreak–Netherlands, April 1999-January 2000. MMWR Morb Mortal Wkly Rep. 2000, 49: 299-303.Google Scholar
- Outbreak of measles among Christian Science students–Missouri and Illinois, 1994. MMWR Morb Mortal Wkly Rep. 1994, 43: 463-465.Google Scholar
- Reis BY, Mandl KD: Time series modeling for syndromic surveillance. BMC Med Inform Decis Mak. 2003, 3: 2-10.1186/1472-6947-3-2.View ArticlePubMedPubMed CentralGoogle Scholar
- Buckeridge DL, Burkom H, Campbell M, Hogan WR, Moore AW: Algorithms for rapid outbreak detection: a research synthesis. J Biomed Inform. 2005, 38: 99-113. 10.1016/j.jbi.2004.11.007.View ArticlePubMedGoogle Scholar
- Buckeridge DL, Burkom H, Moore A, Pavlin J, Cutchis P, Hogan W: Evaluation of syndromic surveillance systems–design of an epidemic simulation model. MMWR Morb Mortal Wkly Rep. 2004, 53 (Suppl): 137-143.Google Scholar
- Stoto MA, Schonlau M, Mariano LT: Syndromic Surveillance: Is It Worth the Effort?. Chance. 2004, 17: 19-24.View ArticleGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1472-6947/7/6/prepub

### Pre-publication history

## Copyright

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.