Accounting for seasonal patterns in syndromic surveillance data for outbreak detection

<p>Abstract</p> <p>Background</p> <p>Syndromic surveillance (SS) can potentially contribute to outbreak detection capability by providing timely, novel data sources. One SS challenge is that some syndrome counts vary with season in a manner that is not identical from ye...

Full description

Bibliographic Details
Main Authors: Picard Richard, Michalak Sarah, Klamann Richard, Graves Todd, Burr Tom, Hengartner Nicolas
Format: Article
Language:English
Published: BMC 2006-12-01
Series:BMC Medical Informatics and Decision Making
Online Access:http://www.biomedcentral.com/1472-6947/6/40
id doaj-7a8f799aa8e9442190734f5d519f9126
record_format Article
spelling doaj-7a8f799aa8e9442190734f5d519f91262020-11-25T00:44:45ZengBMCBMC Medical Informatics and Decision Making1472-69472006-12-01614010.1186/1472-6947-6-40Accounting for seasonal patterns in syndromic surveillance data for outbreak detectionPicard RichardMichalak SarahKlamann RichardGraves ToddBurr TomHengartner Nicolas<p>Abstract</p> <p>Background</p> <p>Syndromic surveillance (SS) can potentially contribute to outbreak detection capability by providing timely, novel data sources. One SS challenge is that some syndrome counts vary with season in a manner that is not identical from year to year.</p> <p>Our goal is to evaluate the impact of inconsistent seasonal effects on performance assessments (false and true positive rates) in the context of detecting anomalous counts in data that exhibit seasonal variation.</p> <p>Methods</p> <p>To evaluate the impact of inconsistent seasonal effects, we injected synthetic outbreaks into real data and into data simulated from each of two models fit to the same real data. Using real respiratory syndrome counts collected in an emergency department from 2/1/94–5/31/03, we varied the length of training data from one to eight years, applied a sequential test to the forecast errors arising from each of eight forecasting methods, and evaluated their detection probabilities (DP) on the basis of 1000 injected synthetic outbreaks. We did the same for each of two corresponding simulated data sets. The less realistic, nonhierarchical model's simulated data set assumed that "one season fits all," meaning that each year's seasonal peak has the same onset, duration, and magnitude. The more realistic simulated data set used a hierarchical model to capture violation of the "one season fits all" assumption.</p> <p>Results</p> <p>This experiment demonstrated optimistic bias in DP estimates for some of the methods when data simulated from the nonhierarchical model was used for DP estimation, thus suggesting that at least for some real data sets and methods, it is not adequate to assume that "one season fits all."</p> <p>Conclusion</p> <p>For the data we analyze, the "one season fits all " assumption is violated, and DP performance claims based on simulated data that assume "one season fits all," for the forecast methods considered, except for moving average methods, tend to be optimistic. Moving average methods based on relatively short amounts of training data are competitive on all three data sets, but are particularly competitive on the real data and on data from the hierarchical model, which are the two data sets that violate the "one season fits all" assumption.</p> http://www.biomedcentral.com/1472-6947/6/40
collection DOAJ
language English
format Article
sources DOAJ
author Picard Richard
Michalak Sarah
Klamann Richard
Graves Todd
Burr Tom
Hengartner Nicolas
spellingShingle Picard Richard
Michalak Sarah
Klamann Richard
Graves Todd
Burr Tom
Hengartner Nicolas
Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
BMC Medical Informatics and Decision Making
author_facet Picard Richard
Michalak Sarah
Klamann Richard
Graves Todd
Burr Tom
Hengartner Nicolas
author_sort Picard Richard
title Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
title_short Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
title_full Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
title_fullStr Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
title_full_unstemmed Accounting for seasonal patterns in syndromic surveillance data for outbreak detection
title_sort accounting for seasonal patterns in syndromic surveillance data for outbreak detection
publisher BMC
series BMC Medical Informatics and Decision Making
issn 1472-6947
publishDate 2006-12-01
description <p>Abstract</p> <p>Background</p> <p>Syndromic surveillance (SS) can potentially contribute to outbreak detection capability by providing timely, novel data sources. One SS challenge is that some syndrome counts vary with season in a manner that is not identical from year to year.</p> <p>Our goal is to evaluate the impact of inconsistent seasonal effects on performance assessments (false and true positive rates) in the context of detecting anomalous counts in data that exhibit seasonal variation.</p> <p>Methods</p> <p>To evaluate the impact of inconsistent seasonal effects, we injected synthetic outbreaks into real data and into data simulated from each of two models fit to the same real data. Using real respiratory syndrome counts collected in an emergency department from 2/1/94–5/31/03, we varied the length of training data from one to eight years, applied a sequential test to the forecast errors arising from each of eight forecasting methods, and evaluated their detection probabilities (DP) on the basis of 1000 injected synthetic outbreaks. We did the same for each of two corresponding simulated data sets. The less realistic, nonhierarchical model's simulated data set assumed that "one season fits all," meaning that each year's seasonal peak has the same onset, duration, and magnitude. The more realistic simulated data set used a hierarchical model to capture violation of the "one season fits all" assumption.</p> <p>Results</p> <p>This experiment demonstrated optimistic bias in DP estimates for some of the methods when data simulated from the nonhierarchical model was used for DP estimation, thus suggesting that at least for some real data sets and methods, it is not adequate to assume that "one season fits all."</p> <p>Conclusion</p> <p>For the data we analyze, the "one season fits all " assumption is violated, and DP performance claims based on simulated data that assume "one season fits all," for the forecast methods considered, except for moving average methods, tend to be optimistic. Moving average methods based on relatively short amounts of training data are competitive on all three data sets, but are particularly competitive on the real data and on data from the hierarchical model, which are the two data sets that violate the "one season fits all" assumption.</p>
url http://www.biomedcentral.com/1472-6947/6/40
work_keys_str_mv AT picardrichard accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection
AT michalaksarah accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection
AT klamannrichard accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection
AT gravestodd accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection
AT burrtom accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection
AT hengartnernicolas accountingforseasonalpatternsinsyndromicsurveillancedataforoutbreakdetection
_version_ 1725273655216177152