Skip to main content

Event-based internet biosurveillance: relation to epidemiological observation



The World Health Organization (WHO) collects and publishes surveillance data and statistics for select diseases, but traditional methods of gathering such data are time and labor intensive. Event-based biosurveillance, which utilizes a variety of Internet sources, complements traditional surveillance. In this study we assess the reliability of Internet biosurveillance and evaluate disease-specific alert criteria against epidemiological data.


We reviewed and compared WHO epidemiological data and Argus biosurveillance system data for pandemic (H1N1) 2009 (April 2009 – January 2010) from 8 regions and 122 countries to: identify reliable alert criteria among 15 Argus-defined categories; determine the degree of data correlation for disease progression; and assess timeliness of Internet information.


Argus generated a total of 1,580 unique alerts; 5 alert categories generated statistically significant (p < 0.05) correlations with WHO case count data; the sum of these 5 categories was highly correlated with WHO case data (r = 0.81, p < 0.0001), with expected differences observed among the 8 regions. Argus reported first confirmed cases on the same day as WHO for 21 of the first 64 countries reporting cases, and 1 to 16 days (average 1.5 days) ahead of WHO for 42 of those countries.


Confirmed pandemic (H1N1) 2009 cases collected by Argus and WHO methods returned consistent results and confirmed the reliability and timeliness of Internet information. Disease-specific alert criteria provide situational awareness and may serve as proxy indicators to event progression and escalation in lieu of traditional surveillance data; alerts may identify early-warning indicators to another pandemic, preparing the public health community for disease events.


The World Health Organization (WHO) collects and publishes databases of statistics on confirmed and suspected disease outbreaks for select infectious diseases. The 2005 International Health Regulations (IHR), designed to ensure timely recognition of outbreaks of infectious disease with the potential to spread widely, requires WHO member nations to report outbreaks of international concern to the WHO within 24 hours of discovery [13]. Consistent with the IHR, during the initial months of the pandemic (H1N1) 2009 WHO requested that countries report the initial cases and thereafter the number of confirmed cases, and deaths in confirmed cases, for as long as feasible [4]. The WHO published weekly updates of pandemic (H1N1) 2009 case and fatality counts based on this reporting [5]. The resulting database represents one of the most comprehensive and timely outbreak reporting databases available to the public on the Internet.

Event-based biosurveillance, relying primarily on Internet sources, is a recognized approach to infectious disease outbreak detection. It complements traditional approaches to public health surveillance and can provide early warning of emerging events relative to such methods, where data may lag behind due to delays in sample collection, laboratory confirmation, and country reporting. There are several active event-based biosurveillance systems: Project Argus (Argus), Biocaster, Global Public Health Intelligence Network (GPHIN), HealthMap, MedISys, ProMED-mail (ProMED) and others [6, 7]. Event reports are generated by automated machine-based processes for Biocaster, HealthMap and MedISys and written by human analysts or subject matter experts for Argus, GPHIN and ProMED. Manual report examination for relevancy typically occurs post-dissemination for the automated systems (e.g., do articles with the word “virus” in the title refer to a biological infection or an attack on computers?). With the exception of ProMed, which utilizes local observers on the ground for some of its outbreak reporting, event-based biosurveillance systems often disseminate reports that are not observer or laboratory verified (e.g., a cluster of unconfirmed human avian influenza cases in Vietnam). Thus the reports provide near real-time cueing and alerting to users, but they may lack specificity.

The specificity and timeliness of outbreak detection using event-based biosurveillance can be assessed by comparison with epidemiological data from official sources, such as WHO, when available. In general, detecting a new epidemic or outbreak (“signal”) amidst a varying background of disease (“noise”, e.g., normal seasonal influenza or influenza-like-illness) from the vast amount of information available on the Internet is difficult. Moreover, event-based biosurveillance systems can generate a sizable amount of information on any given outbreak topic, sometimes overwhelming users with specific interests. For example, Argus alone generated approximately 22,000 reports on pandemic (H1N1) 2009 from April 2009 to March 2010.

Establishing alert criteria can aid users in identifying relevant and anomalous events from such a large amount of information. Argus and other systems have established semi-automated (pushed via email) and customized (user created) email alerts as a method to improve signal detection, to notify users of emerging events of interest, and to allow for easier tracking of outbreaks or the aftermath of natural disasters.

However, establishing criteria for sending email alerts is complex. The WHO pandemic (H1N1) 2009 data provided a means to assess the timeliness of event-based biosurveillance in real-time and retrospectively, as well as to develop and evaluate alert criteria. In this study, a comparison of WHO epidemiological data and Argus reporting data was made in order to: 1) determine to what extent Argus alerting correlated with the epidemiological disease progression by country and region based on WHO data; 2) identify which alert criteria correlate the best with epidemiological data and provide the most reliable situational awareness; and 3) explore the timeliness of biosurveillance reporting.


Project Argus methodology

Project Argus, hosted at the Georgetown University Medical Center, is designed to report and monitor the evolution of biological events threatening human, plant and animal health globally, excluding the United States (US).[69] Argus collects, in an automated process, several thousand local, native-language Internet media articles daily.[10] Bayesian software tools and Boolean search strings, based on a taxonomy of infectious disease, identify candidate relevant articles. Regional experts, collectively fluent in roughly 40 languages, review these articles manually. Relevant media articles are identified based on direct indicators (reports of disease) and indirect indicators (socially disruptive events or precursors to disease, such as preventative measures or adverse enviro-climatic conditions). Regional experts write Argus reports based on these media articles; reports are posted to a password protected Internet portal for users to view.[11] Argus reported on pandemic (H1N1) 2009 from its identification in April 2009 to the post-pandemic period.[12]

Comparing Argus alert data to WHO case counts

Argus employed email alerts to aid users in monitoring pandemic (H1N1) 2009 as it spread. Alert criteria were developed via an iterative process after assessing the progression of 2009 H1N1 between April 2009 – January 2010 and by monitoring WHO guidelines. Employed from August 2009 to January 2010 (see Table 1), email alerts were meant to capture increasing severity in a locale or region as portrayed in the media and were comprised of direct and indirect indicators[9] of disease. Senior staff reviewed Argus reports and reports meeting alert criteria were extracted. The report metadata and a link to the report on the Argus Internet portal were then emailed to users as a means of notification.

Table 1 Alert criteria for 2009 H1N1 pandemic general descriptions

The reporting alert criteria were reviewed and revised (weekly to monthly) during the course of the pandemic, based on reported feedback from system users. By the end of January 2010, 15 alert categories were in use. Some alert criteria were revised during the study period. For example, before November 2009 the number of healthcare workers, military personnel, and officials reported to be infected were recorded, but afterwards only clusters (> = 3) of cases for these categories were recorded. This change reflected the increased frequency of media reporting of individual cases over time and decreased value of monitoring individual case counts. Another example is that before October 2009 reports of overwhelmed ICUs and ventilator shortages were reported, but afterwards only overall hospital/clinic infrastructure strain or collapse was reported. The analysis for this study was performed based on the definition of the alert as specified during the time period of the analysis (Table 1).

The WHO 2009–2010 H1N1 country case count data was retrieved from the WHO website [5] on March 31, 2010. Argus timeliness could have also been assessed using the date of official confirmed case reporting from public Ministry of Health websites or the date of confirmed case reporting by countries to WHO where these sources of information were available. However, for this study we confined our analysis to WHO data only, because it provided comprehensive official information in one location. Argus weekly alert counts and WHO case counts for each alert category were recorded and plotted over time during a 6 month period from August 2009 to January 2010 (week 32, 2009 to week 5, 2010). The remainder of the analysis was performed during a 4 month period from October 2009 to January 2010 (week 41, 2009 to week 5, 2010) during which time 13 of the 15 alert criteria definitions were finalized. (Alert 1 was created and Alert 3 was modified during the first 4 weeks of the study period.) WHO weekly case updates were recorded by country. A time series of all Argus alert data combined was plotted against all WHO case counts from October 2009 to January 2010 (week 41, 2009 to week 5, 2010). The WHO case count and Argus alert counts were normalized between 0 and 1 prior to plotting as follows: count_j/Max{count_j}. Data were only used from countries covered by both WHO and Argus to provide an unbiased comparison (e.g. data from the US was not used since Argus does not cover the US).

A Pearson correlation matrix was generated for all Argus alerts to assess correlation with the data from WHO, the degree to which the alerts are related. Pearson correlation coefficients and corresponding p-values were generated for the combined alert data as well as for each alert individually in comparison to overall WHO case count data. A time series of alerts with WHO case data was plotted for all significant (p < 0.05) correlations.

Eight geographical regions were defined for purposes of our analysis: Africa; East Asia; Europe; Latin America and Caribbean; Middle East; Russia and Central Asia; South Asia; Southeast Asia, Oceania, and Canada. Nations in the WHO dataset were also assigned to these regions. Pearson correlation coefficients and corresponding p-values were generated for the total alert counts and WHO case counts by region.

All statistics were computed using R Version 2.11.0 [13].

First confirmed pandemic (H1N1) 2009 case comparison

Reports of the first confirmed 2009 H1N1 influenza cases were recorded for Argus and/or WHO from April 24, 2009 to June 1, 2009. This time period was chosen due to the international focus on tracking pandemic (H1N1) 2009 cases spread in the initial months of the pandemic and the availability of daily pandemic (H1N1) 2009 WHO situation updates until June 1, 2009. The daily situation updates included WHO pandemic (H1N1) 2009 case data and were available on the WHO Global Alert and Response (GAR) website during this time period [14]. The Argus Internet portal was also monitored daily for media reports of confirmed cases in new countries. Argus timeliness was assessed using the date of first official confirmed case as reported by the WHO relative to the date of first case detected by Argus for a given country from Internet sources during the initial phases of the pandemic.

Study period

The overall study period was from April 2009 to January 2010. August 2009 to January 2010 is the time period used for the overall comparison of WHO case counts and Argus alert counts by week (Figure 1). October 2009 to January 2010 is the time period utilized for the comparison of Argus alerts compared to WHO case data after the alert criteria definitions were nearly finalized (Figures 2, 3, 4, 5, 6, 7, Table 2, Table 3). April 24, 2009 to June 1, 2009 is the study period utilized for the comparison of first confirmed cases for Argus and WHO (Table 4).

Figure 1

Argus alerts and WHO case counts by week. Y axes are all fixed to the same scale in this graph. Note that December 24th 2009 to January 3rd 2010 (week 52, 2009 to week 1, 2010) was a holiday period for Project Argus with substantially reduced reporting volume.

Figure 2

WHO normalized case counts compared to all Argus alerts pooled.

Figure 3

Plots of Argus alert 1 (large increase in case count), 3 (confirmed cases or fatalities in healthcare workers, military personnel and/or national officials), 6 (hospital/clinic infrastructure strain or collapse), 13 (health policy change) and 15 (border closures) versus WHO normalized case counts.

Figure 4

WHO normalized case counts compared to Argus alerts 1, 3, 6, 13 and 15 combined.

Figure 5

WHO normalized case counts compared to Argus alerts 1, 3, and 15 combined.

Figure 6

WHO normalized case counts compared to Argus alerts 3, 6, and 15 combined.

Figure 7

WHO normalized case counts compared to Argus alerts 3, 13 and 15 combined.

Table 2 Correlation coefficients for Argus alerts versus WHO case counts for October 2009 (week 41 2009) to January 2010 (week 5 2010) *
Table 3 WHO case study comparison with Argus alerts by Region*
Table 4 WHO/Argus 1 st date of confirmed 2009 H1N1 case by country


Alert data compared to WHO case counts

Using the alert criteria in Table 1, a total of 1,580 alerts were recorded from 1,499 Argus reports covering 122 countries in the 6 month time period of August 2009 to January 2010 (week 32, 2009 to week 5, 2010). Note that multiple alerts were generated for some Argus reports. WHO recorded 244,196 pandemic (H1N1) 2009 cases during this time period (Figure 1).

Plots of WHO case data and all Argus alerts pooled from October 2009 to January 2010 (week 41, 2009 to week 5, 2010) appear in Figure 2. During this time period, alert categories 1 (i.e. increase in case counts), 3 (i.e. clusters of cases or fatalities of health care workers, military personnel or national officials), 6 (i.e. healthcare facility strain or collapse), 13 (i.e. health policy change) and 15 (i.e. border closure) generated statistically significant (p < 0.05) correlations with WHO case count data with correlation coefficients (r) of 0.80, 0.62, 0.76, 0.67, and 0.58 respectively (Table 2). The time series of the sum of alerts 1, 3, 6, 13 and 15 was highly correlated with WHO case data (r = 0.81, p < 0.0001). The plots for these alerts are illustrated in Figures 3 and 4. Argus alerts 1, 6 and 13 are significantly correlated (alerts 1 and 6: r = 0.87, p < 0.0001; alerts 1 and 13: r=0.80, p<0.0001), alerts 6 and 13: r = 0.81, p < 0.0001). Figures 5, 6, 7 illustrate the combined plots with alerts 3, 15 and 1 or 6 or 13. No reports met the alert 14 (food safety) criteria.

Data was present in both WHO and Argus for 49 countries in the 8 regions. Data from Germany, Portugal, Canada, and Brazil was not available on the WHO website as of March 31, 2010. Alert time series with significant correlation to WHO case data were 1, 3, 5 (i.e. severe manifestation, co-infection or re-infection), 6, 11 (i.e. vaccine failure, severe reaction, or black market sales) and 13 (Table 3).

First confirmed pandemic (H1N1) 2009 case comparison

Argus and WHO collectively identified 64 countries with confirmed cases of pandemic (H1N1) 2009 from May 8, 2009 to June 1, 2009 (Table 4). Argus reported the first confirmed case on the same day as WHO for 21 of the 64 countries. Argus reported from 1 to 16 days ahead of WHO for 42 countries: 1 day ahead for 22 countries; 2 days ahead for 8 countries; 3 days ahead for 8 countries, 4 days ahead for 1 country (Costa Rica) and 5 days ahead for 1 country (India). Two countries were identified by Argus only during the study period. Egypt was identified by WHO on June 3 and United Arab Emirates was identified by WHO on June 8, 16 days and 14 days after Argus, respectively. One country was identified by WHO only during the study period, Bahamas, and was reported by Argus 2 days later. Note that the first case in Egypt was identified by Argus on May 18, but did not appear in the sources monitored again until after the study period on June 2. Both dates are recorded in Table 4.


As the media coverage intensifies during the course of a high profile event, such as pandemic (H1N1) 2009, establishing alert criteria can help guide users of Internet based biosurveillance systems. In this study, alert categories 1 (i.e. increase in case counts), 3 (i.e. cases or fatalities of health care workers, military personnel and/or national officials), 6 (i.e. healthcare facility strain or collapse), 13 (i.e. health policy change) and 15 (i.e. border closure) were significantly correlated with the WHO confirmed case count in the four month study period (October 2009 to January 2010 (week 41, 2009 to week 5, 2010). Thus, alerts targeting direct indicators (Alerts 1 and 3) and indirect indicators (Alerts 6, 13, and 15) provided situational awareness during the pandemic.

Increase in case counts (Alert 1) is the most similar alert category to WHO case count data. The significant correlation suggests that reports of confirmed cases in the media are consistent with confirmed cases identified through public health surveillance and testing. A rising number of cases or fatalities of health care workers, military personnel or national officials (Alert 3), who are often more aware of prevention measures than the general public, is an indication of an emerging or escalating infectious disease outbreak, consistent with a rise in case counts. Health care facility strain or collapse (Alert 6) is an indirect indicator of increasing case counts and/or increase in health care worker cases or fatalities.

Though only six alerts were generated for border closure (Alert 15), it is not surprising that the alert is correlated with WHO case data considering the severity of an event that would warrant such an action. Similarly, massive release of anti-virals or vaccine stockpiles (Alert 12) indicates a severe escalation or perceived escalation in cases or deaths. This alert did not reach significance, however, likely because only two alerts were generated for this category. Alerts 1, 6 and 13 are also correlated with each other and maintain a highly significant correlation with WHO case counts when compared individually to WHO case counts along with alerts 3 and 15. An increase in case counts would lead to healthcare infrastructure strain and health policy change, likely accounting for the intra-alert correlation. Comparison of alerts in pandemic versus non-pandemic years is required for verification; however, this study suggests that Alerts 1, 3, 6, 13 and 15 may all serve as proxy indicators in the media of an emerging or escalating event on the ground and could serve as surveillance measures in conjunction with public health surveillance for a future pandemic.

The other alerts may not have been significantly correlated with WHO case counts due to the relatively mild manifestation of the pandemic (H1N1) 2009 without a virulent secondary wave or changes in transmission patterns.[15] Though reports of atypical clinical manifestations, transmission to other species, anti-viral resistance[16] and failure and viral mutations were prevalent in the media, such mechanisms appear to have not contributed to a significant escalation in case count.[15] These alerts, however, could serve as potential indicators for a future pandemic. A large increase in fatalities (Alert 2) was borderline significant with only 9 alerts generated. Again this is likely due to the mild nature of the pandemic, with an estimated 12,000 deaths, compared to previous pandemics, 1918, 1957 or 1968, with estimated attributable mortality of 50 to 100 million, 1–2 million, and 1 million, respectively.[15, 17, 18]

Correlation analysis by region showed some variation in the significant alerts as was expected based on the differences in severity of the pandemic, capacity for disease detection and capability for response for each region. Alert 5 (i.e. severe manifestation, co-infection or re-infection) and alert 11 (i.e. vaccine failure, severe reaction, or black market sales) emerged as significantly correlated to WHO case counts in Europe and in Europe, South Asia-Canada-Oceana, respectively, though they were not significant when global WHO case counts were considered. These results suggest that regional differences in the evolution of the pandemic are important to consider when developing alert criteria. Alerts 1, 3, 6 and 13 were each significant in one or more regions, which further supports their appropriateness for global surveillance.

Utilizing Internet media sources, Argus identified the first cases of confirmed pandemic (H1N1) 2009 published on the Internet an average of 1.5 days ahead of WHO official reporting (range 1 to 16 days) for all 64 non-US countries reporting by June 1, 2009. This was expected since information from Internet media reports is often timelier than the official reporting of cases to the public after laboratory confirmation. Though in this case the lead-time may be only a few days, this study provides evidence of the validity of using event-based biosurveillance for detecting emerging biological events.

This study had limitations. The alert criteria evolved from initiation in August 2009 through November 2009. However, the study period chosen for the majority of the analysis was after October in order to mitigate any bias from changing alert criteria. In addition, the alert criteria changes were small, geared toward making the alert criteria more specific and did not significantly impact the results (data not shown). In event-based biosurveillance studies there is often a lack of robust gold standard official comparison data. WHO data can be limited by delays in country reporting and under-reporting, however for the 2009 pandemic WHO was considered a timely and accurate source of global data [19]. Finally, the study had a restricted time window. Fears of a virulent resurgence of the virus in a second wave were unfounded and when WHO case counts and Argus alerts decreased to low levels in January 2010, the study was ended. Nonetheless, sufficient data was collected to identify significant indicators of the evolving pandemic.

The pandemic (H1N1) 2009 was of global significance and a main focus of local, national and international public health organizations, particularly during the initial phase. However, there are numerous human, animal and plant diseases that are economically important but are not normally tracked by public health organizations, suggesting that Internet surveillance of such diseases could provide lead-time of an outbreak compared to traditional methods [20]. When surveillance for indirect indicators (suspected cases or prevention measures) is performed in addition to direct reports of disease, the lead-time often increases further.[8, 21] Surveillance of pandemic (H1N1) 2009 serves as an example of the real-time capability of identifying emerging disease events in general, particularly events that may be evident in local media in the regional vernacular.

Other event-based biosurveillance systems have demonstrated the effectiveness of extracting relevant information from Internet media sources as a means for detecting and monitoring disease events.[21] Internet media reporting provides an emerging resource for early detection of new events and for providing situational awareness of evolving events, particularly when official sources may not be available. Alerts based on media reports can provide event situational awareness and cue users of shifts in infectious disease trends. As the number of online news media sources, including social media sources with user-generated content, continues to expand, event-based biosurveillance will play an increasingly important role in disease surveillance. On-going validation and verification of event-based biosurveillance methods with epidemiological and clinical data by users and surveillance system developers will increase the robustness of this approach for detecting and tracking emerging events.

Authors' information

NN holds a MD in International Health and PhD in Epidemiology and is a Research Assistant Professor in the department of Pediatrics at Georgetown University Medical Center. LY is a biostatistician and holds a Masters degree in epidemiology and biostatistics. AR holds a Masters degree in Microbiology and Immunology; she is also an ISDS member and has worked for over four years in the biosurveillance field. JH holds a Masters degree in Emerging Infectious Diseases and Biohazardous Threat Agents and has worked over four years in the biosurveillance field. DH is Research Associate Professor in the department of Microbiology and Immunology at Georgetown University Medical Center; his research applies mathematical modeling and related methodologies to understand the dynamics of disease in human and animal populations.


  1. 1.

    World Health Organization: International Health Regulations (2005). 2008, [cited 2009 May 21]. Geneva: The Organization, 2.

    Google Scholar 

  2. 2.

    Baker MG, Forsyth AM: The new International Health Regulations: a revolutionary change in global health security. N Z Med J. 2007 Dec 14, 120 (1267): U2872.

    PubMed  Google Scholar 

  3. 3.

    Wilson K, McDougall C, Forster A: The responsibility of healthcare institutions to protect global health security. Healthc Q. 2008, 12 (1): 56-60.

    Article  Google Scholar 

  4. 4.

    World Health Organization: Global Alert Response. Human infection with pandemic (H1N1) 2009 virus: updated interim WHO guidance on global surveillance. 10 July 2009, (originally posted on 29 April 2009).

  5. 5.

    WHO FluNet

  6. 6.

    Hartley DM, Nelson NP, Walters R, Arthur R, Yangarber R, Madoff L, et al: The landscape of international event-based biosurveillance. Emerging Health Threats Journal. 2009, 3: e3-10.3134/ehtj.10.003. 10.3134/ehtj.10.003

    Google Scholar 

  7. 7.

    Walters R, Harlan P, Nelson NP, Hartley DM: Data sources for biosurveillance. Wiley handbook of science and technology for homeland security. Edited by: Voeller J. New York: John Wiley & Sons; 2010, 1-17.

    Google Scholar 

  8. 8.

    Nelson NP, Brownstein JS, Hartley DM: Event-based biosurveillance of respiratory disease in Mexico, 2007–2009: connection to the 2009 influenza A(H1N1) pandemic? Euro Surveill 2010, 15 (30): 19626.

    PubMed  Google Scholar 

  9. 9.

    Wilson JM, Polyak MG, Blake JW, Collmann J: A heuristic indication and warning staging model for detection and assessment of biological events. JAMIA. 2008, 15: 158-171.

    PubMed Central  PubMed  Google Scholar 

  10. 10.

    Jeff Collmann, Adam Robinson: Designing Ethical Practice in Biosurveillance The Project Argus Doctrine. Infectious Disease Informatics and Biosurveillance, series: Integrated Series in Information Systems, Volume 27. Edited by: Zeng D, Chen H, Castillo-Chavez C, Lober WB, Thurmond M. New York: Springer;2011, 23-44.

    Google Scholar 

  11. 11.

    Open Source Center.

  12. 12.

    Centers for Disease Control and Prevention (CDC): Swine influenza A (H1N1) infection in two children--southern California, March-April 2009. MMWR Morb Mortal Wkly Rep. 2009, 58 (15): 400-402. http://review/mmwrhtml/mm5815a5.htm

  13. 13.

    R Development Core Team: A language and environment for statistical computing. Vienna: Foundation for Statistical Computing; 2010.

    Google Scholar 

  14. 14.

    World Health Organization: Global Alert Response. Situation updates - Pandemic (H1N1). 2009.

  15. 15.

    Pada S, Tambyah PA: Overview/reflections on the 2009 H1N1 pandemic. Microbes Infect. 2011 May, 13 (5): 470-478. 10.1016/j.micinf.2011.01.009. Epub 2011 Jan 27.

    Article  PubMed  Google Scholar 

  16. 16.

    Hartley DM, Nelson NP, Eli N: Perencevich. Antiviral Drugs for Treatment of Patients Infected with Pandemic (H1N1) 2009. Emerg Infect Dis. 2000, 15: November 2009.

  17. 17.

    Cox NJ, Subbarao K: Global epidemiology of influenza: past and present. Annu Rev Med. 2000, 51: 407-421. Review. 10.1146/

    CAS  Article  PubMed  Google Scholar 

  18. 18.

    Taubenberger JK, Morens DM: 1918 Influenza: the mother of all pandemics. Emerg Infect Dis. 2006 Jan, 12 (1): 15-22.

    PubMed Central  Article  PubMed  Google Scholar 

  19. 19.

    Chan EH, Brewer TF, Madoff LC, Pollack MP, Sonricker AL, Keller M, Freifeld CC, Blench M, Mawudeku A, Brownstein JS: Global capacity for emerging infectious disease detection. Proc Natl Acad Sci U S A. 2010 Dec 14, 107 (50): 21701-21706. 10.1073/pnas.1006219107. Epub 2010 Nov 29.

    PubMed Central  CAS  Article  PubMed  Google Scholar 

  20. 20.

    Thomas CS, Nelson NP, Jahn GC, Tianchan N, Hartley DM: Use of media and public-domain Internet sources for detection and assessment of plant health threats. Emerging Health Threats Journal. 2011, 4: 7157-10.3402/ehtj.v4i0.7157. 10.3402/ehtj.v4i0.7157

    PubMed  Google Scholar 

  21. 21.

    Brownstein JS, Freifeld CC, Chan EH, Keller M, Sonricker AL, Mekaru SR, Buckeridge DL: Information technology and global surveillance of cases of 2009 H1N1 influenza. N Engl J Med. 2010 May 6, 362 (18): 1731-1735. 10.1056/NEJMsr1002707.

    PubMed Central  CAS  Article  PubMed  Google Scholar 

Download references


We acknowledge Sarah Riedl and Emily Iarocci for their work in establishing and revising the Argus alert criteria and Pahini Pandya for her work in tracking the first 2009 H1N1 pandemic cases by country. We also acknowledge the Argus staff at Georgetown University Medical Center for producing the reports that made this study possible. We thank Ronald A. Walters for his critical review of the manuscript and thoughtful comments.

Author information



Corresponding author

Correspondence to Noele P Nelson.

Additional information

Competing interests

The authors have no financial or non-financial competing interests to declare.

Authors’ contributions

NN drafted the manuscript and co-conceived of the research topic. LY performed the statistical analysis and helped to draft the manuscript. AR participated in the design of the study and helped draft the manuscript. JH participated in the design of the study and helped draft the manuscript. DH drafted the manuscript and co-conceived of the research topic. All authors read and approved the final manuscript.

Authors’ original submitted files for images

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and Permissions

About this article

Cite this article

Nelson, N.P., Yang, L., Reilly, A.R. et al. Event-based internet biosurveillance: relation to epidemiological observation. Emerg Themes Epidemiol 9, 4 (2012).

Download citation


  • Biosurveillance
  • Infectious disease
  • Epidemiology
  • Disease-specific alerts
  • Internet media
  • Early warning
  • Pandemic (H1N1) 2009
  • Situational awareness
  • Outbreak detection