In the weeks following the first imported case of Ebola in the U. S. on September 29, 2014, coverage of the very limited outbreak dominated the news media, in a manner quite disproportionate to the actual threat to national public health; by the end of October, 2014, there were only four laboratory confirmed cases of Ebola in the entire nation. Public interest in these events was high, as reflected in the millions of Ebola-related Internet searches and tweets performed in the month following the first confirmed case. Use of trending Internet searches and tweets has been proposed in the past for real-time prediction of outbreaks (a field referred to as “digital epidemiology”), but accounting for the biases of public panic has been problematic. In the case of the limited U. S. Ebola outbreak, we know that the Ebola-related searches and tweets originating the U. S. during the outbreak were due only to public interest or panic, providing an unprecedented means to determine how these dynamics affect such data, and how news media may be driving these trends.
We examine daily Ebola-related Internet search and Twitter data in the U. S. during the six week period ending Oct 31, 2014. TV news coverage data were obtained from the daily number of Ebola-related news videos appearing on two major news networks. We fit the parameters of a mathematical contagion model to the data to determine if the news coverage was a significant factor in the temporal patterns in Ebola-related Internet and Twitter data.
We find significant evidence of contagion, with each Ebola-related news video inspiring tens of thousands of Ebola-related tweets and Internet searches. Between 65% to 76% of the variance in all samples is described by the news media contagion model.
Citation: Towers S, Afzal S, Bernal G, Bliss N, Brown S, Espinoza B, et al. (2015) Mass Media and the Contagion of Fear: The Case of Ebola in America. PLoS ONE 10(6): e0129179. doi:10.1371/journal.pone.0129179
Academic Editor: Christos A. Ouzounis, Hellas, GREECE
Received: November 17, 2014; Accepted: May 5, 2015; Published: June 11, 2015
Copyright: © 2015 Towers et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited
Data Availability: All relevant data are within the paper and its Supporting Information files.
Funding: The parameter optimization portion of the analysis was made possible with education allocation #DMS140043 of super-computing resources from National Science Foundation Extreme Science and Engineering Discovery Environment (XSEDE) high-performance computing initiative (ST GB NB SB BE JJ JJG MK ML RM VMM FN KO MLR CR CCC). This research was also partially supported by the Western Alliance to Expand Student Opportunities (WAESO) Louis Stokes Alliance for Minority Participation (LSAMP) Bridge to the Doctorate (BD) National Science Foundation (NSF) "Multidisciplinary STEM Solutions LSAMP Bridge to the Doctorate" grant #HRD-1401190 (GB SB JJ JJG MLR CR), and the Offices of the President and Provost of Arizona State University. This work was funded in part by the U.S. Department of Homeland Security (DHS) VACCINE Center award #2009-ST-061-CI0001-06 (SA DE), and was also made possible by grant #1R01GM100471-01 from the National Institute of General Medical Sciences (NIGMS) at the National Institutes of Health (CCC). The contents of this manuscript are solely the responsibility of the authors and do not necessarily represent the official views of the funding agencies. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Use of Google search trend data to predict disease outbreaks was first proposed in 2006 , and subsequently developed further in the years prior to the 2009 A/H1N1 outbreak [2, 3]. This resulted in the advent of the Google Flu Trends online application, which uses a combination of flu-related Internet search activity to predict, in real-time, flu outbreaks in the U. S. and other countries worldwide (see http://www.google.org/flutrends/, accessed February 9, 2015).
Twitter is a social networking and micro-blogging service that enables its millions of users to send and read each other’s tweets, consisting of short, 140-character messages. The service currently has over 280 million monthly active users, sending on average around 500 million tweets per day (see http://about.twitter.com/company, accessed February 9, 2015). Twitter was first proposed as a means to track disease in real-time in a 2011 publication that looked retrospectively at flu-related tweets during the 2009 A/H1N1 pandemic .
Since these initial analyses, the field of “digital epidemiology” has rapidly expanded [5–8]. For example, Twitter data has been used to examine cholera outbreaks , vaccination sentiments , and proposed for use for global infectious disease surveillance . Google search data has been used for real-time forecasting of Dengue outbreaks [12, 13], Listeria outbreaks , the spread of tuberculosis , consumer behavior , and the economy .
However, there have been some failings of these methods, leading to some concern about their applicability in an outbreak of an emerging infectious disease. At the beginning of the 2009 A/H1N1 outbreak, Google Flu Trends was not accurate in predicting the progression of the pandemic, due to information seeking behavior that was likely induced by mass media coverage . Indeed, Chew et al (2009) found that some peaks in flu-related Twitter activity during the 2009 A/H1N1 pandemic were correlated to the timing of major news stories about the pandemic . Additionally, during the severe 2012–13 influenza season, Google Flu Trends did not accurately predict the course of the outbreak, seriously overestimating the disease burden in the population. It has been suggested that the problems may have been due to widespread media coverage, including the declaration of a public-health emergency by New York state . It has also been noted that news media is effective in “pulsing” of public opinion during election times , as expressed on social media.
Despite the fact that evidence of the influence of news media on Twitter and Internet search patterns has been repeatedly observed, up until now it has been extraordinarily difficult to correct for this effect because the patterns are also intertwined with temporal patterns due to the other dynamics of interest, such as the actual spread of a disease within a population.
During the recent very limited U. S. Ebola outbreak, the popularity of Ebola-related Google searches in the U. S. rivaled that of flu-related searches during the 2009 A/H1N1 pandemic (see Fig 1), but we can be confident that none (or virtually none) of the Ebola-related U. S. Internet searches or tweets arose from actual victims of Ebola in the U. S. The situation thus provides an excellent means to determine how public interest, curiosity, or panic regarding a certain topic affects social media and Internet search dynamics, and allows us to examine the influence of news media on these trends. The results of this study will thus help to inform future digital epidemiological analyses of outbreak data, possibly allowing for the correction of the effects of benign interest, information-seeking behavior, or public hysteria.
The relative interest in Ebola-related searches during the month of October 2014 rivaled the flu-related searches at the beginning of the A/H1N1 pandemic.
In this analysis we employ a mathematical model of contagion to simulate the potential influence of Ebola-related news videos on peoples’ tendency to perform Ebola-related Internet searches or tweets. The advantage of such a model over the use of a simple statistical regression model to explore these relationships is that a mathematical model can incorporate non-linear dynamics related to people becoming immune to news media induced interest in a topic (i.e. the model incorporates the effect of eventual boredom with a topic, no matter how many news videos are subsequently aired). For comparison purposes, however, we also present the results of statistical linear regression models examining these relationships.
As a cross-check to our analysis, we verify that public interest in Ebola (as reflected by Internet searches and tweets) does not appear to induce news media stories on the topic. We also show that peaks and valleys in the temporal trends in news media coverage tend to precede the same features in the Internet and Twitter data, not the other way around.
In the following sections we describe the sources of data used in this analysis, and describe the contagion model employed, followed by a presentation of results and discussion.
Materials and Methods
Google search data.
Daily trends in Ebola-related Google searches were obtained from Google Trends application program interface (API) (http://www.google.com/trends). The search terms examined were:
- “Ebola symptoms”
- “Do I have Ebola”
The data provided by Google Trends are normalized to the total search volume on Google, and thus the data are relative and not absolute. To estimate the total number of searches, we obtained the total number of searches per month originating in the U. S. for each search term from the Google Adwords subscription service, via the Keyword Planner tool, and scaled the Google Trend data accordingly.
The data are shown in Fig 2.
The samples consist of six weeks of data ending October 31st, 2014. The first case of Ebola confirmed in the U. S. occurred on September 29, 2014. The temporal trends in the data samples are highly inter-correlated, with a minimum of 70% correlation between samples.
Daily Twitter data were obtained from a repository continuously collected by a server maintained by the Purdue center for Visual Analytics for Command, Control, and Interoperability Environments (VACCINE), which uses the Twitter streaming Application Programming Interface (API) to access up to 1% of the global stream of tweets.
From this stream, we selected tweets during the six week period ending October 31, 2014 containing the keyword “Ebola”, or the keywords “Ebola” and “symptoms”. Only English language tweets geo-located in the United States were considered, and retweets were removed from the sample, resulting in a total of 260947 tweets.
Potential bias in the temporal trends due to tweets coming from news agencies, rather than regular (aka “grassroots”) Twitter users needs to be minimized. We thus separated the tweets coming from obvious news agencies, public entities, or reporters; CNN, NBC, ABC, CBS, FOX, MSNBC, New York Times, Washington Post, Reuters, Bloomberg News, New Day, U.N. Spokesperson, and Geraldo Rivera. We also excluded tweets coming from users with usernames that included the words “TV”, “media”, “trend”, “CP” (which often stands for “City Press”), and “report” (all selections were applied such that they were not sensitive to the case of the characters in the username). This yielded a sample of 10224 tweets.
As noted in Reference , the Twitter accounts of news agencies compared to the accounts of grassroots users tend to have a much smaller indegree to outdegree ratio (i.e. they usually follow far fewer accounts than they have followers). Reference  found that this ratio was on average 0.5 for news agencies (with median 0.001), and between 1.1 to 2.6 for non-media Twitter users (with median between 0.8 to 1.6). For our particular sample, we find this ratio is 0.34 for news agency users (with median 0.007), and 1.3 for the remaining users (with median 0.95). Thus, based on both the user name, and the properties of the user account, the remaining 250723 tweets are consistent with having not come from news agencies. These remaining data are shown in Fig 2.
News video data.
According to a 2013 Gallup poll of over 2,000 people, 55% of Americans turn to television to obtain their news , with the next most popular medium being the Internet (22%). According to another study done in 2014 by the Pew Research center , CNN, MSNBC, and Fox News are the three primary cable news networks, with Fox News and MSNBC having an 80% share of that market.
While CNN makes available online a repository of its very recent videos, they are not reliably identifiable by date. However, both Fox News and MSNBC provide tools on their websites to obtain videos related to specific topics within specific date ranges. These videos are aired both on television, and also available for online viewing. We use these data as a proxy for the temporal trends in the total amount of news media coverage related to Ebola.
From these repositories, we downloaded the daily number of Ebola-related videos for the six week period ending October 31st, 2014. The data are shown in Fig 2.
The Granger causality test is a hypothesis test for determining whether or not one time series is able to forecast another . Granger causality is defined on the principle that a cause necessarily must happen prior to its effect, and that the cause contributes unique information regarding the future value of its effect.
A time series X is said to Granger-cause Y if it can be shown that X provides statistically significant information about the future values of Y. This is achieved through F-tests based on linear regression of Y on lagged values of X (including also lagged values of Y). Here, we examine the potential Granger causality of temporal patterns in the daily number of news videos on temporal patterns in Ebola-related tweets and Internet searches the following day.
Significant Granger-causality provides evidence that X may in fact cause Y (although the caveat that correlation does not equal causation applies). Note, however, that in the case that X does indeed cause Y, the time lag between cause and effect must not be too small relative to the time step of the time series in order for the effect to be detectable with the Granger-causality test.
To assess the potential contagion of Ebola-related news media in inspiring Ebola-related Google searches or tweets, we employ a Susceptible, Infected, Recovered, Susceptible, and Vector (SIRSV) compartmental model. News videos, V, can “infect” susceptible individuals with interest in the topic, that is subsequently evidenced by Internet searches or tweets. After an average period of 1/γ days, the individual performs an Internet search or tweet related to the topic, and then “recovers”, subsequently becoming less interested with the topic and no longer feeling a need to perform another Internet search or tweet. The model includes the potential that recovered individuals can flow back into the susceptible compartment (i.e. after a period of time the individual may once again become susceptible to being infected with interest in the topic).
The equations of the model are (1) where N = S + I + R is the U. S. population, and V(t) is the temporal evolution of the number of news videos per day. The parameter β is the number of Internet searches or tweets per unit time inspired by one news video shown to a completely susceptible population. The parameter μ represents the rate of searches inspired by things other than news media, such as people talking among themselves about a particular subject. The parameter α represents the rate at which recovered individuals flow back in to the susceptible compartment.
Our model is somewhat similar to the model used in Reference  to simulate the spread of ideas within a population, with the difference that we include an exogenous infection component, V, rather than assuming that contagion is only due to intrinsic infectious process within the closed population.
The system of equations in Eq 1 is numerically solved to obtain the estimated number Internet searches or tweets made per unit time, γI(t). The number of news videos varies quite dramatically by day, making numerical solutions unreliable due to stiffness of the model equations if the raw data for the daily number of news videos, V(t), are used; we thus first fit a cubic spline to the daily news video data in order to smoothly interpolate the data in finer time steps.
To estimate the model parameters that optimally describe each data sample, we use the Monte Carlo method for solution of inverse problems to randomly sample the parameters of the model from broad uniform distributions, and calculate the Pearson χ2 goodness-of-fit statistic comparing the model to the data sample [27, 28]. The uniform distribution sampling range is chosen to be large enough to ultimately include the parameter optimal value and at least a ±5 standard deviation range about that value. In order to determine the parameter hypotheses that minimize the Pearson χ2 goodness-of-fit statistic, this procedure is repeated at least one million times for each sample with the use of National Science Foundation Extreme Science and Engineering Discovery Environment (NSF XSEDE) high-throughput computing resources (see www.xsede.org, accessed February 9, 2015). To determine the parameter 95% confidence intervals, the Pearson χ2 statistic is corrected for over-dispersion using the ansatz of McCullagh and Nelder (1989) .
If the parameter optimization procedure determines that μ and α are statistically consistent with zero (i.e. that there is no evidence of contagion within the population due to dynamics other than exposure to news videos, and immunity upon recovery is permanent), a more appropriate model is (2) In addition, if the parameter optimization procedure determines that μ, α, and 1/γ are statistically consistent with zero, the most appropriate model is (3) where the move to the recovered class happens essentially immediately after exposure to a news video. Given that our data are aggregated by day, “immediately” in this case essentially implies a time frame smaller than a day.
Note that initially the entire population is not typically susceptible to ideation to do an Ebola-related Internet search or tweet due to exposure to an Ebola-related news story aired by Fox News or MSNBC; the people considered “susceptible” in this analysis are people who watch those networks, have access to a computer, and are also susceptible to various levels of ideated panic, ranging from simple interest or curiosity, to actually wondering if they themselves might have Ebola. In the parameter estimation procedure, we thus also estimate the initial fraction of susceptible people in the population for each kind of Ebola-related tweet or Internet search.
As a validation cross-check of the model, we optimize the model parameters to the first half of the time series, and use the resulting model to predict the temporal patterns of the remaining half. We also perform this model validation procedure for a simple linear regression fit, where the time series of the number of Internet searches and tweets are regressed on V. Unlike the mathematical model, the linear regression model does not include the effect of eventual boredom with a topic.
As an additional cross-check of the analysis, we swap I and V in the model, and examine the potential that Ebola-related Internet searches or tweets inspire the temporal patterns in news media coverage, rather than the other way around.
Results and Discussion
For all data samples considered in this analysis, we find that the contagion model parameter μ is statistically consistent with zero (i.e. we find that there is no statistically significant evidence of contagion due to effects other than news videos). We also find that the parameter α is statistically consistent with zero for all samples (i.e. there is no statistically significant evidence that people return to the susceptible class after recovery). In addition, for the Twitter data samples, 1/γ is statistically consistent with zero, indicating that after viewing a news video the movement to the recovered compartment after doing an Ebola-related tweet occurs within a day. For these data samples, we thus fit the reduced contagion model of Eq 3.
In Table 1, we show the best-fit parameters of the contagion model fit to the data samples. In Table 2, we show the percentage of the variance, R2, in the data samples described by the best-fit contagion model. We also show the results of the model validation procedure, showing the R2 of the model fit to the first half of each data sample, and the R2 of the subsequent model prediction extrapolated to the last half of the sample. The R2 of a simple statistical model, linearly regressing the data samples on the daily number of Ebola-related news videos, is also shown. In all cases, the contagion model has better predictive power than the linear regression model.
As a cross-check, we also use the contagion model to examine the possibility that the Ebola-related Internet searches or tweets inspire news videos on the topic. The results are summarized in Table 3. In all cases, we find that the R2 is much worse for this model compared to the model where we assume that the news videos inspire the Internet searches and tweets.
Also shown in Table 3 are the p-values of the Granger causality test that tests the hypothesis that temporal patterns in news videos Granger-cause temporal patterns in the Internet searches and tweets on the next day, and vice versa. In all cases, there is no statistically significant evidence that Ebola-related Internet searches and tweets Granger-cause temporal patterns in Ebola-related news videos, but there is evidence in several cases that the reverse is true.
Fig 3 shows the best-fit contagion and linear regression models overlaid on the data.
The fits of the linear regression model (shown in blue) tend to be generally too low in the beginning and too high at the end. In contrast, the contagion model (red line) accounts for the boredom effect, where people become more and more disinclined to perform Ebola-related searches or tweets after an extended period of exposure to Ebola-related news-coverage. Incorporation of this dynamic in the model yields significantly better fits to the data compared to the regression model.
As shown in Table 2, we find that a large fraction of the variance in the data is described by the contagion model (R2 at least 65% in all cases), and the model validation procedure shows good predictive capability for the extrapolated model to the test samples. The performance of a simple linear regression model is not as good; for all but one sample, the linear regression model yields a lower R2 than the contagion model, and the model validation procedure reveals that the regression model has generally much poorer predictive capabilities compared to the contagion model. It is thus apparent that in modeling the dynamics of ideation due to exposure to news media, “recovery” and ultimate immunity to further ideation should be taken into account.
As shown in the best-fit model parameters Table 1, we find that a relatively large percentage of the population (up to 18%) is susceptible to ideation with what likely is simple curiosity prompting a Google search for “Ebola”, and we find that one Ebola-related news video aired by Fox News or MSNBC on average inspires tens of thousands of such searches in the American population. In contrast, a much smaller percentage of the population appears to be susceptible to ideation to perform more specific searches indicating personal concern about actually having the disease, such as Google searches for “do I have Ebola”.
The contagion model we employ assumes that people recover once performing an Ebola-related Internet search or tweeting Ebola-related information, and do not feel the need to seek or disseminate information again. Through comparison of the fit of a contagion model to the fit of a simple linear regression model that does not include the dynamics of the recovery process, we find evidence that indeed recovery and immunity to further ideation does play a role in the overall dynamics of peoples’ information seeking behavior. This is concordance with the conclusions of a previous study which examined the information seeking behavior of people during the 2009 A/H1N1 pandemic , and determined that information seeking became less common as the pandemic progressed.
Our model assumes that people perform just a single tweet or Internet search before moving to the “recovered” class with permanent immunity. In reality, this simplifying assumption may be violated in some cases, with some people performing several related Internet searches or tweets in a short time period. For instance, the 250,723 tweets related to Ebola used in this analysis were produced by 118,705 unique Twitter users. However, the average time between the first tweet and becoming disinterested with the topic and never tweeting again was 3 days.
In our analysis, we did not examine the sentiment expressed in the Twitter data. Previous studies have shown that negative sentiments tend to be more infectious on social networks than positive sentiments ; while we found no evidence of contagion of Ebola-related sentiments within the Twitter community itself in this particular analysis, it may be that sentiment, in addition to temporal patterns, may be used in a real outbreak situation to disentangle the effect of news media contagion, contagion within the social media platform, and effects due to the spread of the disease.
The vast majority of published digital epidemiology results show a positive correlation between digital data and the temporal evolution of epidemics or outbreaks; but what typically are not seen are the analyses that show no significant correlation, likely due to the “file drawer” effect where uninteresting or null results simply are not published [32, 33]. Indeed, several of the authors (ST, DE, SA) can attest to the fact that use of Twitter data to predict outbreaks is fraught with difficulties (unpublished data), and accounting for potential sources of bias is extraordinarily difficult. However, in this new age of readily accessible, and rapidly evolving, temporal and geo-spatial information in social media, digital epidemiology has a hopeful future as a tool to detect newly emerging infectious diseases, and track the spread of established diseases. The methodology is being continually refined to reduce both Type I and Type II errors (i.e. false positives and false negatives), and efforts are being made to understand the potential biases of the methods. With our analysis, we have explored a source of major potential bias when applying digital epidemiological methodology to emerging disease outbreaks; while media-induced panic is certainly not the only source of bias in such situations, we hope the results of our study will be informative for future analyses.
S1 Data. The file S1_Data.csv is a file containing the data used in this study.
The authors are grateful to Jonathan Dushoff for useful discussions related to this work. The authors also wish to thank the National Science Foundation Extreme Science and Engineering Discovery Environment (NSF XSEDE) high-performance computing initiative for education allocation #DMS140043, which made possible the parameter optimization portion of this analysis. This work was funded in part by the U. S. Department of Homeland Security (DHS) VACCINE Center award #2009-ST-061-CI0001-06, and was also made possible by grant #1R01GM100471-01 from the National Institute of General Medical Sciences (NIGMS) at the National Institutes of Health. This research was also partially supported by the Western Alliance to Expand Student Opportunities (WAESO) Louis Stokes Alliance for Minority Participation (LSAMP) Bridge to the Doctorate (BD) National Science Foundation (NSF) “Multidisciplinary STEM Solutions LSAMP Bridge to the Doctorate” Grant #HRD-1401190, and the Offices of the President and Provost of Arizona State University. The contents of this manuscript are solely the responsibility of the authors and do not necessarily represent the official views of the funding agencies. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Conceived and designed the experiments: ST CCC. Performed the experiments: ST GB NB SB BE JJ JJG MK ML RM VMM FN KO MLR CR. Analyzed the data: ST GB NB SB BE JJ JJG MK ML RM VMM FN KO MLR CR. Contributed reagents/materials/analysis tools: ST DE SA CCC JM. Wrote the paper: ST SA GB NB SB BE JJ JJG MK ML RM VMM FN KO MLR CR.
- 1. Eysenbach G (2006) Infodemiology: tracking flu-related searches on the web for syndromic surveil-lance. In: AMIA Annual Symposium Proceedings. American Medical Informatics Association, volume 2006, p. 244.
- 2. Polgreen PM, Chen Y, Pennock DM, Nelson FD, Weinstein RA (2008) Using Internet searches for influenza surveillance. Clinical infectious diseases 47: 1443–1448. doi: 10.1086/593098. pmid:18954267
- 3. Ginsberg J, Mohebbi MH, Patel RS, Brammer L, Smolinski MS, Brilliant L (2008) Detecting influenza epidemics using search engine query data. Nature 457: 1012–1014. doi: 10.1038/nature07634.
- 4. Signorini A, Segre AM, Polgreen PM (2011) The use of Twitter to track levels of disease activity and public concern in the US during the influenza A H1N1 pandemic. PloS one 6: e19467. doi: 10.1371/journal.pone.0019467. pmid:21573238
- 5. Brownstein JS, Freifeld CC, Madoff LC (2009) Digital disease detection: harnessing the Web for public health surveillance. New England Journal of Medicine 360: 2153–2157. doi: 10.1056/NEJMp0900702. pmid:19423867
- 6. Salathe M, Bengtsson L, Bodnar TJ, Brewer DD, Brownstein JS, Buckee C, et al. (2012) Digital epidemiology. PLoS computational biology 8: e1002616. doi: 10.1371/journal.pcbi.1002616. pmid:22844241
- 7. Schmidt CW (2012) Trending now: using social media to predict and track disease outbreaks. Environ Health Perspect 120: a30–a33. doi: 10.1289/ehp.120-a30. pmid:22214548
- 8. Eysenbach G (2009) Infodemiology and infoveillance: framework for an emerging set of public health informatics methods to analyze search, communication and publication behavior on the Internet. Journal of medical Internet research 11. doi: 10.2196/jmir.1157. pmid:19329408
- 9. Chunara R, Andrews JR, Brownstein JS (2012) Social and news media enable estimation of epidemiological patterns early in the 2010 Haitian cholera outbreak. The American Journal of Tropical Medicine and Hygiene 86: 39–45. doi: 10.4269/ajtmh.2012.11-0597. pmid:22232449
- 10. Salathé M, Khandelwal S (2011) Assessing vaccination sentiments with online social media: implications for infectious disease dynamics and control. PLoS computational biology 7: e1002199. doi: 10.1371/journal.pcbi.1002199. pmid:22022249
- 11. Hay SI, George DB, Moyes CL, Brownstein JS (2013) Big data opportunities for global infectious disease surveillance. PLoS medicine 10: e1001413. doi: 10.1371/journal.pmed.1001413. pmid:23565065
- 12. Balmaseda A, Standish K, Mercado JC, Matute JC, Tellez Y, Saborio S, et al. (2010) Trends in patterns of dengue transmission over 4 years in a pediatric cohort study in Nicaragua. Journal of Infectious Diseases 201: 5–14. doi: 10.1086/648592. pmid:19929380
- 13. Chan EH, Sahai V, Conrad C, Brownstein JS (2011) Using web search query data to monitor dengue epidemics: a new model for neglected tropical disease surveillance. PLoS neglected tropical diseases 5: e1206. doi: 10.1371/journal.pntd.0001206. pmid:21647308
- 14. Wilson K, Brownstein JS (2009) Early detection of disease outbreaks using the Internet. Canadian Medical Association Journal 180: 829–831. doi: 10.1503/cmaj.090215. pmid:19364791
- 15. Zhou X, Ye J, Feng Y (2011) Tuberculosis surveillance by analyzing Google trends. Biomedical Engineering, IEEE Transactions on 58: 2247–2254. doi: 10.1109/TBME.2011.2132132.
- 16. Goel S, Hofman JM, Lahaie S, Pennock DM, Watts DJ (2010) Predicting consumer behavior with Web search. Proceedings of the National Academy of Sciences 107: 17486–17490. doi: 10.1073/pnas.1005962107.
- 17. Choi H, Varian H (2012) Predicting the present with Google Trends. Economic Record 88: 2–9. doi: 10.1111/j.1475-4932.2012.00809.x.
- 18. Cook S, Conrad C, Fowlkes AL, Mohebbi MH (2011) Assessing Google flu trends performance in the united states during the 2009 influenza virus a (H1N1) pandemic. PloS one 6: e23610. doi: 10.1371/journal.pone.0023610. pmid:21886802
- 19. Chew C, Eysenbach G (2010) Pandemics in the age of Twitter: content analysis of tweets during the 2009 H1N1 outbreak. PloS one 5: e14118. doi: 10.1371/journal.pone.0014118. pmid:21124761
- 20. Olson DR, Konty KJ, Paladini M, Viboud C, Simonsen L (2013) Reassessing Google Flu Trends data for detection of seasonal and pandemic influenza: a comparative epidemiological study at three geographic scales. PLoS computational biology 9: e1003256. doi: 10.1371/journal.pcbi.1003256. pmid:24146603
- 21. Metaxas PT, Mustafaraj E, Gayo-Avello D (2011) How (not) to predict elections. In: Privacy, security, risk and trust (PASSAT), 2011 IEEE third international conference on and 2011 IEEE third international conference on social computing (SocialCom). IEEE, pp. 165–171.
- 22. Cha M, Benevenuto F, Haddadi H, Gummadi K (2012) The world of connections and information flow in twitter. Systems, Man and Cybernetics, Part A: Systems and Humans, IEEE Transactions on 42: 991–998. doi: 10.1109/TSMCA.2012.2183359.
- 23. Gallup (2013) TV is Americans’ main source of news See also: www.gallup.com/poll/163412/americans-main-source-news.aspx (accessed February 28, 2015).
- 24. Pew Research Center (2014) Key indicators in media and news See also: www.journalism.org/2014/03/26/state-of-the-news-media-2014-key-indicators-in-media-and-news (accessed February 28, 2015).
- 25. Granger CW (1969) Investigating causal relations by econometric models and cross-spectral methods. Econometrica: Journal of the Econometric Society: 424–438. doi: 10.2307/1912791.
- 26. Bettencourt L, Cintron-Arias A, Kaiser DI, Castillo-Chavez C (2006) The power of a good idea: Quantitative modeling of the spread of ideas from epidemiological models. Physica A: Statistical Mechanics and its Applications 364: 513–536. doi: 10.1016/j.physa.2005.08.083.
- 27. Cowan G (1998) Statistical Data Analysis. Oxford University Press 1998.
- 28. Mosegaard K and Tarantola A (1995) Monte Carlo sampling of solutions to inverse problems. Journal of Geophysical Research: Solid Earth (1978–2012) 100 (B7): 12431–12447.
- 29. McCullagh P, Nelder JA (1989) Generalized linear models. London England Chapman and Hall 1983.
- 30. Tausczik Y, Faasse K, Pennebaker JW, Petrie KJ (2012) Public anxiety and information seeking following the H1N1 outbreak: blogs, newspaper articles, and Wikipedia visits. Health communication 27: 179–185. doi: 10.1080/10410236.2011.571759. pmid:21827326
- 31. Salathé M, Vu DQ, Khandelwal S, Hunter DR (2013) The dynamics of health behavior sentiments on a large online social network. EPJ Data Science 2: 1–12. doi: 10.1140/epjds16
- 32. Rosenthal R (1979) The file drawer problem and tolerance for null results. Psychological bulletin 86: 638. doi: 10.1037/0033-2909.86.3.638.
- 33. Dickersin K (1990) The existence of publication bias and risk factors for its occurrence. Jama 263: 1385–1389. doi: 10.1001/jama.1990.03440100097014. pmid:2406472