11 research outputs found

    Forecasting seasonal influenza fusing digital indicators and a mechanistic disease model

    Get PDF
    The availability of novel digital data streams that can be used as proxy for monitoring infectious disease incidence is ushering in a new era for real-time forecast approaches to disease spreading. Here, we propose the first seasonal influenza forecast framework based on a stochastic, spatially structured mechanistic model (individual level microsimulation) initialized with geo-localized microblogging data. The framework provides for more than 600 census areas in the United States, Italy and Spain, the initial conditions for a stochastic epidemic computational model that generates an ensemble of forecasts for the main indicators of the epidemic season: peak time and intensity. We evaluate the forecasts accuracy and reliability by comparing the results from our framework with the data from the official influenza surveillance systems in the US, Italy and Spain in the seasons 2014/15 and 2015/16. In all countries studied, the proposed framework provides reliable results with leads of up to 6 weeks that became more stable and accurate with progression of the season. The results for the United States have been generated in real-time in the context of the Centers for Disease Control and Prevention “Forecasting the Influenza Season Challenge". A characteristic feature of the mechanistic modeling approach is in the explicit estimate of key epidemiological parameters relevant for public health decision-making that cannot be achieved with statistical models not considering the disease dynamic. Furthermore, the presented framework allows the fusion of multiple data streams in the initialization stage and can be enriched with census, weather and socioeconomic data

    Steering a Historical Disease Forecasting Model Under a Pandemic: Case of Flu and COVID-19

    Full text link
    Forecasting influenza in a timely manner aids health organizations and policymakers in adequate preparation and decision making. However, effective influenza forecasting still remains a challenge despite increasing research interest. It is even more challenging amidst the COVID pandemic, when the influenza-like illness (ILI) counts are affected by various factors such as symptomatic similarities with COVID-19 and shift in healthcare seeking patterns of the general population. Under the current pandemic, historical influenza models carry valuable expertise about the disease dynamics but face difficulties adapting. Therefore, we propose CALI-Net, a neural transfer learning architecture which allows us to 'steer' a historical disease forecasting model to new scenarios where flu and COVID co-exist. Our framework enables this adaptation by automatically learning when it should emphasize learning from COVID-related signals and when it should learn from the historical model. Thus, we exploit representations learned from historical ILI data as well as the limited COVID-related signals. Our experiments demonstrate that our approach is successful in adapting a historical forecasting model to the current pandemic. In addition, we show that success in our primary goal, adaptation, does not sacrifice overall performance as compared with state-of-the-art influenza forecasting approaches.Comment: Appears in AAAI-2

    Collective Response to Media Coverage of the COVID-19 Pandemic on Reddit and Wikipedia: Mixed-Methods Analysis

    Get PDF
    Background: The exposure and consumption of information during epidemic outbreaks may alter people’s risk perception and trigger behavioral changes, which can ultimately affect the evolution of the disease. It is thus of utmost importance to map the dissemination of information by mainstream media outlets and the public response to this information. However, our understanding of this exposure-response dynamic during the COVID-19 pandemic is still limited. Objective: The goal of this study is to characterize the media coverage and collective internet response to the COVID-19 pandemic in four countries: Italy, the United Kingdom, the United States, and Canada. Methods: We collected a heterogeneous data set including 227,768 web-based news articles and 13,448 YouTube videos published by mainstream media outlets, 107,898 user posts and 3,829,309 comments on the social media platform Reddit, and 278,456,892 views of COVID-19–related Wikipedia pages. To analyze the relationship between media coverage, epidemic progression, and users’ collective web-based response, we considered a linear regression model that predicts the public response for each country given the amount of news exposure. We also applied topic modelling to the data set using nonnegative matrix factorization. Results: Our results show that public attention, quantified as user activity on Reddit and active searches on Wikipedia pages, is mainly driven by media coverage; meanwhile, this activity declines rapidly while news exposure and COVID-19 incidence remain high. Furthermore, using an unsupervised, dynamic topic modeling approach, we show that while the levels of attention dedicated to different topics by media outlets and internet users are in good accordance, interesting deviations emerge in their temporal patterns. Conclusions: Overall, our findings offer an additional key to interpret public perception and response to the current global health emergency and raise questions about the effects of attention saturation on people’s collective awareness and risk perception and thus on their tendencies toward behavioral change.Peer ReviewedPostprint (published version

    The contribution of unstructured data from social media for prediction in marketing management

    Get PDF
    The capacity to obtain market insights is a strategic need for companies to remain competitive. Despite this and the massive volume of data generated by consumers every second, companies rarely have the culture of making marketing decisions based on data and, when they do, rarely use consumer data widely available online, specially on social networks. One reason is that these data (e.g. texts) tend to be “dirty”, disorganized and bulky, a so-called unstructured data. Despite the complexity involved in extracting informational value from this data, companies can gain insights that can improve decision making and result in greater competitive performance. The purpose of this article is to discuss the benefits of new types of data that have become more abundant and accessible in Web 3.0, as well as new methods of analysis, particularly learning methods. For this, an extensive literature review was carried out and a topic modeling was conducted to get an overview of the data and methods. At the end, the article suggests six main marketing challenges that unstructured data analytics can contribute, improving companies’ competitiveness. The capacity to obtain market insights is a strategic need for companies to remain competitive. Despite this and the massive volume of data generated by consumers every second, companies rarely have the culture of making marketing decisions based on data and, when they do, rarely use consumer data widely available online, especially on social networks. One reason is that these data (e.g. texts) tend to be “dirty”, disorganized and bulky, a so-called unstructured data. The purpose of this article is to discuss the benefits of new types of data that have become more abundant and accessible in Web 3.0 through popular social networks, as well as new methods of analysis, particularly learning methods for prediction. For this, an extensive literature review was carried out and a topic modeling was conducted to get an overview of the data and methods. At the end, the article suggests six main marketing challenges that unstructured data analytics can contribute to overcome, improving companies’ competitiveness
    corecore