12 research outputs found

    On Identifying Disaster-Related Tweets: Matching-based or Learning-based?

    Full text link
    Social media such as tweets are emerging as platforms contributing to situational awareness during disasters. Information shared on Twitter by both affected population (e.g., requesting assistance, warning) and those outside the impact zone (e.g., providing assistance) would help first responders, decision makers, and the public to understand the situation first-hand. Effective use of such information requires timely selection and analysis of tweets that are relevant to a particular disaster. Even though abundant tweets are promising as a data source, it is challenging to automatically identify relevant messages since tweet are short and unstructured, resulting to unsatisfactory classification performance of conventional learning-based approaches. Thus, we propose a simple yet effective algorithm to identify relevant messages based on matching keywords and hashtags, and provide a comparison between matching-based and learning-based approaches. To evaluate the two approaches, we put them into a framework specifically proposed for analyzing disaster-related tweets. Analysis results on eleven datasets with various disaster types show that our technique provides relevant tweets of higher quality and more interpretable results of sentiment analysis tasks when compared to learning approach

    Rethinking Infrastructure Resilience Assessment with Human Sentiment Reactions on Social Media in Disasters

    Get PDF
    The objective of this study is to propose and test a theoretical framework which integrates the human sentiment reactions on social media in disasters into infrastructure resilience assessment. Infrastructure resilience assessment is important for reducing adverse consequences of infrastructure failures and promoting human well-being in natural disasters. Integrating societal impacts of infrastructure disruptions can enable a better understanding of infrastructure performance in disasters and human capacities under the stress of disruptions. However, the consideration of societal impacts of infrastructure disruptions is limited in existing studies for infrastructure resilience assessment. The reasons are twofold: first, an integrative theoretical framework for connecting the societal impacts to infrastructure resilience is missing; and second, gathering empirical data for capturing societal impacts of disaster disruptions is challenging. This study proposed a theoretical framework to examine the relationship between the societal impacts and infrastructure performance in disasters using social media data. Sentiments of human messages for relevant infrastructure systems are adopted as an indicator of societal impacts of infrastructure disruptions. A case study for electricity and transportation systems in Houston during the 2017 Hurricane Harvey was conducted to illustrate the application of the proposed framework. We find a relation between human sentiment and infrastructure status and validate it by extracting situational information from relevant tweets and official public data. The findings enable a better understanding of societal expectations and collective sentiments regarding the infrastructure disruptions. Practically, the findings also improve the ability of infrastructure management agencies in infrastructure prioritization and planning decisions

    Community Segmentation and Inclusive Social Media Listening

    Get PDF
    Social media analytics provide a generalized picture of situational awareness from the conversations happening among communities present in social media channels that are that are, or risk being affected by crises. The generalized nature of results from these analytics leaves underrepresented communities in the background. When considering social media analytics, concerns, sentiment, and needs are perceived as homogenous. However, offline, the community is diverse, often segmented by age group, occupation, or language, to name a few. Through our analysis of interviews from professionals using social media as a source of information in public service organizations, we argue that practitioners might not be perceiving this segmentation from the social media conversation. In addition, practitioners who are aware of this limitation, agree that there is room for improvement and resort to alternative mechanisms to understand, reach, and provide services to these communities in need. Thus, we analyze current perceptions and activities around segmentation and provide suggestions that could inform the design of social media analytics tools that support inclusive public services for all, including persons with disabilities and from other disadvantaged groups.publishedVersionPaid open acces

    Bot-Based Emergency Software Applications for Natural Disaster Situations

    Get PDF
    Upon a serious emergency situation such as a natural disaster, people quickly try to call their friends and family with the software they use every day. On the other hand, people also tend to participate as a volunteer for rescue purposes. It is unlikely and impractical for these people to download and learn to use an application specially designed for aid processes. In this work, we investigate the feasibility of including bots, which provide a mechanism to get inside the software that people use daily, to develop emergency software applications designed to be used by victims and volunteers during stressful situations. In such situations, it is necessary to achieve efficiency, scalability, fault tolerance, elasticity, and mobility between data centers. We evaluate three bot-based applications. The first one, named Jayma, sends information about affected people during the natural disaster to a network of contacts. The second bot-based application, Ayni, manages and assigns tasks to volunteers. The third bot-based application named Rimay registers volunteers and manages campaigns and emergency tasks. The applications are built using common practice for distributed software architecture design. Most of the components forming the architecture are from existing public domain software, and some components are even consumed as an external service as in the case of Telegram. Moreover, the applications are executed on commodity hardware usually available from universities. We evaluate the applications to detect critical tasks, bottlenecks, and the most critical resource. Results show that Ayni and Rimay tend to saturate the CPU faster than other resources. Meanwhile, the RAM memory tends to reach the highest utilization level in the Jayma application.Fil: Ovando Leon, Gabriel. Universidad de Santiago de Chile; ChileFil: Veas Castillo, Luis. Universidad de Santiago de Chile; ChileFil: Gil Costa, Graciela Verónica. Universidad Nacional de San Luis; Argentina. Consejo Nacional de Investigaciones Científicas y Técnicas. Centro Científico Tecnológico Conicet - San Luis; ArgentinaFil: Marin, Mauricio. Universidad de Santiago de Chile; Chil

    COVIDHealth: A Benchmark Twitter Dataset and Machine Learning based Web Application for Classifying COVID-19 Discussions

    Full text link
    The COVID-19 pandemic has had adverse effects on both physical and mental health. During this pandemic, numerous studies have focused on gaining insights into health-related perspectives from social media. In this study, our primary objective is to develop a machine learning-based web application for automatically classifying COVID-19-related discussions on social media. To achieve this, we label COVID-19-related Twitter data, provide benchmark classification results, and develop a web application. We collected data using the Twitter API and labeled a total of 6,667 tweets into five different classes: health risks, prevention, symptoms, transmission, and treatment. We extracted features using various feature extraction methods and applied them to seven different traditional machine learning algorithms, including Decision Tree, Random Forest, Stochastic Gradient Descent, Adaboost, K-Nearest Neighbour, Logistic Regression, and Linear SVC. Additionally, we used four deep learning algorithms: LSTM, CNN, RNN, and BERT, for classification. Overall, we achieved a maximum F1 score of 90.43% with the CNN algorithm in deep learning. The Linear SVC algorithm exhibited the highest F1 score at 86.13%, surpassing other traditional machine learning approaches. Our study not only contributes to the field of health-related data analysis but also provides a valuable resource in the form of a web-based tool for efficient data classification, which can aid in addressing public health challenges and increasing awareness during pandemics. We made the dataset and application publicly available, which can be downloaded from this link https://github.com/Bishal16/COVID19-Health-Related-Data-Classification-Website.Comment: 27 pages, 6 figure

    Visual Event Cueing in Linked Spatiotemporal Data

    Get PDF
    abstract: The media disperses a large amount of information daily pertaining to political events social movements, and societal conflicts. Media pertaining to these topics, no matter the format of publication used, are framed a particular way. Framing is used not for just guiding audiences to desired beliefs, but also to fuel societal change or legitimize/delegitimize social movements. For this reason, tools that can help to clarify when changes in social discourse occur and identify their causes are of great use. This thesis presents a visual analytics framework that allows for the exploration and visualization of changes that occur in social climate with respect to space and time. Focusing on the links between data from the Armed Conflict Location and Event Data Project (ACLED) and a streaming RSS news data set, users can be cued into interesting events enabling them to form and explore hypothesis. This visual analytics framework also focuses on improving intervention detection, allowing users to hypothesize about correlations between events and happiness levels, and supports collaborative analysis.Dissertation/ThesisMasters Thesis Computer Science 201

    General supervised learning framework for open world classification

    Get PDF
    In machine learning, the most common scenario for classification modeling is when the training set contains all possible classes and the algorithm learns to identify these classes. The problem setting in which the training data does not contain all classes is referred to as open-world classification problem. Hence, when the model is applied to new data, it is imperative to identify the instances belonging to unknown classes. While literature addresses this issue, most of the work in this field has been limited to the domain of computer vision and the solution approaches are specific to a particular type of machine learning algorithm. Furthermore, it is equally important to categorize the identified instances into their classes, to facilitate retraining and to the best of our knowledge, there is no generalized approach that provides a complete solution to the problem. This work proposes a framework that can identify instances from unseen data and also categorize the identified instances into their respective classes. We claim that this methodology works irrespective of the nature of the data and also the type of classifier under consideration. To validate our claim, the methodology is tested on different types of data such as image, text, sensor, etc. Furthermore, the proposed framework is demonstrated on the case study; social media analytics for community resilience. Our results show that the performance of the methodology is successful and consistent across the data sets and the case study considered
    corecore