March 28, 2024
With the advent of the era of big data, massive information, expert experience, and high-accuracy models bring great opportunities to the information cascade prediction of public emergencies. However, the involvement of specialist knowledge from various disciplines has resulted in a primarily application-specific focus (e.g., earthquakes, floods, infectious diseases) for information cascade prediction of public emergencies. The lack of a unified prediction framework poses a challenge for classifying intersectional prediction methods across different application fields. This survey paper offers a systematic classification and summary of information cascade modeling, prediction, and application. We aim to help researchers identify cutting-edge research and comprehend models and methods of information cascade prediction under public emergencies. By summarizing open issues and outlining future directions in this field, this paper has the potential to be a valuable resource for researchers conducting further studies on predicting information cascades.
<ccs2012> <concept> <concept_id>10010405.10010481.10010487</concept_id> <concept_desc>Applied computing Forecasting</concept_desc> <concept_significance>500</concept_significance> </concept> <concept> <concept_id>10010405.10010481.10010484</concept_id> <concept_desc>Applied computing Decision analysis</concept_desc> <concept_significance>500</concept_significance> </concept> <concept> <concept_id>10002951.10002952.10003219</concept_id> <concept_desc>Information systems Information integration</concept_desc> <concept_significance>300</concept_significance> </concept> </ccs2012>
Information Cascade Prediction is a critical research field that has implications in a variety of domains, such as healthcare, business, cyber domain, politics, and entertainment, ultimately impacting nearly every aspect of our lives [1]. These emergencies are unexpected events that occur suddenly and result in or have the potential to result in significant casualties, property damage, ecological harm, and serious social consequences [2]. Throughout history, natural disasters (such as earthquakes, tsunamis, volcanic eruptions, storms, floods, avalanches, droughts, and wildfires) and accident disasters (including environmental disasters, traffic accidents, explosions, and gas leaks) have caused numerous fatalities, infrastructure damage, and extensive economic loss. According to the Emergencies Database (EM-DAT), between 2000 and 2023, 5,922 public emergencies occurred, leading to 480,000 casualties and 3.5 trillion in economic losses, as shown in Figure 1 [3]. Therefore, it is increasingly vital to use data, information, and various models to predict potential public emergencies that jeopardize public safety and well-being. Predicting the cascade of information in the event deduction process under public emergencies assists governments, organizations, and individuals in taking proactive measures to mitigate the impact of emergencies and minimize damage.
Public emergencies are classified into different categories. The most common categories of public emergencies include (1) Natural disasters, (2) Accident disasters.
Figure 1: Global Distribution of Public Emergencies.
Natural disasters pose a significant and recurring threat to human populations and infrastructure worldwide. These disasters range from sudden, catastrophic events such as volcanic eruptions, earthquakes, floods, hurricanes, and storms to slower, gradual processes like land desertification, soil erosion, environmental degradation, and major infectious disease outbreaks, mass unexplained diseases. For instance, the COVID-19 pandemic, which began in December 2019, has had a devastating impact globally, with over 6.5 million deaths reported to the World Health Organization as of 21 October 2022, and more than 623 million confirmed cases. In addition to the immediate impact of natural disasters, derivative disasters often follow, which exacerbate the initial damage caused by the event. Therefore, it is crucial to recognize the complex and interconnected nature of natural disasters, as one disaster trigger or contribute to several others.
Accident Disaster is a term that refers to unexpected events occurring during people’s production or daily life, typically caused by human activities, and resulting in a substantial number of casualties, economic losses, or environmental pollution. Examples of Accident Disasters include fires, explosions, and toxic leaks during industrial production; collapses, gas and coal dust explosions, and flooding during mining; and shipwrecks, major traffic accidents, and aircraft accidents during river-sea transportation. Accident Disasters are a type of public emergency with universality, randomness, inevitability, causal correlation, mutation, latency, and harmfulness. Disasters significantly impact people’s lives and production, resulting in severe social consequences.
Public emergencies often create uncertainty and panic, leading individuals to be influenced by the behavior of those around them. The phenomenon where people rely on the actions of others rather than their own judgment or the development of events to form impressions, often due to external factors, is referred to as information cascades. The cascading information predicts events, anticipate their progress and human behavior, and predicts their impact on other events or sub-events. Information Cascade Prediction involves analyzing and predicting how information spreads and affects decision-making during crises or emergencies. Researchers and practitioners utilize diverse algorithms to predict and manage information cascades during public emergencies.
Decision Trees are a type of supervised learning algorithm that works by recursively partitioning the data based on various features or parameters until a stopping criterion is met. Bayesian Networks are probabilistic graphical models that represent and reason about uncertain relationships between various factors. They are used to predict the likelihood of cascading events or hazards by incorporating various factors and their probabilistic relationships.
Matrix Factorization algorithms, such as Singular Value Decomposition and Non-negative Matrix Factorization, are effective tools for analyzing and predicting the spread of a pandemic or other public health crises. Support Vector Machines are commonly used to predict the path and intensity of natural disasters like hurricanes or typhoons, taking factors like wind speed and atmospheric pressure into account. Similarly, Random Forests predict the likelihood of cascading failures during a cyber attack on critical infrastructure by analyzing various factors related to the attack and the resilience of the infrastructure.
Neural Networks, including Long Short-Term Memory Networks (LSTMs) and Graph Neural Networks (GNNs), are capable of learning complex patterns and relationships in data, which leads to improved prediction accuracy. LSTMs predict future actions based on past behavior patterns, while GNNs incorporate social network data to predict the spread of behavior patterns. Other algorithms like Deep Learning models, Generative Adversarial Networks (GANs), Autoencoders, Attention mechanisms, and Transformer models are also valuable for learning complex patterns and relationships in data and improving prediction accuracy. Each of these algorithms is used to predict various events and hazards using different types of data and relationships between them.
In summary, the choice of algorithm for information cascade prediction in public emergencies depends on the specific task and the level of complexity required. Each algorithm has its strengths and limitations, and the appropriate choice depends on the nature of the problem and the available data.
It is well-known that data-driven approaches are commonly applied at 4 stages of public emergencies, including mitigation, preparedness, response, and recovery. However, predicting public emergencies is challenging due to the uncertain characteristics of events. In recent years, the rapid development of machine learning methods has led to significant progress in the development and application of information cascade prediction technology. The various data-driven methods are applied at each stage of a public emergency as shown in Table 1.
0.2cm
Surveys | Regular Events | Public Emergency | |||||||
---|---|---|---|---|---|---|---|---|---|
Model | Prediction | Multi—stage Application | |||||||
Direct Risk | Sequence Risk | Cascading Risk | Before | During | After | ||||
[4] | ✔ | ||||||||
[5] | ✔ | ||||||||
[6] | ✔ | ||||||||
[7] | ✔ | ||||||||
[8] | ✔ | ✔ | |||||||
[9] | ✔ | ✔ | ✔ | ||||||
[10] | ✔ | ✔ | |||||||
[11] | ✔ | ||||||||
Our Work | ✔ | ✔ | ✔ | ✔ | ✔ | ✔ | ✔ |
In recent years, several studies[4]–[11] have reviewed prediction algorithms for public emergencies. However, none of these studies have presented a comprehensive overview of methods that effectively incorporate information cascade prediction modeling, prediction, and application. Gao and Duan et al. [4], [5] have emphasized event modeling and simple risk forecasting using 1D structured data. Nuno, Alexandru, and Gabor et al. [6]–[9] have focused on summarizing information cascade prediction methods under public emergencies and analyzing the application of various feature engineering methods. Ali and Robert et al. [10], [11] have considered information cascade prediction for emergency public events and summarized application methods for different stages of emergency events. Although these reviews have approached public emergency prediction methods from different perspectives, they fail to consider the influence of each stage of a public emergency and the event evolution caused by the information cascade, as well as time, place, semantics, deduction, and other features.
To address this gap, our work focuses on developing systematic and standardized summary methods to evaluate various information cascade prediction methods. These efforts help identify bottlenecks, pitfalls, open problems, and potentially fruitful future research directions.
Figure 2: The Overall Structure Of the Survey
The motivation behind conducting a comprehensive review of the field of information cascade prediction is to gain a deeper understanding of the methods and techniques used to predict information cascades and analyze the diffusion of individuals. By examining current research in these areas, we identify gaps in knowledge, and potential areas for improvement, and determine the most effective approaches and people flow at different levels of interaction. Different from previous surveys, our work focuses on information cascade prediction methods under emergencies. This review aims to provide a comprehensive overview of the field, highlighting key findings and advancements in information cascade prediction and people flow analysis. Compared to previous investigations, this paper makes several unique contributions.
Firstly, it systematically classifies and summarizes existing technologies, categorizing them according to event aspects, problem formulation, and corresponding techniques. This taxonomy helps domain experts find the most useful techniques for their target problem settings.
Secondly, the paper conducts a finer-grained analysis of existing information cascade prediction methods for emergencies. It also considers the information cascades and prevalence prediction of emergencies, explaining and analyzing their characteristics and methods.
Thirdly, the paper provides a comprehensive and up-to-date literature review of information cascade prediction methods, covering both the macro-model of collective behavior and the micro-model of individual user responses. These approaches incorporate recent advances in modeling and predicting information popularity.
Lastly, the paper discusses the research status and future trends in the field, outlining the overall situation and shape of the current research front. It concludes with new insights into bottlenecks, pitfalls, and open issues, and discusses possible future directions.
The remainder of this paper is organized as shown in Figure 2. Section 2 presents Problem Formulation and Performance Evaluation. Section 3 introduces the Information Cascade Model under Public Emergencies. Section 4 introduces the Deduction of Public Emergencies by joint information cascade prediction, including joint time and semantics (Section 4.1), joint time and location (Section 4.2), joint time, location, and semantics (Section 4.3), Vulnerability Prediction (Section 4.4), Association-based Impact Prediction (Section 4.5), and Causality-based Prediction (Section 4.6). Section 5 introduces the Application to Public Emergencies and points out directions for future work, and the survey concludes in Section 5.
Due to the particularity of public emergencies, there are usually information cascade items between public emergencies. This information is usually heterogeneous, sparse, and biased data due to different acquisition channels and methods. This section introduces the formulas and classifications used in the information cascade prediction process of public emergencies: (1) Information Cascade; (2) Prediction.
Information-level correlation in public emergencies often refers to the information correlation between sub-events or between events and events in public emergencies. The information cascade prediction of public emergencies is essentially a regression problem. The development trend of public emergencies is represented by information diffusion.
Definition 2.1 (Information Cascade) Assume that there are \(M\) information items \({I_1, I_2,..., I_M}\) for event \(y\). \(I_i\) represents the cascading degree of events at time \(t\), that is, the probability that event \(y\) may affect subsequent events. and information cascade prediction aims to predict the occurrence probability \(P_i(t)\) at time \(t\) in the future.
Information Cascade Prediction aims to predict whether a cascading event occurs given a predefined absolute/relative threshold. For example, whether the event occurs [12], [13], and whether the public emergency causes other cascading events in the future [14], [15], in addition, the classification task is also used to predict the probability of the occurrence of the emergency. It is possible to fall into which interval [16]–[19], which is the classic multi-class classification task. Since the classification task is similar to the low-level version of the regression task, it has lower dimensions in the data parameters, making it easier to achieve good results. Developing regression strategies relies on expert knowledge and requires a fine-grained scope to analyze which factors affect predicting cascading events in public emergencies. Due to the complexity of cascades among public emergencies, accurate regression predictions usually require more information about events as well as cascades [20]. Furthermore, it suffers from undesirable problems such as overfitting, inductive bias, and accumulation of prediction errors [21].
Public emergencies contain three important information items: time, location, and semantic [22]. Let sudden public emergencies be represented by \(y = (t,l,s)\), \(t \in T\) represents the time factor, \(l \in L\) represents the positioning element, and \(s \in S\) represents the semantic element. \(T\), \(L\), and S represent the time, location, and semantic domains. Among them, the semantic domain contains any type of semantic feature useful in detailing the semantics of various aspects of an event, including its actors, objects, actions, size, textual description, and other analytical information.
In the context of predicting information cascades during public emergencies, it is possible to independently predict the time, location, and semantics of disasters while predicting other factors of the emergency through different elements of the three. Alternatively, the cascading nature of public emergencies is used to predict the occurrence of public emergencies, with precursor events and subsequent events being jointly predicted. The emergency information is represented by \(X \subseteq T*L*F\), and F represents the set of semantic information. The current time is expressed as \(t\), past time and future time are expressed as \(T^-\equiv \{ t|t_{now} \le t,t \in T\}\) and \(T^+ \equiv \{ t |t > t_{now},t \in T \}\), the public emergencies formula is as follows:
Definition 2.2 (Public Emergency Prediction) Let the set of precursory events be expressed as \(X\subseteq T^-*L*S\), and the data of subsequent precursory events be expressed as \(Y_0\subseteq T^-*L*S\). Based on the information of precursor events and subsequent events, the information of public emergencies is expressed as \(\hat{Y} \subseteq T^-*L*S\). So that each future emergency is expressed as \(\hat{y} = (t,l,s)\in \hat{Y}\) where \(t>t_{now }\).
The Information Cascade Model is a predictive model that utilizes feature extraction techniques to predict the spread and severity of public emergencies. This model is designed to identify relevant factors that influence the occurrence and spread of emergencies and to incorporate these factors into a machine learning framework for predictive modeling. The Information Cascade Model incorporates various types of features, including Temporal Features, Structural Features, User/Ttem Features, and Content32Features, to capture different aspects of public emergency occurrences. Temporal features capture changes in emergency occurrences over time, while structural features capture relationships between emergency response networks. User/item features capture relevant demographic and socioeconomic factors, while content features capture emergency data.
Temporal features play a crucial role in predicting the spread of information items or cascades during public emergencies. In this section, we discuss the importance of different types of temporal features in feature engineering for prediction. Specifically, we focus on four subtopics: Observation Time, Occurrence Time, Participation Time, and Evolving Trends as shown in Table 2.
0.2cm
Feature | Definition | References |
---|---|---|
Observation Time | The observation time of early features and precursor events | [8], [14], [23] |
Occurrence Time | The time at which a public emergency begins | [8], [16], [24]–[27] |
Participation Time | The time at which an individual participates in a public emergency | [13], [28], [29] |
Evolving Trends | The Developments in public opinion, behavior, or Events over Time | [23], [30], [31] |
Cascade Graph | A directed graph representing the spread of information from an initial node | [32]–[35] |
Global Graph | A fundamental representation of the relationships between nodes | [36]–[48] |
r-reachable Graph | A sub-graph extracted from the global graph based on cascade nodes | [38], [49] |
Temporal features play a critical role in predicting the popularity of public emergencies. These features are typically extracted based on the peeking strategy, which involves observing a small number of early features and their action time to obtain a sequence of timestamps that are utilized for feature selection. However, the length of the time series is highly irregular, and directly utilizing timestamps as a feature is often ineffective in practice. To address this, transformations are often applied in advance [23]. For example, the period may be divided into evenly distributed intervals, and the cumulative or incremental popularity may be calculated, or a fixed number of early features may be observed [14]. To predict the popularity \(P_i(t_p)\) at prediction time \(t_p\) based on the information observed at time \(t_o\), previous studies have analyzed the relationships between the log-transformed popularity \(P_i(t_p)\) and \(P_i(t_o)\). One such study [8] found a high correlation between early-stage and future popularity and used a simple linear prediction model that takes the early observed popularity as input to predict future popularity.
Temporal features are crucial in predicting the popularity of public emergencies, with occurrence time (\(t_0\)) being a particularly important factor. Previous studies [8], [24]–[26], [29] have highlighted the strong relationship between the popularity of public emergencies and the time they occurred, with items posted during the day generally more popular than those posted at midnight. To overcome the effect of user activity periodicity, researchers have proposed various solutions. For instance, local models were designed with each model trained on samples published at a specific hour during the day [13]. Tweet time was utilized to eliminate the imbalanced diurnal effect of user activities [16]. Other temporal factors, such as source time [29] and activeness variability [27], have also been employed to improve the robustness of prediction models. However, to improve the accuracy of prediction models, other temporal features, such as the age of the tweet and the time since the event started, should also be considered. For example, a study [50] showed that the age of a tweet has a negative correlation with its popularity, indicating that tweets posted early during an event are more likely to be popular than those posted later. Therefore, future research should explore the use of various temporal features to improve the accuracy of prediction models for public emergencies.
The first arrival time \(t_1\) is another crucial temporal feature to consider in forecasting public emergencies. Moreover, various more sophisticated temporal features have been proposed in the literature, such as mean arrival time \(1/M {\textstyle \sum_{M}^{j=1}} t_j\), mean reaction time \(1/M {\textstyle \sum_{M}^{j=1}} (t_j-t_{j-1})\), change rate, dormant period, and peek fraction. For instance, research has shown that human reaction time typically follows a log-normal distribution, such as people’s reactions to calls, emails, and social networks [28]. These temporal features provide useful insights for predicting the spread and impact of public emergencies.
Analyzing the changing trends of public emergencies has been shown to provide valuable insights for forecasting their severity and impact. Previous studies [23], [30], [31] have demonstrated that identifying the temporal patterns of these events helps predict their popularity and impact on society. These patterns are categorized into various types, such as steadily increasing or rapidly fluctuating, depending on the clustering algorithm.
The spread of public emergencies, such as pandemics, has garnered significant attention from scholars across various fields. In particular, the structure of cascades, also known as information diffusion, has been the subject of extensive research. It explains how information and contagions spread through populations [14], [20], [51]–[54]. Scholars have approached the modeling of cascades in different ways, which are categorized into three types. The first type is Cascade Graph [32]–[35], which involves the analysis of cascade graphs to understand how information spreads among participants in a network. The second type is Global Graph [36]–[48], which considers both participants and non-participants in the network, considering the external factors that influence the spread of the emergency. Finally, the third type is r-reachable modeling [38], [49], which seeks to strike a balance between the first two types, by expanding the cascade graph to include non-participants within the global graph as shown in Table 2.
A cascade graph is constructed based on its participants and their interactions:
Definition3.1 (Cascade Graph) Given an information item \(I_i\) and the corresponding cascade \(C_i\), a cascade graph is defined as \(G_c\) \(=\{V_c, E_c\}\), where nodes \(Vc\) \(=\{u_0,u_1,...,u_N\}\)are participants of cascade \(C_i\), and matrix \(E_c\) \(\subset V_c * V_c\) contains a set of edges representing immediate relationships between \(V_c\) in a cascade.
During public emergencies, the spread of information is crucial to inform and mobilize individuals to take necessary actions. The study of information diffusion through cascades has been widely researched, with a cascade graph being a fundamental concept in the field [34], [35]. A cascade graph, denoted as \(G_c ={V_c, E_c}\), consists of a set of nodes, \(V_c\), representing participants of the cascade, and a set of edges, \(E_c\), representing the immediate relationships between the participants as shown in Table 3. Cascade graphs play a vital role in characterizing the process of information diffusion of a particular item, such as the spreading directions and graph topology. For instance, one study examined the correlation between cascade popularity and two structural features, namely edge density, and depth, among early participants in microblogging networks [33]. However, it is essential to note that the topological structure of cascade graphs varies significantly, even when the number of nodes is in the same [35]. The depth, structural virality, and other structural measurements, such as node degree or PageRank, may not accurately predict whether an information item would be popular. As cascades grow over time, the initial structural features may become less important in determining their popularity [32]. Therefore, further research is needed to explore the complex mechanisms of information diffusion during public emergencies and develop effective information dissemination strategies.
0.2cm
Structural features | Definition | Examples | References |
---|---|---|---|
Follower/Followee Graph | A graph where nodes represent users and edges represent the relationships | Twitter, Instagram, Facebook | [42], [48] |
Hidden Friend Graph | A graph representing less obvious or non-publicly visible relationships | Twitter, Facebook | [36], [37], [39], [45], [46] |
Co-Participation Graph | A graph that captures the relationships between users based on their participation | Digg, Reddit | [43], [55], [56] |
Interaction Global Graph | A graph that captures interactions between users | Twitter, YouTube, Instagram | [37], [38], [41], [44], [47] |
Understanding the spread of information across different types of networks is crucial for effective crisis communication as shown in Table 3. While cascade graphs provide insight into the local spread of information, exploring global graphs is also important [36], [37], [39], [45], [46]. These interaction global graphs are useful in various prediction tasks, particularly when the explicit social graph is not available, and historical behaviors serve as a suitable representation of the actual diffusion of information [48]. Therefore, understanding the characteristics of different types of global graphs can aid in designing effective crisis communication strategies during public emergencies.
Definition 3.2 (Global Graph) A global graph \(G_g=(V_g, E_g)\) is a fundamental representation of the relationships between nodes, where \(V_g\) is a set of nodes, and \(E_g \subset V_g * V_g\) is a set of edges that connect nodes based on their relationships. The global graph is further defined based on additional characteristics such as edge direction, edge weight, node/edge attributes, and node/edge features.
By providing a macro perspective of the relationships between nodes, a global graph offers a valuable tool to analyze the spread of information to individuals and communities. In contrast to the cascade graph, which shows the local spread patterns for information cascade, the global graph describes the relationships between users and potential routes for diffusion [57]. In social networking platforms, the discovery and dissemination of information items primarily occur through the users’ social networks. This highlights the importance of understanding the relationships between users in the global graph to effectively analyze and mitigate public emergencies such as the spread of misinformation, epidemics, and natural disasters. [58]
Drawing upon the concepts of cascade and global graphs, a sub-graph extracted from the global graph, termed an r-reachable graph, is defined as follows:
Definition 3.3 (r-reachable Graph) Consider a global graph \(G_g\) and its cascade sub-graph \(G_c\). An r-reachable graph of \(G_c\), denoted by \(G_c^r ={V_c^r, E_c^r}\), is a graph comprising the nodes in \(V_c\) and those in \(V_g\) that are within \(r\)-hops of nodes in \(V_c\). For instance, if \(r\) = 1, then \(V_c^r\) contains all nodes in \(V_c\) and their immediate neighbors.
The r-reachable graph has been employed in several studies to aid in predicting the popularity of cascades in social media [49]. A global graph \(G_g\) and several 1-reachable graphs \(G_c^1\) were constructed from retweet cascades in Weibo, and structural features were extracted from \(G_g\) and \(G_c^1\) to forecast tweet popularity [38]. A comparative analysis of several content and structural features revealed that the size of the cascade graph and the 1-reachable graph are the two most predictive of 53 features for predicting the popularity of Twitter hashtags [49]. Nevertheless, constructing and computing an r-reachable graph are computationally intensive; for instance, the 2-reachable graph of a cascade graph with dozens of nodes may contain tens of thousands of nodes [59].
Understanding the early structure and patterns of information diffusion is crucial for effective crisis management and response [36]. As the spread of misinformation and rumors exacerbate the impact of a crisis, identifying influential users and predicting the popularity of information items are essential tasks [60]–[62]. Previous studies have shown that the structure of cascade graphs is not a reliable predictor of final item popularity, and alternative approaches such as identifying influential users must be considered. Furthermore, the types of items and their contents also affect information diffusion behavior during crises [63]. Therefore, analyzing the structural patterns of information diffusion and identifying influential users in the context of public emergencies provide valuable insights for crisis management and response [32], [44].
The unprecedented occurrence of public emergencies has heightened the need for accurate and timely predictions of their impact on society. However, obtaining early observations and monitoring temporal and structural features is impractical in the context of rapidly unfolding events. As such, some works try to examine the features inherent to users and information items, which possess distinct characteristics and inherent appeal that render them valuable in forecasting popularity before dissemination.
During public emergencies, such as natural disasters or pandemics, information dissemination becomes crucial in managing and mitigating the situation. User behaviors play a critical role in this process, as they determine the speed and reach of information spread. However, it is important to note that large cascades of information are not produced by influential users such as celebrities and news organizations but also originate from normal users. Therefore, it is essential to study and analyze large cascades produced by many types of users during public emergencies. Various other features have been extensively explored and studied for analyzing and predicting the popularity of information items during public emergencies. These features include user profiles, historical behaviors, user interests, collectivity, similarity, activity/passivity, discoveries, affinities, and responsiveness [64]–[70]. Understanding and utilizing these features predict and manage the spread of information during public emergencies, ultimately leading to more effective communication and response.
Understanding the impact of project characteristics on information dissemination has a significant impact on the dissemination of key messages as shown in Figure 3. Previous literature explored various item characteristics and their impact on information dissemination. For example, one study analyzed how user interfaces on social media platforms affect item visibility [71]. Furthermore, using entropy computed across information categories and topics provides insight into the diversity of information and its impact on dissemination [43]. Using popularity variability to analyze how the popularity of different types of items changes over time helps identify patterns and trends in information dissemination during public emergencies [72].
Figure 3: From Emergency Time-line to Transactions
During public emergencies, understanding the factors that contribute to information diffusion is critical. User and item characteristics play a significant role, with some features being more easily understood than others. However, factors like user influence, preferences, and similarities require more sophisticated algorithms and calculations. Advanced techniques enable better predictions and are particularly useful in such situations where timely and accurate information dissemination is crucial [73]–[75].
From the perspective of predicting public emergencies, content is widely recognized as an essential driving force and one of the key factors that contribute to the success of any emergency response effort. For instance, breaking news, rumors, and fake news, as well as hot spots and controversial or peculiar topics, disinformation, and misinformation, tend to attract considerably more attention than regular content.
From the perspective of predicting public emergencies, text feature is a crucial component of existing prediction models. Textual information is pervasive in articles, microblogs, image/audio/video captions/descriptions, and even retrieved from multimedia sources. Researchers have employed various language models such as Term Frequency-Inverse Document Frequency (TF-IDF) and Latent Dirichlet Allocation (LDA) [76], along with typical Machine Learning models like naive Bayes, SVM, and linear regression to predict item popularity. TF-IDF and LDA are used to learn the topic distributions of tweets [64]. TF-IDF is utilized to estimate the importance of keywords in user tweets and to calculate the mutual correlation between a user’s historical content and a specific item to measure their likelihood of adopting that item as shown in Figure 4 (a) [69]. Authors analyze various semantic and statistical content features of Digg comments to identify the characteristics of content that people prefer to retweet [77].
Figure 4: Content32Features.
The retrieval and analysis of image features require techniques from computer vision learning, distinct from text-based methods. Basic features of an image, such as size, date, orientation, dominant color, resolution, location, caption, and tag. For example, during the COVID-19 pandemic, computer vision techniques were utilized to analyze images from social media platforms and identify violations of social distancing protocols, such as large gatherings or people not wearing masks [78]. Lu et al. [79] provide an analysis of these attributes, while Michael et al. [80] study the correlations between Flickr image features and their normalized popularity as shown in Figure 4 (b). Content features of images, categorized as simple human-interpretable, low-, and high-level image features, contribute significantly to improving prediction performance. During natural disasters like hurricanes or earthquakes, computer vision techniques are utilized to analyze satellite images and identify areas that have been affected and need assistance. Li et al. [78] extract texture and color features and use VGG19 [81] to extract deep features. During a pandemic, computer vision techniques are utilized to analyze chest X-rays and identify patterns that could indicate COVID-19 infection.
The challenge of predicting item features based solely on content features is still prevalent, despite previous studies exploring the linguistic and visual characteristics of items. This difficulty arises because even items with the same content have varying features, making it challenging to distinguish the effects of descriptive factors from intrinsic content. This unpredictability is particularly relevant in predicting the spread of misinformation during a public health crisis.
Predicting the joint information of public emergencies is a crucial task in disaster management. Research in this area are categorized into six types: (1) Joint Time and Semantic Prediction; (2) Joint Time and Location Prediction; (3) Joint Time, Location, and Semantic Prediction; (4) Vulnerability Prediction; (5) Association-based Impact Prediction; (6) Causality-based Prediction.
In the prediction of public emergencies, various methods have been developed to jointly predict the joint time and semantics. Existing work in this area is broadly categorized into the following three types: Temporal Association Rule; Spatial-Temporal Methods; Time Series Forecasting-based Methods.
Temporal association rules are used to embed additional temporal information into the occurrence of events and to redefine the meaning of co-occurrence and association with temporal constraints as shown in Table 4. In that case, they define the Left-Hand Side (LHS) of the association rule as a tuple \((E_L,\tau)\), where \(\tau\) represents the time window before the target flood emergency occurrence in the Right-Hand Side (RHS) predefined by the user. The events that occur within this time window before the flood emergency satisfy the LHS. However, defining the time window beforehand is challenging and may not suit different target events. To overcome this, researchers have proposed a way to automatically identify information on a continuous time interval from the data [68]. Here, the transaction comprises items and continuous time duration information as shown in figure 5. LHS is a set of items (e.g., previous emergencies), and RHS is a tuple \((E_R,[t_1,t_2])\) consisting of a future emergency’s semantic representation and its time interval of occurrence. To learn the time interval in RHS, two different methods have been proposed: confidence-interval-based and minimal temporal region selection [68].
0.2cm
Category | Direction | Method | References |
---|---|---|---|
Risk Occurrence | Binary Classification | Simple Threshold-based | [82]–[85] |
3-4 | Logistic regression | [86]–[88] | |
3-4 | Support Vector Machines | [43], [55], [89] | |
3-4 | Neural Networks | [62], [90], [91] | |
3-4 | Decision trees | [10], [40], [92] | |
2-4 | Anomaly detection | One-classification, hypotheses testing | [42], [93] |
2-4 | Regression | Auto-regression, linear regression, ordinal regression | [37], [94], [95] |
Discrete-time | Time windows | Auto-regression, ARIMA, ordinal regression or classification | [96]–[98] |
2-4 | Time scales | Regression or classification | [67], [99], [100] |
2-4 | Time series | Autoregressive, burstiness detection, change detection | [101]–[103] |
3-4 | Learning event characterization | [73], [104]–[106] | |
Rule-based | Associative learning | Frequent set mining | [103], [107]–[109] |
Cross-chain | Generalizing event graphs | Graph representation | [110], [111] |
Semantic causation | Event representation | RDF-based | [36], [112]–[115] |
2-4 | Event inference | probability estimate | [116]–[118] |
2-4 | Future event inference | Circular binary search tree | [115], [118]–[120] |
Semantic model | Feature-based | Aggregation-based | [48], [121]–[123] |
3-4 | Compositional-based | [56], [124] | |
3-4 | Markov models | [32], [44], [125]–[127] | |
2-4 | Prototype-based | Clustering-based | [128]–[130] |
2-4 | Attribute-based | Feature extraction-based | [97], [131]–[133] |
2-4 | Descriptive-based | Text-based | [111], [134]–[138] |
Spatial-temporal methods are used to predict public emergencies by incorporating spatial and temporal information about emergencies. For instance, a spatio-temporal model is proposed to predict emergency incidence [139]. It considers the correlation between the spatial and temporal dimensions of emergencies and incorporates spatial factors such as population density and road network density. A multi-view learning-based spatio-temporal model is proposed, which learns the features of spatial and temporal domains separately and then fuses them for prediction [140]. Another approach is to leverage the deep learning model for prediction [95]. A spatiotemporal Graph Convolutional Network is proposed that learns the spatial and temporal correlations of emergencies using graph convolutional networks as shown in Figure [fig:my95label].
Figure 5: Tensor Decomposition and Forecasting for Complex Time-stamped events
Methods for forecasting public emergencies based on time series are divided into two categories:
Direct methods typically approach the problem of predicting emergency types as a multivariate time series forecasting problem, where each variable corresponds to an emergency type \(E_i(i=1, ...)\) and the predicted emergency type at future time \(t\) is calculated as \(\hat{s}_{\hat{t}} = argmax{E_i}\) \(f(s_{\hat{t}} = E_i|X)\). For instance, in [141], a longitudinal support vector regressor is utilized to forecast multi-dimensional emergencies. They build \(n\) Support Vector Regressors, each of which corresponds to a specific attribute to achieve the goal of predicting the next time point’s attribute value. To predict multiple emergency types, Weiss and Page [142] use multiple-point process models. To improve the accuracy of their predictions, Biloš et al. [143] employ RNN to learn the historical representation of emergencies, and then input the results into a Gaussian process model to predict future emergency types. To better capture the dynamics across multiple variables in the time series, Brandt et al. [144] extend this to Bayesian vector autoregression.
Indirect methods focus on learning a mapping from observed emergency types to low-dimensional latent-topic space using tensor decomposition-based techniques. Regarding information cascade prediction under public emergencies, Matsubara et al. [145] propose a 3-way analysis of the original observed emergency tensor \(Y_0 \in R^{D_oD_aD_e}\), which consists of three factors: actors, objects, and time. They decompose this tensor into latent variables via three corresponding low-rank matrices \(P_o\in R^{D_kD_o}\), \(P_a \in R_{D_kD_a}\), and \(P_e \in R_{D_k*D_e}\), where \(D_k\) is the number of latent topics. Through multivariate time series forecasting, the time matrices \(P_e\) are predicted into the future to estimate future emergency tensors by recovering a "future emergency tensor" \(\hat{Y}\) through the multiplication of the predicted time matrix \(P_e\) with the known actor matrix \(P_a\) and object matrix \(P_o\). This approach provides insights into the different types of emergencies based on latent topics.
The prediction of the location and time of future public emergencies is a critical area of research for emergency management and disaster response. One relevant category of methods for such predictions is based on the joint consideration of time and location. These methods are classified into two subtypes based on their approach as shown in Figure 6.
Figure 6: The Type Of Location-based Prediction
Grid-based methods are useful in predicting the location and time of future public emergencies. In recent years, several techniques have been proposed to capture spatial and temporal information for information cascade prediction under public emergencies as shown in Table 5.
0.1cm -0.2cm
Category | Description | References |
---|---|---|
Spatial Clustering | Grouping contiguous regions that collectively exhibit significant patterns. | [88], [146], [147] |
Spatial Interpolation | Estimating the probability of event occurrence at locations without historical data, resulting in spatial smoothness. | [71], [81], [130], [146], [148] |
Spatial Convolution | Learning and representing spatial patterns using Convolutional Neural Networks (CNNs). | [46], [74], [86], [132], [149], [150] |
Trajectory Destination | Predicting population-based events by interpreting the collective behaviors of individuals. | [48], [84], [91], [147] |
A straightforward approach to incorporating spatial information is to include location data as an input feature and use it in predictive models, such as Linear Regression [151], LSTM [152], and Gaussian Processes [62]. Zhao et al. [151] utilized the spatiotemporal dependency to regularize their model parameters during training. However, most of the methods in this domain aim to jointly consider the spatial and temporal dependencies for predictions [153]. To achieve this, the multi-attributed spatial information for each time point is organized as a series of multi-channel images that are encoded using convolution-based operations[154].
The multi-attributed spatial information is represented as a spatiotemporal grid, where each cell in the grid represents a specific location and time point. Let \(X_{i,j,t}\) denote the value of the \(i\)-th attribute at the \((i,j)\)-th location and \(t\)-th time point. The spatiotemporal grid is represented as: \[X = [X_{i,j,t}]_{i=1,\dots,n; j=1,\dots,m; t=1,\dots,T}\] where \(n\) and \(m\) represent the number of locations in the grid, and \(T\) represents the number of time points.
To encode the spatiotemporal grid using convolution-based operations, it first converts it into a series of multi-channel images, where each channel corresponds to a different attribute. Let \(X^{(k)}\) denote the \(k\)-th channel of the image, which is a 2D matrix of size \(n \times m\): \[X^{(k)} = [X_{i,j,t}^{(k)}]_{i=1,\dots,n; j=1,\dots,m}\]
The 2D convolution operation is then applied to each channel of the image to learn spatial features. The output of the convolution operation is a 2D feature map, which is then passed through a nonlinear activation function. Let \(W^{(k)}\) denote the weight matrix for the \(k\)-th channel, and \(b^{(k)}\) denote the bias vector. The output of the convolution operation is expressed as: \[H^{(k)}_{i,j,t} = f\left(\sum_{p=1}^{P}\sum_{q=1}^{Q} W^{(k)}_{p,q} X_{i+p-1,j+q-1,t} + b^{(k)}\right)\] where \(P\) and \(Q\) represent the size of the convolution kernel, and \(f\) is the activation function.
The output of the convolution operation is then passed through a pooling layer to reduce the spatial dimensionality of the feature map. The pooling operation computes a summary statistic over a local neighborhood of the feature map. The most common pooling operations are max pooling and average pooling. Let \(H^{(k)}_{i,j,t}\) denote the output of the pooling operation for the \(k\)-th channel. The output of the pooling layer is expressed as: \[Y^{(k)}_{i,j,t} = \text{pool}\left(\{H^{(k)}_{p,q,t}\}_{p=i,\dots,i+K-1; q=j,\dots,j+K-1}\right)\] where \(K\) represents the size of the pooling window, and \(\text{pool}\) is the pooling function.
The outputs of the pooling layer for channels are concatenated to form a spatiotemporal feature vector. Let \(Y_t\) denote the spatiotemporal feature vector at time \(t\). The spatiotemporal feature vector is expressed as: \[Y_t = [Y^{(1)}_{1,1,t}, \dots, Y^{(1)}_{n,m,t}, Y^{(2)}_{1,1,t}, \dots, Y^{(2)}_{n,m,t}, \dots, Y^{(K)}_{1,1,t}, \dots, Y^{(K)}_{n,m,t}]\]
Spatio-temporal point process modeling is a valuable technique for predicting the time and location of public emergencies. This method models the rate of event occurrence in both space and time, providing a comprehensive understanding of the spatiotemporal distribution of events. The technique is defined as: \[\lambda (t,l|X)=\lim_{|dt|\rightarrow 0,|dl|\rightarrow 0} \frac{E[N(dt*dl)|X]}{(|dt||dl|)}\] Several models have been proposed to instantiate this framework, which includes different facets of input data such as location, time, and other semantic features. Liu and Brown et al. [155] assumed conditional independence among spatial and temporal factors and decomposed the rate of event occurrence as follows: \[\lambda (t,l|X)=\lambda (t,l|L,T,F)=\lambda_1 (l|L,T,F,t)*\lambda_2(t|T)\] In this equation, \(\lambda_1(x)\) is modeled using the Markov spatial point process, while \(\lambda_2(x)\) is characterized using temporal autoregressive models. To handle situations where explicit assumptions for model distributions are difficult, several methods have been proposed to involve deep architecture during the point process. Recently, Okawa et al. [156] proposed the following: \[\lambda(t, l \mid X)=\int g_{\theta}\left(t^{\prime}, l^{\prime}, \mathcal{F}\left(t^{\prime}, l^{\prime}\right)\right) \cdot \mathcal{K}\left((t, l),\left(t^{\prime}, l^{\prime}\right)\right) \mathrm{d} t^{\prime} \mathrm{d} l^{\prime}\] Here, \(K(x,y)\) is a kernel function, such as a Gaussian kernel [157], which measures the similarity in time and location dimensions. \(F(t,l)\subseteq F\) denotes the feature values for the data at location \(l^{'}\) and time \(t^{'}\). \(g_{\theta}(x)\) is a deep neural network that parameterized by \(\theta\) and returns a nonnegative scalar. The model selection of \(g_{\theta}(x)\) depends on the data types.
Category | Directions | Methods | References |
---|---|---|---|
*System-based | Event prediction | Model-fusion system | [105], [158], [159] |
2-4 | Crowd-sourced system | Recommender system | [67], [160], [161] |
2-4 | Future event detection | Prediction market system | [96], [141], [162] |
*Model-based | *Planned event detection | Mathematical models | [61], [163], [164] |
3-4 | Content filtering | [165], [166] | |
3-4 | Time expression identification | [165], [167] | |
3-4 | Future reference sentence extraction | [90], [166] | |
3-4 | Location identification | [40], [73], [168], [169] | |
*Tensor-based | *Tensor extrapolation | Tensor decomposition | [109], [163], [170] |
3-4 | Tensor completion | [151], [170] |
In this section, we discuss various techniques that are utilized for predicting the time, location, and Semantics of public emergencies. These methods are broadly classified into two categories: System-based approaches rely on the development and deployment of complex systems that are designed to gather and analyze data from various sources in real time. On the other hand, Future Event Detection Methods approaches involve the development of mathematical models that use historical data to predict future emergencies. Both approaches have their own strengths and limitations, and the choice of strategy depends on the specific needs and constraints of the emergency management system.
Public emergencies are predicted using system-based or model-based strategies. In this section, we discuss system-based approaches. One such approach is the model-fusion system, which integrates various techniques for predicting the time, location, and semantics of public emergencies into a unified emergencies prediction system [171].
Cascaded prediction of emergency events is crucial for timely and effective response. EMBERS [158] is an online warning system that predicts the time, location, and type of future events, as well as the population affected. To maximize precision, EMBERS prioritizes individual prediction models by suppressing their recall [172]. The fusion of predictions from different models eventually results in high recall. Bayesian fusion-based strategies have been investigated [159], and similar strategies are used in systems like Carbon [105]. Crowd-sourced systems are another approach that implements fusion strategies to generate predictions made by human predictors. For instance, Rostami et al. [160] proposed a recommender system that matches event-predicting tasks to human predictors with suitable skills to maximize the accuracy of their fused predictions. This approach addresses the heterogeneity and diversity of human predictors’ skill sets and background knowledge under limited human resources.
Planned public emergencies are detected using methods that rely on Natural Language Processing (NLP) techniques and Linguistic Principles to analyze various media sources such as social media and news. These methods typically consist of four main steps.
Content filtering is employed to retain the texts that are relevant to the event of interest. Existing works utilize either supervised methods or unsupervised methods.
Time expression identification is used to identify future reference expressions and determine the time of the event. This step leverage existing tools such as the Rosetta text analyzer [167] or propose dedicated strategies based on linguistic rules [165].
Future reference sentence extraction is the core of planned event detection and is implemented either by designing regular expression-based rules [166] or by textual classification [90].
Location identification is essential to infer the event’s location accurately. The expression of locations is typically highly heterogeneous and noisy. Existing works have relied heavily on geocoding techniques to resolve the event location accurately. Various types of locations are considered, such as article locations, locations mentioned in the articles [168], and authors’ neighbors’ locations [73]. Multiple locations have been selected using a geometric median [40] or fused using logical rules such as probabilistic soft logic [169].
Public emergencies are often analyzed using tensor decomposition and extrapolation techniques. These methods involve formulating the data into a tensor form, which includes dimensions such as location, time, and semantics. Tensor decomposition is then applied to approximate the original tensor by using multiple low-rank matrices, each of which represents a mapping from latent topics to each dimension [173]. Specifically, given a tensor \(\mathcal{T}\) with dimensions \(I\), \(J\), and \(K\), the low-rank tensor approximation is represented as: \[\mathcal{T} \approx \sum_{r=1}^{R} \mathbf{A}^{(1)}_{r} \circ \mathbf{A}^{(2)}_{r} \circ \mathbf{A}^{(3)}_{r}\] where \(\mathbf{A}^{(1)}_{r} \in \mathbb{R}^{I \times R}\), \(\mathbf{A}^{(2)}_{r} \in \mathbb{R}^{J \times R}\), and \(\mathbf{A}^{(3)}_{r} \in \mathbb{R}^{K \times R}\) are low-rank matrices, \(\circ\) denotes the outer product, and \(R\) is the rank of the tensor.
After the tensor is decomposed, various strategies are employed to extrapolate the tensor toward future periods. For instance, Mirtaheri [163] extrapolated the time dimension matrix and then multiplied it with the other dimensions’ matrices to recover the estimated extrapolated tensor into the future: \[\mathbf{A}^{(3)}_{r}(t_f) \approx \mathbf{A}^{(3)}_{r}(t_n) + (t_f - t_n) \Delta \mathbf{A}^{(3)}_{r}\] where \(\mathbf{A}^{(3)}_{r}(t_f)\) and \(\mathbf{A}^{(3)}_{r}(t_n)\) are the time dimension matrices at the future period \(t_f\) and the last observed period \(t_n\), respectively, and \(\Delta \mathbf{A}^{(3)}_{r}\) is the first-order difference matrix of the time dimension matrix.
On the other hand, Zhou et al. [170] adopted a different approach, where they added "empty values" for the entries corresponding to a future time in the original tensor. Then, they utilized tensor completion techniques to infer the missing values that correspond to future events. These techniques include low-rank tensor completion and tensor regression, which aim to recover the missing entries by exploiting the low-rank structure and the relationships between the dimensions of the tensor. By extrapolating the tensor towards future periods, decision-makers better understand the potential impact of such events and prepare appropriate response plans.
Vulnerability prediction in the context of public emergencies refers to the use of machine learning models to identify and assess potential risks and vulnerabilities associated with various hazards, such as earthquakes, floods, landslides, and extreme weather events.
In public emergencies, innovative solutions are required, and machine learning (ML) models have emerged as a promising approach to addressing the challenges posed by these events.
Prasad et al. [174] utilized an ML-based ensemble technique that combines multiple models to improve the accuracy of flood vulnerability mapping. The ensemble method is expressed mathematically as follows: \[F(x) = \sum_{t=1}^{T} w_t f_t(x)\] where \(x\) represents the input data, \(F(x)\) is the output of the ensemble model, \(f_t(x)\) is the output of the \(t\)-th base model, and \(w_t\) is the weight assigned to the \(t\)-th base model. The weights are typically computed based on the performance of the base models on the training data.
Similarly, Nsengiyumva and Valentino [175] employed the logistic model tree (LMT) to predict the vulnerability of landslide areas. The LMT is expressed mathematically as follows: \[p(C_k|x) = \frac{N_k(x)}{N(x)}\] where \(p(C_k|x)\) is the probability of the \(k\)-th class given the input data \(x\), \(N_k(x)\) is the number of training instances in class \(k\) that satisfy the conditions specified by the tree, and \(N(x)\) is the total number of training instances that satisfy the conditions specified by the tree.
Association-based impact prediction in the context of public emergencies involves the use of association Rule-based methods to predict future events accurately. These methods rely on discovering associations between precursor and target events to predict the onset of various hazards, such as pandemics.
The Apriori algorithm is a popular algorithm for frequent set mining, which generates frequent item sets by iteratively joining smaller itemsets and pruning infrequent ones. Mathematically, the support of an itemset \(I\) is defined as: \[Supp(I) = \frac{freq(I)}{N}\] where \(freq(I)\) is the number of transactions that contain the itemset \(I\), and \(N\) is the total number of transactions.
Association rule mining involves finding rules of the form \(X\rightarrow Y\), where \(X\) is a set of items (the precursor event), and \(Y\) is a set of items (the target event). The association rules \(X\rightarrow Y\) are said to be valid if their support and confidence are above a given threshold. Pruning strategies are employed to remove invalid rules and retain the most accurate rules. The lift measure is another commonly used metric that measures the strength of association between the precursor event \(X\) and the target event \(Y\) is defined as: \[lift(X\rightarrow Y) = \frac{supp(X\cup Y)}{supp(X)supp(Y)}\] where \(supp(X\cup Y)\) is the support of the itemset \(X\cup Y\). A lift value greater than 1 indicates a positive association between \(X\) and \(Y\). In conclusion, association-based methods effectively predict events during public emergencies by discovering associations between precursor and target events.
Causality-based prediction involves inferring cause-effect relationships among historical emergency events and utilizing this knowledge to predict future emergencies. This approach involves Emergency Semantic Representation, Emergency Causality Inference, and Future Emergency Inference.
The first step in applying causality-based prediction to public emergencies is to extract relevant emergency events from various sources, including news articles, social media, and government reports, using natural language processing techniques. This process involves several approaches to represent emergency events in a meaningful way. One approach is the Event Phrase-based method, where the emergency event is represented as a phrase extracted from the text [176]. For instance, if the text mentions a "terrorist attack in downtown," the phrase "terrorist attack in downtown" represents the event. Another approach is the Event Keywords-based method, where keywords extracted from the text are used to represent the emergency event [56], [145]. A third approach is the Tuple-based method [72], where each emergency event is represented by a tuple consisting of objects, a relationship, and time. For instance, if the text mentions a "car crash involving two vehicles on the highway at 4 pm," the tuple representation could be (car crash, involving, two vehicles, highway, 4 pm). An RDF-based format is utilized in some cases [149]. Mathematically, a tuple-based representation of an emergency event is expressed as \(E = (O, R, T)\). where \(O\) represents the objects involved in the event, \(R\) represents the relationship between the objects, and \(T\) represents the time of the event. In the case of the car crash example mentioned earlier, the tuple representation is expressed as \(E = (\text{car crash, two vehicles, highway}, \text{involving}, \text{4 pm})\)
In this stage, the focus is on inferring cause-effect relationships among historical emergencies. The first step is to cluster the emergency events into emergency chains, sequences of time-ordered events with the same topics, actors, and objects [145]. Once these chains are established, various approaches are used to infer the causal relationships among the emergency pairs. One common approach is to use natural language processing techniques to identify causal mentions, such as causal connectives, prepositions, and verbs [145], and then extract the causal relationships from the text. Another approach is to formulate causal-effect relationship identification as a classification task, where the inputs are the candidate events, and contextual information is incorporated, including related background knowledge from web texts [135].
Future Emergency Inference After learning the cause-effect relationships among historical emergency events and representing them semantically, they use this knowledge to predict future emergencies. This enables us to anticipate potential emergency situations and take preventative measures to mitigate their impact.
Retrieve similar emergencies To predict future emergencies, first search for similar events in the historical emergency event pool. They use various similarity measures, such as the Euclidean distance or cosine similarity, to compare the query event to historical events. Contextual information, such as the emergency time, location, and other relevant environmental and descriptive information, is taken into account. Specifically, use the following formula to calculate the Euclidean distance: \[d(E_1, E_2) = \sqrt{\sum_{i=1}^{n}(x_{1,i}-x_{2,i})^2}\] where \(E_1\) and \(E_2\) are two emergency events being compared, \(x_{1,i}\) and \(x_{2,i}\) are the values of feature \(i\) for \(E_1\) and \(E_2\), respectively, and \(n\) is the total number of features being compared.
Similarly, use the cosine similarity to calculate the similarity between two emergency events: \[sim(E_1, E_2) = \frac{\sum_{i=1}^{n} x_{1,i} x_{2,i}}{\sqrt{\sum_{i=1}^{n} x_{1,i}^2} \sqrt{\sum_{i=1}^{n} x_{2,i}^2}}\] where \(E_1\) and \(E_2\) are two emergency events being compared, \(x_{1,i}\) and \(x_{2,i}\) are the values of feature \(i\) for \(E_1\) and \(E_2\), respectively, and \(n\) is the total number of features being compared. Then rank the retrieved emergency events based on their similarity to the query event. This approach allows us to identify potential future emergencies and take proactive steps to prevent or mitigate their impact.
Infer the future emergenciesThe next step is to determine the potential consequences caused by the query emergency based on the causality of emergencies learned from historical data. They traverse the abstraction tree that represents the learned causal relationships, starting from the root that corresponds to the most general emergency rule. The search frontier then moves across the tree if the child node is more similar, culminating in the nodes that are the least general but still similar to the new event being retrieved. Specifically, use the following formula to determine the similarity between two nodes in the abstraction tree: \[sim(n_1, n_2) = \frac{w_{1,2}}{\sqrt{w_{1,1}} \sqrt{w_{2,2}}}\] where \(n_1\) and \(n_2\) are two nodes being compared, \(w_{1,2}\) is the weight of the edge connecting nodes \(n_1\) and \(n_2\), \(w_{1,1}\) is the sum of the weights of edges connected to node \(n_1\), and \(w_{2,2}\) is the sum of the weights of edges connected to node \(n_2\) [177].
Since each case, event lead to multiple emergency events, use various approaches to determine the final prediction. For example, calculate the support or conditional probability of the rules using the following formula: \[P(E|C) = \frac{P(C|E)P(E)}{P(C)}\] where \(P(E|C)\) is the probability of emergency event \(E\) given case event \(C\), \(P(C|E)\) is the probability of case event \(C\) given emergency event \(E\), \(P(E)\) is the prior probability of emergency event \(E\), and \(P(C)\) is the prior probability of case event \(C\).
A ranking approach based on the similarity between the new emergency event and historical emergency events is employed. The similarity is defined by the length of their minimal generalization path, which is calculated using the following formula: \[d(E_1, E_2) = \sum_{i=1}^{k} \frac{1}{2^i}\]
where \(E_1\) and \(E_2\) are two emergency events being compared, and \(k\) is the length of their minimal generalization path. The formula gives greater weight to earlier levels in the path, as the denominator increases exponentially with each level. This approach allows us to rank emergency events based on their similarity to the new event, and use this ranking to make a prediction [72], [145].
Large Language Models (LLMs) have great potential in addressing public emergencies, such as natural disasters and pandemics. LLMs provide real-time reports, predict outcomes, and analyze social media data to identify areas that are most affected. They also identify and address misinformation and fake news by analyzing social media data and news articles. Large Language Models, including the autoregressive integrated moving average (ARIMA) model, have demonstrated significant potential in predicting natural disasters through data analysis.
In public emergency prediction, it would be predicting the outbreak of a disease based on the language used in social media posts. A formula that could be used in this category is the autoregressive integrated moving average (ARIMA) model: \[Y_t = c + \sum_{i=1}^p\phi_iY_{t-i} + \sum_{i=1}^q\theta_i\varepsilon_{t-i} + \varepsilon_t\] where \(Y_t\) is the number of social media posts at time \(t\), \(c\) is a constant, \(\phi_i\) and \(\theta_i\) are the autoregressive and moving average parameters, respectively, and \(\varepsilon_t\) is the error term at time \(t\).
Then, it would be predicting the location and time of a flood based on historical data on weather patterns and water levels. A formula that could be used in this category is the space-time autoregressive integrated moving average (STARIMA) model:
\[Y_{s,t} = c + \sum_{i=1}^p\sum_{j=1}^k\sum_{l=1}^m\phi_{i,j,l}Y_{s-i,t-j,l} + \sum_{i=1}^q\sum_{j=1}^k\sum_{l=1}^m\theta_{i,j,l}\varepsilon_{s-i,t-j,l} + \varepsilon_{s,t}\] where \(Y_{s,t}\) is the water level at spatial location \(s\) and time \(t\), \(\phi_{i,j,l}\) and \(\theta_{i,j,l}\) are the autoregressive and moving average parameters, respectively, and \(\varepsilon_{s,t}\) is the error term at spatial location \(s\) and time \(t\).
After that, it would be predicting a wildfire’s location, time, and nature based on satellite imagery and textual data on weather patterns. A formula that could be used in this category is the space-time-text autoregressive model (STTAR) model:
\[Y_{s,t} = c + \sum_{i=1}^p\sum_{j=1}^k\sum_{l=1}^m\phi_{i,j,l}Y_{s-i,t-j,l} + \sum_{i=1}^q\sum_{j=1}^k\sum_{l=1}^m\theta_{i,j,l}\varepsilon_{s-i,t-j,l} + \beta X_{s,t} + \varepsilon_{s,t}\] where \(Y_{s,t}\) is the severity of the wildfire at spatial location \(s\) and time \(t\), \(\phi_{i,j,l}\) and \(\theta_{i,j,l}\) are the autoregressive and moving average parameters, respectively, \(X_{s,t}\) is a vector of weather and environmental variables at spatial location \(s\) and time \(t\), \(\beta\) is a vector of regression coefficients, and \(\varepsilon_{s,t}\) is the error term at spatial location \(s\) and time \(t\).
In Vulnerability Prediction, it would be predicting a community’s vulnerability to a hurricane based on demographic and socio-economic data. A formula that could be used in this category is the vulnerability index:
\[VI = \sum_{i=1}^n w_i \cdot (1 - z_i)\] where \(w_i\) is the weight assigned to vulnerability factor \(i\), \(z_i\) is the standardized value of vulnerability factor \(i\), and \(n\) is the number of vulnerability factors.
In Association-based Impact Prediction, it would be predicting the impact of a hurricane based on its association with other factors, such as wind speed and storm surge [178]. A formula that could be used in this category is the association rule [179]: \(\text{hurricane} \rightarrow \text{damage}\), where the hurricane is the antecedent and the damage is the consequent. In Causality-based Prediction, it would be predicting the likelihood of a wildfire based on its causal factors, such as drought and lightning strikes [180]. A formula that could be used in this category is the Bayesian network:
\[P(W|D,L) = \frac{P(W)P(D,L|W)}{P(D,L)}\] where \(W\) is the event of a wildfire occurring, \(D\) is the event of a drought occurring, \(L\) is the event of lightning strikes occurring and \(P(W|D,L)\) is the probability of a wildfire occurring given the occurrence of a drought and lightning strikes. \(P(W)\) is the prior probability of a wildfire occurring, \(P(D,L|W)\) is the conditional probability of a drought and lightning strikes occurring given that a wildfire has happened, and \(P(D,L)\) is the joint probability of a drought and lightning strikes occurring.
This section introduces the applications of public emergency management. By dividing the applications of public emergency management into different phases, this section is divided into (1) Early Warning, (2) Disaster Monitoring, (3) Damage Assessment, and (4) Disaster Response.
0.1cm
Category | Direction | Method | References |
---|---|---|---|
*Early Warning | Sensing devices process data | PCA, Logistic Regression, CNN, RNN | [181]–[184] |
2-4 | Sense the location of emergencies | GAN, RF, Image recognition | [185]–[188] |
2-4 | Enhance the accuracy and speed | TLS, ANN, Fuzzy Deep Neural Network | [189]–[191] |
2-4 | Utilize communication channels | Mining information dissemination data | [188], [189] |
*Disaster Monitoring | *Element Sensing | UAV-based sensing | [192]–[194] |
3-4 | Visual sensing | [195]–[197] | |
2-4 | *Situation Sensing | Crowdsourced sensing | [198], [199] |
3-4 | Sensor Network-based System | [194], [196], [200] |
Early warning is a critical phase in public emergency management, aimed at detecting and predicting potential emergencies and providing timely alerts to relevant authorities and the public. To achieve early warning, various methods have been developed, including data mining, machine learning, and predictive modeling. For example, in the case of a flood, data from sources such as rainfall sensors, river level gauges, and weather forecasts are analyzed to predict the flood’s location, intensity, and possible impacts [201]. Early identification and warning of public emergencies are performed through Sensor Networks, Remote Sensing, Social Media, and other communication channels.
Effective early warning of public emergencies relies on processing data acquired by sensing devices promptly. Moon et al. [183] proposed a machine learning method for effective early warning of short-term rainfall, while He et al. [182] utilized machine learning techniques to extract data from tweets and perform fusion analysis for rainstorm disasters in real-time. Perol et al. [184] proposed a Convolutional Neural Network-based method for early detection and warning of earthquake disasters, and Chin et al. [181] adopted a Recurrent Neural Network model for earthquake early warning systems. Zheng et al. [191] proposed a fuzzy deep neural network for early warning of industrial accidents.
Early warning of public emergencies is critical in minimizing their impact. Seismic wave analysis has been used to sense the location of earthquakes, where Li et al. [186] used a Generative Adversarial Network. Ethan et al. [188] performed image recognition on natural disaster images from social media to provide early warning of public emergencies such as earthquakes, floods, and wildfires. Machine learning models have also been used to sense the severity of flood events in videos. Huang et al. [185] proposed a deep belief network method for meteorological early warning of precipitation-induced landslides, achieving precise sensing of emergencies.
Early warnings provide valuable information for governments, communities, and individuals to take appropriate actions to mitigate the impact of public emergencies. It also enables real-time sensing of conditions on land and sea using advanced computer numerical models. Ultimately, continuous improvement of early warning systems for public emergencies such as weather, climate, traffic, and epidemics is essential for effective disaster response.
Continuous monitoring and assessment of public emergencies are crucial for providing emergency responders with crucial information to make informed decisions on response strategies. Disaster monitoring tools and technologies, such as Remote Sensing, Geographic Information Systems (GIS), and Unmanned Aerial Vehicles (UAVs), have been used to track and assess natural and man-made disasters.
UAV-based sensing is a promising solution for sensing in public emergencies, with research focusing on detecting and monitoring gas leaks, floods, and forest fires [194]. Sensor networks are used to collect data on gas concentrations, water levels, and temperature and humidity levels, which are transmitted to a central server for analysis, providing real-time information to emergency responders [192], [193]. Visual sensing using VR and MR technologies, as well as crowdsourced sensing through social media platforms, are also valuable for sensing in public emergencies [94], [105], [195]. Advances in mobile technologies and social media have made it easier to collect data from citizens in affected areas, providing valuable information to emergency responders and improving the response time and effectiveness of emergency services [198].
0.1cm
Category | Directions | Methods | References |
---|---|---|---|
Obtaining event data | Weakly supervised method | [200], [202], [203] | |
2-4 | Vulnerabilities in sensing devices | UAV-based sensing | [192], [194] |
2-4 Element sensing | Energy issues in fixed sensors | Crowdsourced sensing | [193], [198] |
2-4 | Privacy issues in social media sensing | UAV-based sensing | [196], [197] |
2-4 | Regional restrictions for sensing | Visual sensing | [94], [105] |
Efficient search for accurate information | Non-negative matrix factorization | [192], [193], [204], [205] | |
2-4 | Identifying sub-events | Unsupervised learning framework | [194], [203], [206] |
2-4 Situational | Sensing data during GPS failures | Social media sensor monitoring | [195], [207], [208] |
2-4 understanding | Real-time image processing | Image processing pipeline | [200], [209], [210] |
2-4 | Combining heterogeneous data sources | Context-aware fusion method | [197], [211], [212] |
2-4 | Transfer learning for COVID-19 | Convolutional neural networks | [196], [205], [209], [213] |
Situational understanding is critical in emergency response, as it improves the level of sensing and the depth of understanding, ultimately helping in managing public emergencies. However, situational sensing does not exist in isolation and must be assisted by element sensing to obtain event data.
To deal with noisy data during public emergencies, weakly supervised methods and contextual messages are used to enrich message representations and achieve situational understanding. Cheng et al. [192] studied a novel topic-tracking problem and enabled efficient search for accurate information through an online non-negative matrix factorization scheme. Situation-sensing algorithms have been designed to automatically identify important sub-events during public emergencies. Social media sensors are useful for monitoring natural disasters and real-time understanding of image content, but filtering out irrelevant parts of images and combining heterogeneous data sources is required [200], [202], [206], [207], [211].
0.1cm
Category | Directions | Methods | References |
---|---|---|---|
Social Media Analysis | DL-based multimodal approach | [214], [215] | |
2-4 | Temporal and Spatial Analysis | Latent Dirichlet Allocation (LDA) | [216], [217] |
2-4 | Aerial Image Analysis | Pre-trained DL CNN and ML | [218], [219] |
2-4 Damage | Flood Damage Classification | K-means clustering and SVM | [220], [221] |
2-4 Assessment | Damage Classification | CNN architecture | [222], [223] |
2-4 | Deep CNN-based | Deep CNN | [215], [224] |
2-4 | Multi-modal Classification | Two-stage multi-modal | [219], [225] |
2-4 | Damage Identification | CNNs and class activation maps | [226], [227] |
*Disaster Response | Public emergencies response | Drones, telemedicine, mobile apps | [73], [146], [228] |
2-4 | Relief aid supply | Dynamic calculation | [214], [229], [230] |
2-4 | Social media data analysis | ML-based supervised models | [231]–[233] |
2-4 | Medical rescue | Decision table, genetic algorithm | [221], [234], [235] |
2-4 | Image classification | DL methods, ML algorithms | [227], [236], [237] |
2-4 | Social media data classification | Adaptation classifiers, LSTM, CNN | [225], [232], [238], [239] |
Damage assessment is crucial for effective emergency management. Techniques such as social media analysis and image processing provide real-time information to evaluate the extent of damage and plan resource deployment to mitigate the impact of the emergency on the affected population, infrastructure, and environment.
Researchers have proposed various ML and DL techniques to minimize the impact of natural disasters. These techniques include analyzing social media data, aerial images, and satellite data to evaluate disaster situations and assess damage caused by natural disasters. Some of the techniques proposed include DL-based multimodal approach [214], ML techniques combined with temporal and spatial analysis of social media posts [216], hybrid approaches that classify areas affected by floods [220], CNN architectures for damage classification [219], [222], [226], and DL-based approaches to assess the impact of water-related disasters using satellite image data [224]. These techniques have shown promising results in identifying and assessing damage in disaster-hit areas. Overall, these proposed approaches using ML and DL techniques demonstrate their potential to improve disaster management and minimize the impact of natural disasters on the environment, infrastructure, and human lives. Damage assessment techniques provide accurate and timely information that is critical for planning and deploying resources to mitigate the impact of public emergencies.
Public emergency response is the phase of emergency management that involves the deployment of resources to mitigate the impact of the emergency and assist affected populations. This phase involves various activities, including medical assistance, evacuation, and infrastructure management.
ML and DL techniques have been used to improve public emergency response and disaster management. These techniques include the use of MLP NN to estimate relief supplies [146], decision tables to manage medical rescue [234], and unmanned aerial vehicles for recognizing the status of individuals in disaster-struck areas [221]. ML algorithms have also been used to analyze social media data related to public emergencies [225], [231], [232], [235], [238], [239] and classify images from earthquake-hit areas [227]. These technologies have been tested on various datasets and for different types of disasters, aiding decision-making and response efforts. Technologies such as unmanned aerial vehicles, telemedicine, and mobile applications play a vital role in improving the efficiency and effectiveness of public emergency response and disaster management.
This study provides a comprehensive review of research papers focused on the prediction of public emergencies. The reviewed articles covered a broad range of research areas related to the use of social media data, machine learning models, and natural language processing techniques for analyzing emergency events and developing predictive models for different types of disasters. Despite significant progress in predicting information cascades during public emergencies, there are still several open questions and directions for future research to explore:
Future work in emergency response includes developing predictive algorithms that consider ethical and social implications. This requires using fairness, transparency, and accountability metrics to evaluate the impact on different populations, and creating ethical frameworks to guide AI use. Predictive algorithms must also account for complex and dynamic public emergencies, using advanced modeling techniques to capture interactions between emergency responders, the public, and stakeholders.
Predicting the popularity and spread of emergency information cascades is a challenging task, and there are fundamental questions that have yet to be answered. Additionally, it is crucial to understand the mechanisms that govern the success of emergency information dissemination [240]. Interpretability of the predictions is also essential for building trust and accountability. However, these models are complex and challenging to interpret, which leads to issues with trust and accountability. Future research should focus on developing more interpretable models that are easily understood by human operators and provide a clear explanation of the reasoning behind their predictions.
Information cascades during public emergencies are easily influenced or manipulated by malicious actors, leading to inaccurate predictions and potentially harmful outcomes. External stimuli such as breaking news and rumors significantly affect the spread of emergency information. Modeling these external stimuli is important for improving the robustness of prediction models. Cross-domain real-time transfer learning and retrieval of information from other platforms are used to model external stimuli. Additionally, analyzing the sources of external stimuli provide insights into the future evolution of emergency information dissemination. Future research should aim to develop more robust models that handle noisy and adversarial input data and provide accurate predictions even in the presence of interference.
The field of Human-AI teaming has enormous potential in addressing public emergencies, such as natural disasters, pandemics, and terrorist attacks. This requires using fairness, transparency, and accountability metrics to evaluate the impact on different populations, and creating ethical frameworks to guide AI use. Predictive algorithms must also account for complex and dynamic public emergencies, using advanced modeling techniques to capture interactions between emergency responders, the public, and stakeholders.
Future research should focus on developing multimodal, large-scale models that can effectively integrate and process diverse sources of information to generate accurate and timely predictions during sudden public events. These models should leverage deep learning techniques, such as natural language processing, computer vision, and speech recognition, to analyze different data modalities and identify patterns and trends. Incorporating contextual information and domain knowledge could further enhance their performance, ultimately improving our ability to anticipate and respond to sudden public events.
This paper offers a comprehensive and systematic overview of existing techniques and methods for predicting emergency information cascades. The presented taxonomy serves as a valuable resource for domain experts when selecting the appropriate technique for a specific problem set. In particular, this paper analyzes the characteristics and methods of predicting information cascades during public emergencies from three perspectives: information cascade modeling, prediction, and application. These methods encompass a range of features and models, including time, location, and semantics, and incorporate the latest advances in modeling and predicting information cascades to provide a comprehensive and up-to-date overview of the field. Finally, we also discuss current research fronts, identify bottlenecks, pitfalls, and unresolved issues, and outline potential future research directions.