164 research outputs found

    Spatiotemporal Propagation Learning for Network-Wide Flight Delay Prediction

    Full text link
    Demystifying the delay propagation mechanisms among multiple airports is fundamental to precise and interpretable delay prediction, which is crucial during decision-making for all aviation industry stakeholders. The principal challenge lies in effectively leveraging the spatiotemporal dependencies and exogenous factors related to the delay propagation. However, previous works only consider limited spatiotemporal patterns with few factors. To promote more comprehensive propagation modeling for delay prediction, we propose SpatioTemporal Propagation Network (STPN), a space-time separable graph convolutional network, which is novel in spatiotemporal dependency capturing. From the aspect of spatial relation modeling, we propose a multi-graph convolution model considering both geographic proximity and airline schedule. From the aspect of temporal dependency capturing, we propose a multi-head self-attentional mechanism that can be learned end-to-end and explicitly reason multiple kinds of temporal dependency of delay time series. We show that the joint spatial and temporal learning models yield a sum of the Kronecker product, which factors the spatiotemporal dependence into the sum of several spatial and temporal adjacency matrices. By this means, STPN allows cross-talk of spatial and temporal factors for modeling delay propagation. Furthermore, a squeeze and excitation module is added to each layer of STPN to boost meaningful spatiotemporal features. To this end, we apply STPN to multi-step ahead arrival and departure delay prediction in large-scale airport networks. To validate the effectiveness of our model, we experiment with two real-world delay datasets, including U.S and China flight delays; and we show that STPN outperforms state-of-the-art methods. In addition, counterfactuals produced by STPN show that it learns explainable delay propagation patterns.Comment: 14 pages,8 figure

    Using Machine Learning to Predict Damaging Straight-line Convective Winds

    Get PDF
    Thunderstorms, including straight-line (non-tornadic) winds, cause an average of over 100 deaths and $10 billion of insured damage per year in the United States. In the past decade machine learning has led to significant improvements in the prediction of other convective hazards, such as tornadoes, hail, lightning, and convectively induced aircraft turbulence. However, very few studies have used machine learning specifically to predict damaging straight-line winds. We have developed machine-learning models to predict the probability of damaging straight-line wind, defined as a gust ≥ 50 kt (25.72 m s-1), for a given storm cell. Predictions are made for three buffer distances around the storm cell (0, 5, and 10 km) and five lead-time windows ([0, 15]; [15, 30]; [30, 45]; [45, 60]; and [60, 90] minutes). Three types of data are used to train models: radar images from the Multi-year Reanalysis of Remotely Sensed Storms (MYRORSS); atmospheric soundings from the Rapid Update Cycle (RUC) model and North American Regional Reanalysis (NARR); and near-surface wind observations from the Meteorological Assimilation Data Ingest System (MADIS), Oklahoma Mesonet, one-minute meteorological aerodrome reports (METARs), and National Weather Service local storm reports. Radar images are used to determine the structural and hydrometeorological properties of storm cells, while soundings are used to determine properties of the near-storm environment, which are important for storm evolution. Both of these data types are used to create predictor variables. Meanwhile, near-surface wind observations are used as verification data (to determine which storm cells produced damaging straight-line winds). For each buffer distance and lead-time window, we experiment with five machine-learning algorithms: logistic regression, logistic regression with an elastic net, feed-forward neural nets, random forests, and gradient-boosted tree (GBT) ensembles. Forecast probabilities from each model are calibrated with isotonic regression, which makes them more reliable. Forecasts are verified mainly with three numbers: area under the receiver-operating-characteristic curve (AUC), maximum critical success index (CSI), and Brier skill score (BSS). AUC and maximum CSI range from [0, 1], where 0 is the worst score and 1 is a perfect score. BSS ranges from (−∞, 1], where −∞ is the worst score; 1 is a perfect score; and > 0 means that the model is better than climatology. Models are ranked by AUC. The best model (for a buffer distance of 0 km and lead time of [15, 30] minutes) has an AUC of 0.996, maximum CSI of 0.99, and BSS of 0.88. The worst model (for a buffer distance of 10 km and lead time of [60, 90] minutes) has an AUC of 0.89, maximum CSI of 0.20, and BSS of 0.12. All models outperform climatology. Finally, for each buffer distance and lead-time window, we use three methods to select the most important predictor variables: sequential forward selection, J-measures, and decision trees

    Storm-scale Ensemble-based Severe Weather Guidance: Development of an Object-based Verification Framework and Applications of Machine Learning

    Get PDF
    A goal of the National Oceanic and Atmospheric Administration (NOAA) Warn-on-Forecast (WoF) project is to provide rapidly updating probabilistic guidance to human forecasters for short-term (e.g., 0-3 h) severe weather forecasts. Several case studies have shown that experimental WoF systems (WoFS) can produce accurate short-term probabilistic guidance for hazards such as tornadoes, hail, and heavy rainfall. However, without an appropriate probabilistic verification method for WoFS-style forecasts (which provide guidance for individual thunderstorms), a robust evaluation of WoFS performance has been lacking. In this dissertation, I develop a novel object-based verification method for short-term, storm-scale probabilistic forecasts and apply it to WoFS probabilistic mesocyclone guidance and further adapted to evaluate machine learning-based calibrations of WoFS severe weather probabilistic guidance. The probabilistic mesocyclone guidance was generated by calculating grid-scale ensemble probabilities from WoFS forecasts of updraft helicity (UH) in layers 2—5 km (mid-level) and 0—2 km (low-level) above ground level (AGL) aggregated over 60-min periods. The resulting ensemble probability swaths are associated with individual thunderstorms and treated as objects. Each ensemble track object is assigned a single representative probability value. A mesocyclone probability object, conceptually, is a region bounded by the ensemble forecast envelope of a mesocyclone track for a thunderstorm over 1 hour. The mesocyclone probability objects were matched against rotation track objects in Multi-Radar Multi-Sensor data using the total interest score, but with the maximum displacement varied between 0, 9, 15, and 30 km. Forecast accuracy and reliability were assessed at four different forecast lead time periods: 0-60 min, 30-90 min, 60-120 min, and 90-150 min. In the 0-60 minute forecast period, the low-level UH probabilistic forecasts had a POD, FAR, and CSI of 0.46, 0.45, and 0.31, respectively, with a probability threshold of 22.2% (the threshold of maximum CSI). In the 90-150 minute forecast period, the POD and CSI dropped to 0.39 and 0.27 while FAR remained relatively unchanged. Forecast probabilities >60% over-predicted the likelihood of observed mesocyclones in the 0-60 min period; however, reliability improved when allowing larger maximum displacements for object matching and at longer lead times. To evaluate the ability of machine learning (ML) models to calibrate WoFS severe weather guidance, the probability object-based method was generalized for identifying any ensemble storm track (based on individual ensemble updraft tracks rather than mesocyclone tracks). Using these ensemble storm tracks, three sets of predictors were extracted from the WoFS forecasts: intra-storm state variables, near-storm environment variables, and morphological attributes of the ensemble storm tracks. Random forests, gradient-boosted trees, and logistic regression algorithms were then trained to predict which WoFS 30-min ensemble storm tracks will produce a tornado, severe hail, and/or severe wind report. To provide a baseline against which to test the ML models’ performance, I extracted the probability of mid-level UH exceeding a threshold (tuned per severe weather hazard) from each ensemble storm track. The three ML algorithms discriminated well for all three hazards and produced far more reliable probabilities than the UH-based predictions. Using state-of-the-art ML interpretability methods, I found that the ML models learned sound physical relationships and the appropriate responses to the ensemble statistics. Intra-storm predictors were found to be more important than environmental predictors for all three ML models, but environmental predictors made positive contributions to severe weather likelihood in situations where the WoFS fails to analyze ongoing convection. Overall, the results suggest that ML-based calibrations of dynamical ensemble output can improve short term, storm-scale severe weather probabilistic guidance

    Analysis, Characterization, Prediction and Attribution of Extreme Atmospheric Events with Machine Learning: a Review

    Full text link
    Atmospheric Extreme Events (EEs) cause severe damages to human societies and ecosystems. The frequency and intensity of EEs and other associated events are increasing in the current climate change and global warming risk. The accurate prediction, characterization, and attribution of atmospheric EEs is therefore a key research field, in which many groups are currently working by applying different methodologies and computational tools. Machine Learning (ML) methods have arisen in the last years as powerful techniques to tackle many of the problems related to atmospheric EEs. This paper reviews the ML algorithms applied to the analysis, characterization, prediction, and attribution of the most important atmospheric EEs. A summary of the most used ML techniques in this area, and a comprehensive critical review of literature related to ML in EEs, are provided. A number of examples is discussed and perspectives and outlooks on the field are drawn.Comment: 93 pages, 18 figures, under revie

    Coupling Data Science Techniques and Numerical Weather Prediction Models for High-Impact Weather Prediction

    Get PDF
    Meteorologists have access to more model guidance and observations than ever before, but this additional information does not necessarily lead to better forecasts. New tools are needed to reduce the cognitive load on forecasters and to provide them with accurate, reliable consensus guidance. Techniques from the data science community, such as machine learning and image processing, have the potential to summarize and calibrate numerical weather prediction model output and to generate deterministic and probabilistic forecasts of high-impact weather. In this dissertation, I developed data-science-based approaches to improve the predictions of two high-impact weather domains: hail and solar irradiance. Both hail and solar irradiance produce large economic impacts, have non-Gaussian distributions of occurrence, are poorly observed, and are partially driven by processes too small to be resolved by numerical weather prediction models. Hail forecasts were produced with convection-allowing model output from the Center for Analysis and Prediction of Storms and National Center for Atmospheric Research ensembles. The machine learning hail forecasts were compared against storm surrogate variables and physics-based diagnostic models of hail size. Initial machine learning hail forecasts reduced size errors but struggled with predicting extreme events. By coupling the machine learning model to predicting hail size distributions and estimating the distribution parameters jointly, the machine learning methods were able to show skill and reliability in predicting both severe and significant hail. Machine learning model and data configurations for gridded solar irradiance forecasting were evaluated on two numerical modeling systems. The evaluation determined how machine learning model choice, closeness of fit to training data, training data aggregation, and interpolation method affected forecasts of clearness index at Oklahoma Mesonet sites not included in the training data. The choice of machine learning model, interpolation scheme, and loss function had the biggest impacts on performance. Errors tended to be lower at testing sites with sunnier weather and those that were closer to training sites. All of the machine learning methods produced reliable predictions but underestimated the frequency of cloudiness compared to observations

    Low-latency, query-driven analytics over voluminous multidimensional, spatiotemporal datasets

    Get PDF
    2017 Summer.Includes bibliographical references.Ubiquitous data collection from sources such as remote sensing equipment, networked observational devices, location-based services, and sales tracking has led to the accumulation of voluminous datasets; IDC projects that by 2020 we will generate 40 zettabytes of data per year, while Gartner and ABI estimate 20-35 billion new devices will be connected to the Internet in the same time frame. The storage and processing requirements of these datasets far exceed the capabilities of modern computing hardware, which has led to the development of distributed storage frameworks that can scale out by assimilating more computing resources as necessary. While challenging in its own right, storing and managing voluminous datasets is only the precursor to a broader field of study: extracting knowledge, insights, and relationships from the underlying datasets. The basic building block of this knowledge discovery process is analytic queries, encompassing both query instrumentation and evaluation. This dissertation is centered around query-driven exploratory and predictive analytics over voluminous, multidimensional datasets. Both of these types of analysis represent a higher-level abstraction over classical query models; rather than indexing every discrete value for subsequent retrieval, our framework autonomously learns the relationships and interactions between dimensions in the dataset (including time series and geospatial aspects), and makes the information readily available to users. This functionality includes statistical synopses, correlation analysis, hypothesis testing, probabilistic structures, and predictive models that not only enable the discovery of nuanced relationships between dimensions, but also allow future events and trends to be predicted. This requires specialized data structures and partitioning algorithms, along with adaptive reductions in the search space and management of the inherent trade-off between timeliness and accuracy. The algorithms presented in this dissertation were evaluated empirically on real-world geospatial time-series datasets in a production environment, and are broadly applicable across other storage frameworks

    Traffic Prediction using Artificial Intelligence: Review of Recent Advances and Emerging Opportunities

    Full text link
    Traffic prediction plays a crucial role in alleviating traffic congestion which represents a critical problem globally, resulting in negative consequences such as lost hours of additional travel time and increased fuel consumption. Integrating emerging technologies into transportation systems provides opportunities for improving traffic prediction significantly and brings about new research problems. In order to lay the foundation for understanding the open research challenges in traffic prediction, this survey aims to provide a comprehensive overview of traffic prediction methodologies. Specifically, we focus on the recent advances and emerging research opportunities in Artificial Intelligence (AI)-based traffic prediction methods, due to their recent success and potential in traffic prediction, with an emphasis on multivariate traffic time series modeling. We first provide a list and explanation of the various data types and resources used in the literature. Next, the essential data preprocessing methods within the traffic prediction context are categorized, and the prediction methods and applications are subsequently summarized. Lastly, we present primary research challenges in traffic prediction and discuss some directions for future research.Comment: Published in Transportation Research Part C: Emerging Technologies (TR_C), Volume 145, 202

    Effects of environmental shear and buoyancy on simulated supercell interactions

    Get PDF
    This study examines the role that buoyancy and vertical wind shear play in modulating the relationship between storm interactions and storm severity. Using an idealized numerical model, 240 supercell interactions are simulated under systematically varied amounts of buoyancy and vertical wind shear. Small changes in buoyancy or vertical wind shear have signi cant impacts on post-interaction storm morphology. A wide amount of variation in low-level rotation is seen across the simulation suite. Two-cell storm simulations are not always stronger than one-cell control simulations. Migration of low-level vertical vorticity centers is ubiquitous through all runs, but orientation of and interaction between two storms' gust fronts modulates where the vortic- ity center will end up. Gust fronts in better alignment have more vorticity centers reach an updraft where they are stretched and intensi ed. With re- spect to storm mode, higher buoyancy produced less classic supercells while higher shear produced more classic supercells. High precipitation supercells were favored with two-cell simulations where the second cell was directly to the southwest of the control cell. Secondary cells that were close to the con- trol merged quickly and were often stronger than simulations with large cell separation distance. Further questions remain with trajectories and machine learning algorithms are tthe next steps for a more detailed analysis of this large data set

    WAZE Data Reporting

    Get PDF
    This study evaluated the quality of crowdsourced Waze data (including reports and speed) and explored promising use scenarios of Waze data to facilitate the development of intelligent transportation in Tennessee. To this end, the thoroughly assessed Waze reports quality in terms of spatiotemporal accuracy and coverage. The study found Waze users reported crash events about 2.2 minutes sooner, on average, than reports of the same events recorded in the state\u2019s Locate/IM incident log. The reported crash locations per Waze are on average 6 feet from the Locate/IM log reported by the officials. It is found that 26% of crashes reported in Waze was matched with 67% Locate/IM crash reports, with the rest 74% reports pointing to unreported incidents. Waze speed is affected by the Wazers behaviors and tends to be slightly higher than detector speed in free-flow status. This study evaluated several novel use scenarios such as secondary crash detection, end of queue detection and tracking, level of service evaluation, work zone monitoring, wildlife hazards and crashes, and pothole detection and maintenance. Results show that Waze is a suitable data source for incident management, level of service evaluation, work zone management, roadway maintenance management, etc. when properly used and in cooperation with the agency\u2019s other information sources
    • …
    corecore