131 research outputs found
A computer vision system for detecting and analysing critical events in cities
Whether for commuting or leisure, cycling is a growing transport mode in many cities worldwide. However, it is still perceived as a dangerous activity. Although serious incidents related to cycling leading to major injuries are rare, the fear of getting hit or falling hinders the expansion of cycling as a major transport mode. Indeed, it has been shown that focusing on serious injuries only touches the tip of the iceberg. Near miss data can provide much more information about potential problems and how to avoid risky situations that may lead to serious incidents. Unfortunately, there is a gap in the knowledge in identifying and analysing near misses. This hinders drawing statistically significant conclusions to provide measures for the built-environment that ensure a safer environment for people on bikes. In this research, we develop a method to detect and analyse near misses and their risk factors using artificial intelligence. This is accomplished by analysing video streams linked to near miss incidents within a novel framework relying on deep learning and computer vision. This framework automatically detects near misses and extracts their risk factors from video streams before analysing their statistical significance. It also provides practical solutions implemented in a camera with embedded AI (URBAN-i Box) and a cloud-based service (URBAN-i Cloud) to tackle the stated issue in the real-world settings for use by researchers, policy-makers, or citizens. The research aims to provide human-centred evidence that may enable policy-makers and planners to provide a safer built environment for cycling in London, or elsewhere. More broadly, this research aims to contribute to the scientific literature with the theoretical and empirical foundations of a computer vision system that can be utilised for detecting and analysing other critical events in a complex environment. Such a system can be applied to a wide range of events, such as traffic incidents, crime or overcrowding
Recommended from our members
Estimation and Sparse Selection of Conditional Probability Models for Vector Time Series
Diverse scientific fields collect multiple time series data to investigate the dynamical behavior of complex systems: atmospheric and climate science, geophysics, neuroscience, epidemiology, ecology, and environmental science. Identifying patterns of mutual dependence among such data generates valuable knowledge that can be applied either for inferential or forecasting purposes. Vector autoregressive (VAR) processes provide a flexible class of statistical models for multiple time series that are easy to estimate using regression techniques. However, scaling to large data sets and extension to more general processes stretch the framework's capacity: due to the dense parametrization of VAR models, which have one parameter for every possible pairwise relationship between components (i.e., between each univariate time series in a collection), high-dimensional data generate difficulties associated with model selection and parametric regularization; and modeling more general processes requires data transformations that complicate inference, forecasting, and model interpretation. Fields such as epidemiology, neuroscience, and ecology generate high-dimensional time series of count vectors, which incur both sets of challenges at once. Autoregressive conditional probability models --- models in which the conditional means of a time series follow an autoregressive structure in the process history --- are natural generalizations that preserve ease of estimation and, in conjunction with selection methods in regression, promise to address challenges associated with modeling large multiple time series of count (and other discrete) data. This thesis focuses on developing empirical methodology for sparse selection of nonlinear VAR-type conditional Poisson models. Chapter 1 provides an overview of related existing work. Chapter 2 develops an empirical method for sparse selection in VAR models based on resampling methods. Chapter 3 presents a conditional probability generalization of the VAR model and analyzes the stability properties of Poisson generalized vector autoregressive (GVAR) processes. Chapter 4 combines the work of the preceding two chapters and develops a resampling-based method for sparse estimation of Poisson GVAR models. Finally, Chapter 5 summarizes key findings, challenges, and future work
Theory and Application of Dynamic Spatial Time Series Models
Stochastic economic processes are often characterized by dynamic interactions between variables that are dependent in both space and time. Analyzing these processes raises a number of questions about the econometric methods used that are both practically and theoretically interesting. This work studies econometric approaches to analyze spatial data that evolves dynamically over time. The book provides a background on least squares and maximum likelihood estimators, and discusses some of the limits of basic econometric theory. It then discusses the importance of addressing spatial heterogeneity in policies. The next chapters cover parametric modeling of linear and nonlinear spatial time series, non-parametric modeling of nonlinearities in panel data, modeling of multiple spatial time series variables that exhibit long and short memory, and probabilistic causality in spatial time series settings
MODELING THE LEADERSHIP OF LANGUAGE CHANGE FROM DIACHRONIC TEXT
Natural languages constantly change over time. These changes are modulated by social factors such as influence which are not always directly observable. However, large-scale computational modeling of language change using timestamped text can uncover the latent organization and social structure. In turn, the social dynamics of language change can potentially illuminate our understanding of innovation, influence, and identity: Who leads? Who follows? Who diverges?
This thesis contributes to the growing body of research on using computational methods to model language change with a focus on quantifying linguistic leadership of change. A series of studies highlight the unique contributions of this thesis: methods that scale to huge volumes of data; measures that quantify leadership at the level of individuals or in aggregate; and analysis that links linguistic leadership to other forms of influence. First, temporal and predictive models of event cascades on a network of millions of Twitter users are used to show that lexical change spreads in the form of a contagion and influence from densely embedded ties is crucial for the adoption of non-standard terms. A Granger-causal test for detecting social influence in event cascades on a network is then presented, which is robust to both the presence of confounds such as homophily and can be applied to model both linguistic or non-linguistic change in a network. Next, a novel scheme to score and identify documents that lead semantic change in progress is introduced. This linguistic measure of influence on the documents is strongly predictive of their influence in terms of the number of citations that they receive for both US court opinions and scientific articles. Subsequently, a measure of lead on any semantic change between a pair of document sources (e.g. newspapers) and a method to aggregate multiple lead-lag relationships into a network is presented. Analysis on an induced network of nineteenth century abolitionist newspapers, following the proposed method, reveals the important yet understated role of women and Black editors in shaping the discourse on abolitionism. Finally, a method to induce an aggregate semantic leadership network using contextual word representations is proposed to investigate the link between semantic leadership and influence in the form of citations among publication venues that are part of the Association of Computational Linguistics. Taken together, these studies illustrate the utility of finding leaders of language change to gain insights in sociolinguistics and for applications in social science and digital humanities.Ph.D
Analyzing Granger causality in climate data with time series classification methods
Attribution studies in climate science aim for scientifically ascertaining the influence of climatic variations on natural or anthropogenic factors. Many of those studies adopt the concept of Granger causality to infer statistical cause-effect relationships, while utilizing traditional autoregressive models. In this article, we investigate the potential of state-of-the-art time series classification techniques to enhance causal inference in climate science. We conduct a comparative experimental study of different types of algorithms on a large test suite that comprises a unique collection of datasets from the area of climate-vegetation dynamics. The results indicate that specialized time series classification methods are able to improve existing inference procedures. Substantial differences are observed among the methods that were tested
A New Framework for Decomposing Multivariate Information
What are the distinct ways in which a set of predictor variables can provide information about a target variable? When does a variable provide unique information, when do variables share redundant information, and when do variables combine synergistically to provide complementary information? The redundancy lattice from the partial information decomposition of Williams and Beer provided a promising glimpse at the answer to these questions. However, this structure was constructed using a much-criticised measure of redundant information, and despite sustained research, no completely satisfactory replacement measure has been proposed. This thesis presents a new framework for information decomposition that is based upon the decomposition of pointwise mutual information rather than mutual information. The framework is derived in two separate ways. The first of these derivations is based upon a modified version of the original axiomatic approach taken by Williams and Beer. However, to overcome the difficulty associated with signed pointwise mutual information, the decomposition is applied separately to the unsigned entropic components of pointwise mutual information which are referred to as the specificity and ambiguity. This yields a separate redundancy lattice for each component. Based upon an operational interpretation of redundancy, measures of redundant specificity and redundant ambiguity are defined which enables one to evaluate the partial information atoms separately for each lattice. These separate atoms can then be recombined to yield the sought-after multivariate information decomposition. This framework is applied to canonical examples from the literature and the results and various properties of the decomposition are discussed. In particular, the pointwise decomposition using specificity and ambiguity is shown to satisfy a chain rule over target variables, which provides new insights into the so-called two-bit-copy example. The second approach begins by considering the distinct ways in which two marginal observers can share their information with the non-observing individual third party. Several novel measures of information content are introduced, namely the union, intersection and unique information contents. Next, the algebraic structure of these new measures of shared marginal information is explored, and it is shown that the structure of shared marginal information is that of a distributive lattice. Furthermore, by using the fundamental theorem of distributive lattices, it is shown that these new measures are isomorphic to a ring of sets. Finally, by combining this structure together with the semi-lattice of joint information, the redundancy lattice form partial information decomposition is found to be embedded within this larger algebraic structure. However, since this structure considers information contents, it is actually equivalent to the specificity lattice from the first derivation of pointwise partial information decomposition. The thesis then closes with a discussion about whether or not one should combine the information contents from the specificity and ambiguity lattices
Recommended from our members
Bayesian dynamic graphical models for high-dimensional flow forecasting in road traffic networks
Congestion on roads is a crucial problem which affects our lives in many ways: As a consequence, there is a strong effort to improve road networks in order to keep the traffic flowing. Flow forecasting models based on the large amount of traffic data, which are now available, can be a very useful tool to support decisions and actions when managing traffic networks. Although many forecasting models have been developed to this end, very few of them capture important features of high dimensional traffic data and, moreover, operating most of these models is a hard task when considering on-line traffic management environments.
Dynamic graphical models can be a suitable choice to address the challenge of forecasting high-dimensional traffic flows in real-time. These models represent network flows by a graph, which not only is a useful pictorial representation of multivariate time series of traffic flow data, but it also ensures that model computation is always simple, even for very complex road networks. One example of such a model is the multiregression dynamic model (MDM).
This thesis focuses on the development of two classes of dynamic graphical models to forecast traffic flows . Firstly, the linear multiregression dynamic model (LMDM), which is an MDM particular case, is extended to allow important traffic characteristics in its structure, such as the heterocedasticity of daily traffic flows, measurement errors due to malfunctions in data collection devices, and the use of extra traffic variables as predictors to forecast flows. Due to its graphical structure, the MDM assumes independence of flows at the entrances of a road network. This thesis therefore introduces a new class of dynamic graphical models where the correlation across road network entrances is accommodated, resulting in better forecasts when compared to the LMDM.
All the methodology proposed in this thesis is illustrated using data collected at the intersection of three busy motorways near Manchester, UK
Feature Papers of Forecasting
Nowadays, forecast applications are receiving unprecedent attention thanks to their capability to improve the decision-making processes by providing useful indications. A large number of forecast approaches related to different forecast horizons and to the specific problem that have to be predicted have been proposed in recent scientific literature, from physical models to data-driven statistic and machine learning approaches. In this Special Issue, the most recent and high-quality researches about forecast are collected. A total of nine papers have been selected to represent a wide range of applications, from weather and environmental predictions to economic and management forecasts. Finally, some applications related to the forecasting of the different phases of COVID in Spain and the photovoltaic power production have been presented
- …