1,783 research outputs found

    Data-driven action-value functions for evaluating players in professional team sports

    Get PDF
    As more and larger event stream datasets for professional sports become available, there is growing interest in modeling the complex play dynamics to evaluate player performance. Among these models, a common player evaluation method is assigning values to player actions. Traditional action-values metrics, however, consider very limited game context and player information. Furthermore, they provide directly related to goals (e.g., shots), not all actions. Recent work has shown that reinforcement learning provided powerful methods for addressing quantifying the value of player actions in sports. This dissertation develops deep reinforcement learning (DRL) methods for estimating action values in sports. We make several contributions to DRL for sports. First, we develop neural network architectures that learn an action-value Q-function from sports events logs to estimate each team\u27s expected success given the current match context. Specifically, our architecture models the game history with a recurrent network and predicts the probability that a team scores the next goal. From the learned Q-values, we derive a Goal Impact Metric (GIM) for evaluating a player\u27s performance over a game season. We show that the resulting player rankings are consistent with standard player metrics and temporally consistent within and across seasons. Second, we address the interpretability of the learned Q-values. While neural networks provided accurate estimates, the black-box structure prohibits understanding the influence of different game features on the action values. To interpret the Q-function and understand the influence of game features on action values, we design an interpretable mimic learning framework for the DRL. The framework is based on a Linear Model U-Tree (LMUT) as a transparent mimic model, which facilitates extracting the function rules and computing the feature importance for action values. Third, we incorporate information about specific players into the action values, by introducing a deep player representation framework. In this framework, each player is assigned a latent feature vector called an embedding, with the property that statistically similar players are mapped to nearby embeddings. To compute embeddings that summarize the statistical information about players, we implement a Variational Recurrent Ladder Agent Encoder (VaRLAE) to learn a contextualized representation for when and how players are likely to act. We learn and evaluate deep Q-functions from event data for both ice hockey and soccer. These are challenging continuous-flow games where game context and medium-term consequences are crucial for properly assessing the impact of a player\u27s actions

    Use of the Electronic Nose as a Screening Tool for the Recognition of Durum Wheat Naturally Contaminated by Deoxynivalenol: A Preliminary Approach

    Get PDF
    Fungal contamination and the presence of related toxins is a widespread problem. Mycotoxin contamination has prompted many countries to establish appropriate tolerance levels. For instance, with the Commission Regulation (EC) N. 1881/2006, the European Commission fixed the limits for the main mycotoxins (and other contaminants) in food. Although valid analytical methods are being developed for regulatory purposes, a need exists for alternative screening methods that can detect mould and mycotoxin contamination of cereal grains with high sample throughput. In this study, a commercial electronic nose (EN) equipped with metal-oxide-semiconductor (MOS) sensors was used in combination with a trap and the thermal desorption technique, with the adoption of Tenax TA as an adsorbent material to discriminate between durum wheat whole-grain samples naturally contaminated with deoxynivalenol (DON) and non-contaminated samples. Each wheat sample was analysed with the EN at four different desorption temperatures (i.e., 180 °C, 200 °C, 220 °C, and 240 °C) and without a desorption pre-treatment. A 20-sample and a 122-sample dataset were processed by means of principal component analysis (PCA) and classified via classification and regression trees (CART). Results, validated with two different methods, showed that it was possible to classify wheat samples into three clusters based on the DON content proposed by the European legislation: (a) non-contaminated; (b) contaminated below the limit (DON < 1,750 μg/kg); (c) contaminated above the limit (DON > 1,750 μg/kg), with a classification error rate in prediction of 0% (for the 20-sample dataset) and 3.28% (for the 122-sample dataset)

    New Framework and Decision Support Tool to Warrant Detour Operations During Freeway Corridor Incident Management

    Get PDF
    As reported in the literature, the mobility and reliability of the highway systems in the United States have been significantly undermined by traffic delays on freeway corridors due to non-recurrent traffic congestion. Many of those delays are caused by the reduced capacity and overwhelming demand on critical metropolitan corridors coupled with long incident durations. In most scenarios, if proper detour strategies could be implemented in time, motorists could circumvent the congested segments by detouring through parallel arterials, which will significantly improve the mobility of all vehicles in the corridor system. Nevertheless, prior to implementation of any detour strategy, traffic managers need a set of well-justified warrants, as implementing detour operations usually demand substantial amount of resources and manpower. To contend with the aforementioned issues, this study is focused on developing a new multi-criteria framework along with an advanced and computation-friendly tool for traffic managers to decide whether or not and when to implement corridor detour operations. The expected contributions of this study are: * Proposing a well-calibrated corridor simulation network and a comprehensive set of experimental scenarios to take into account many potential affecting factors on traffic manager\u27s decision making process and ensure the effectiveness of the proposed detour warrant tool; * Developing detour decision models, including a two-choice model and a multi-choice model, based on generated optima detour traffic flow rates for each scenario from a diversion control model to allow responsible traffic managers to make best detour decisions during real-time incident management; and * Estimating the resulting benefits for comparison with the operational costs using the output from the diversion control model to further validate the developed detour decision model from the overall societal perspective

    Application of Artificial Intelligence Approaches in the Flood Management Process for Assessing Blockage at Cross-Drainage Hydraulic Structures

    Get PDF
    Floods are the most recurrent, widespread and damaging natural disasters, and are ex-pected to become further devastating because of global warming. Blockage of cross-drainage hydraulic structures (e.g., culverts, bridges) by flood-borne debris is an influen-tial factor which usually results in reducing hydraulic capacity, diverting the flows, dam-aging structures and downstream scouring. Australia is among the countries adversely impacted by blockage issues (e.g., 1998 floods in Wollongong, 2007 floods in Newcas-tle). In this context, Wollongong City Council (WCC), under the Australian Rainfall and Runoff (ARR), investigated the impact of blockage on floods and proposed guidelines to consider blockage in the design process for the first time. However, existing WCC guide-lines are based on various assumptions (i.e., visual inspections as representative of hy-draulic behaviour, post-flood blockage as representative of peak floods, blockage remains constant during the whole flooding event), that are not supported by scientific research while also being criticised by hydraulic design engineers. This suggests the need to per-form detailed investigations of blockage from both visual and hydraulic perspectives, in order to develop quantifiable relationships and incorporate blockage into design guide-lines of hydraulic structures. However, because of the complex nature of blockage as a process and the lack of blockage-related data from actual floods, conventional numerical modelling-based approaches have not achieved much success. The research in this thesis applies artificial intelligence (AI) approaches to assess the blockage at cross-drainage hydraulic structures, motivated by recent success achieved by AI in addressing complex real-world problems (e.g., scour depth estimation and flood inundation monitoring). The research has been carried out in three phases: (a) litera-ture review, (b) hydraulic blockage assessment, and (c) visual blockage assessment. The first phase investigates the use of computer vision in the flood management domain and provides context for blockage. The second phase investigates hydraulic blockage using lab scale experiments and the implementation of multiple machine learning approaches on datasets collected from lab experiments (i.e., Hydraulics-Lab Dataset (HD), Visual Hydraulics-Lab Dataset (VHD)). The artificial neural network (ANN) and end-to-end deep learning approaches reported top performers among the implemented approaches and demonstrated the potential of learning-based approaches in addressing blockage is-sues. The third phase assesses visual blockage at culverts using deep learning classifi-cation, detection and segmentation approaches for two types of visual assessments (i.e., blockage status classification, percentage visual blockage estimation). Firstly, a range of existing convolutional neural network (CNN) image classification models are imple-mented and compared using visual datasets (i.e., Images of Culvert Openings and Block-age (ICOB), VHD, Synthetic Images of Culverts (SIC)), with the aim to automate the process of manual visual blockage classification of culverts. The Neural Architecture Search Network (NASNet) model achieved best classification results among those im-plemented. Furthermore, the study identified background noise and simplified labelling criteria as two contributing factors in degraded performance of existing CNN models for blockage classification. To address the background clutter issue, a detection-classification pipeline is proposed and achieved improved visual blockage classification performance. The proposed pipeline has been deployed using edge computing hardware for blockage monitoring of actual culverts. The role of synthetic data (i.e., SIC) on the performance of culvert opening detection is also investigated. Secondly, an automated segmentation-classification deep learning pipeline is proposed to estimate the percentage of visual blockage at circular culverts to better prioritise culvert maintenance. The AI solutions proposed in this thesis are integrated into a blockage assessment framework, designed to be deployed through edge computing to monitor, record and assess blockage at cross-drainage hydraulic structures

    Prediction of Housing Price and Forest Cover Using Mosaics with Uncertain Satellite Imagery

    Get PDF
    The growing world is more expensive to estimate land use, road length, and forest cover using a plant-scaled ground monitoring system. Satellite imaging contains a significant amount of detailed uncertain information. Combining this with machine learning aids in the organization of these data and the estimation of each variable separately. The resources necessary to deploy Machine learning technologies for Remote sensing images, on the other hand, restrict their reach ability and application. Based on satellite observations which are notably underutilised in impoverished nations, while practical competence to implement SIML might be restricted. Encoded forms of images are shared across tasks, and they will be calculated and sent to an infinite number of researchers who can achieve top-tier SIML performance by training a regression analysis onto the actual data. By separating the duties, the proposed SIML solution, MOSAIKS, shapes SIML approachable and global. A Featurization stage turns remote sensing data into concise vector representations, and a regression step makes it possible to learn the correlations which are specific to its particular task which link the obtained characteristics to the set of uncertain data

    The Emerging Trends of Multi-Label Learning

    Full text link
    Exabytes of data are generated daily by humans, leading to the growing need for new efforts in dealing with the grand challenges for multi-label learning brought by big data. For example, extreme multi-label classification is an active and rapidly growing research area that deals with classification tasks with an extremely large number of classes or labels; utilizing massive data with limited supervision to build a multi-label classification model becomes valuable for practical applications, etc. Besides these, there are tremendous efforts on how to harvest the strong learning capability of deep learning to better capture the label dependencies in multi-label learning, which is the key for deep learning to address real-world classification tasks. However, it is noted that there has been a lack of systemic studies that focus explicitly on analyzing the emerging trends and new challenges of multi-label learning in the era of big data. It is imperative to call for a comprehensive survey to fulfill this mission and delineate future research directions and new applications.Comment: Accepted to TPAMI 202

    Data Driven Approaches for Image & Video Understanding: from Traditional to Zero-shot Supervised Learning

    Get PDF
    In the present age of advanced computer vision, the necessity of (user-annotated) data is a key factor in image & video understanding. Recent success of deep learning on large scale data has only acted as a catalyst. There are certain problems that exist in this regard: 1) scarcity of (annotated) data, 2) need of expensive manual annotation, 3) problem of change in domain, 4) knowledge base not exhaustive. To make efficient learning systems, one has to be prepared to deal with such diverse set of problems. In terms of data availability, extensive manual annotation can be beneficial in obtaining category specific knowledge. Even then, learning efficient representation for the related task is challenging and requires special attention. On the other hand, when labelled data is scarce, learning category specific representation itself becomes challenging. In this work, I investigate data driven approaches that cater to traditional supervised learning setup as well as an extreme case of data scarcity where no data from test classes are available during training, known as zero-shot learning. First, I look into supervised learning setup with ample annotations and propose efficient dictionary learning technique for better learning of data representation for the task of action classification in images & videos. Then I propose robust mid-level feature representations for action videos that are equally effective in traditional supervised learning as well as zero-shot learning. Finally, I come up with novel approach that cater to zero-shot learning specifically. Thorough discussions followed by experimental validations establish the worth of these novel techniques in solving computer vision related tasks under varying data-dependent scenarios

    ImageNet Large Scale Visual Recognition Challenge

    Get PDF
    The ImageNet Large Scale Visual Recognition Challenge is a benchmark in object category classification and detection on hundreds of object categories and millions of images. The challenge has been run annually from 2010 to present, attracting participation from more than fifty institutions. This paper describes the creation of this benchmark dataset and the advances in object recognition that have been possible as a result. We discuss the challenges of collecting large-scale ground truth annotation, highlight key breakthroughs in categorical object recognition, provide a detailed analysis of the current state of the field of large-scale image classification and object detection, and compare the state-of-the-art computer vision accuracy with human accuracy. We conclude with lessons learned in the five years of the challenge, and propose future directions and improvements.Comment: 43 pages, 16 figures. v3 includes additional comparisons with PASCAL VOC (per-category comparisons in Table 3, distribution of localization difficulty in Fig 16), a list of queries used for obtaining object detection images (Appendix C), and some additional reference

    Approximation and Relaxation Approaches for Parallel and Distributed Machine Learning

    Get PDF
    Large scale machine learning requires tradeoffs. Commonly this tradeoff has led practitioners to choose simpler, less powerful models, e.g. linear models, in order to process more training examples in a limited time. In this work, we introduce parallelism to the training of non-linear models by leveraging a different tradeoff--approximation. We demonstrate various techniques by which non-linear models can be made amenable to larger data sets and significantly more training parallelism by strategically introducing approximation in certain optimization steps. For gradient boosted regression tree ensembles, we replace precise selection of tree splits with a coarse-grained, approximate split selection, yielding both faster sequential training and a significant increase in parallelism, in the distributed setting in particular. For metric learning with nearest neighbor classification, rather than explicitly train a neighborhood structure we leverage the implicit neighborhood structure induced by task-specific random forest classifiers, yielding a highly parallel method for metric learning. For support vector machines, we follow existing work to learn a reduced basis set with extremely high parallelism, particularly on GPUs, via existing linear algebra libraries. We believe these optimization tradeoffs are widely applicable wherever machine learning is put in practice in large scale settings. By carefully introducing approximation, we also introduce significantly higher parallelism and consequently can process more training examples for more iterations than competing exact methods. While seemingly learning the model with less precision, this tradeoff often yields noticeably higher accuracy under a restricted training time budget
    corecore