197 research outputs found

    A novel Big Data analytics and intelligent technique to predict driver's intent

    Get PDF
    Modern age offers a great potential for automatically predicting the driver's intent through the increasing miniaturization of computing technologies, rapid advancements in communication technologies and continuous connectivity of heterogeneous smart objects. Inside the cabin and engine of modern cars, dedicated computer systems need to possess the ability to exploit the wealth of information generated by heterogeneous data sources with different contextual and conceptual representations. Processing and utilizing this diverse and voluminous data, involves many challenges concerning the design of the computational technique used to perform this task. In this paper, we investigate the various data sources available in the car and the surrounding environment, which can be utilized as inputs in order to predict driver's intent and behavior. As part of investigating these potential data sources, we conducted experiments on e-calendars for a large number of employees, and have reviewed a number of available geo referencing systems. Through the results of a statistical analysis and by computing location recognition accuracy results, we explored in detail the potential utilization of calendar location data to detect the driver's intentions. In order to exploit the numerous diverse data inputs available in modern vehicles, we investigate the suitability of different Computational Intelligence (CI) techniques, and propose a novel fuzzy computational modelling methodology. Finally, we outline the impact of applying advanced CI and Big Data analytics techniques in modern vehicles on the driver and society in general, and discuss ethical and legal issues arising from the deployment of intelligent self-learning cars

    Exploiting the Statistics of Learning and Inference

    Full text link
    When dealing with datasets containing a billion instances or with simulations that require a supercomputer to execute, computational resources become part of the equation. We can improve the efficiency of learning and inference by exploiting their inherent statistical nature. We propose algorithms that exploit the redundancy of data relative to a model by subsampling data-cases for every update and reasoning about the uncertainty created in this process. In the context of learning we propose to test for the probability that a stochastically estimated gradient points more than 180 degrees in the wrong direction. In the context of MCMC sampling we use stochastic gradients to improve the efficiency of MCMC updates, and hypothesis tests based on adaptive mini-batches to decide whether to accept or reject a proposed parameter update. Finally, we argue that in the context of likelihood free MCMC one needs to store all the information revealed by all simulations, for instance in a Gaussian process. We conclude that Bayesian methods will remain to play a crucial role in the era of big data and big simulations, but only if we overcome a number of computational challenges.Comment: Proceedings of the NIPS workshop on "Probabilistic Models for Big Data

    Prediction Techniques in Internet of Things (IoT) Environment: A Comparative Study

    Get PDF
    Socialization and Personalization in Internet of Things (IOT) environment are the current trends in computing research. Most of the research work stresses the importance of predicting the service & providing socialized and personalized services. This paper presents a survey report on different techniques used for predicting user intention in wide variety of IOT based applications like smart mobile, smart television, web mining, weather forecasting, health-care/medical, robotics, road-traffic, educational data mining, natural calamities, retail banking, e-commerce, wireless networks & social networking. As per the survey made the prediction techniques are used for: predicting the application that can be accessed by the mobile user, predicting the next page to be accessed by web user, predicting the users favorite TV program, predicting user navigational patterns and usage needs on websites & also to extract the users browsing behavior, predicting future climate conditions, predicting whether a patient is suffering from a disease, predicting user intention to make implicit and human-like interactions possible by accepting implicit commands, predicting the amount of traffic occurring at a particular location, predicting student performance in schools & colleges, predicting & estimating the frequency of natural calamities occurrences like floods, earthquakes over a long period of time & also to take precautionary measures, predicting & detecting false user trying to make transaction in the name of genuine user, predicting the actions performed by the user to improve the business, predicting & detecting the intruder acting in the network, predicting the mood transition information of the user by using context history, etc. This paper also discusses different techniques like Decision Tree algorithm, Artificial Intelligence and Data Mining based Machine learning techniques, Content and Collaborative based Recommender algorithms used for prediction

    Reconstructing parton distribution functions from Ioffe time data: from Bayesian methods to Neural Networks

    Full text link
    The computation of the parton distribution functions (PDF) or distribution amplitudes (DA) of hadrons from first principles lattice QCD constitutes a central open problem. In this study, we present and evaluate the efficiency of a selection of methods for inverse problems to reconstruct the full xx-dependence of PDFs. Our starting point are the so called Ioffe time PDFs, which are accessible from Euclidean time calculations in conjunction with a matching procedure. Using realistic mock data tests, we find that the ill-posed incomplete Fourier transform underlying the reconstruction requires careful regularization, for which both the Bayesian approach as well as neural networks are efficient and flexible choices.Comment: 1+41 pages, 20 figure

    Can Generative Adversarial Networks Help Us Fight Financial Fraud?

    Get PDF
    Transactional fraud datasets exhibit extreme class imbalance. Learners cannot make accurate generalizations without sufficient data. Researchers can account for imbalance at the data level, algorithmic level or both. This paper focuses on techniques at the data level. We evaluate the evidence of the optimal technique and potential enhancements. Global fraud losses totalled more than 80 % of the UK’s GDP in 2019. The improvement of preprocessing is inherently valuable in fighting these losses. Synthetic minority oversampling technique (SMOTE) and extensions of SMOTE are currently the most common preprocessing strategies. SMOTE oversamples the minority classes by randomly generating a point between a minority instance and its nearest neighbour. Recent papers adopt generative adversarial networks (GAN) for data synthetic creation. Since 2014 there had been several GAN extensions, from improved training mechanisms to frameworks specifically for tabular data. The primary aim of the research is to understand the benefits of GANs built specifically for tabular data on supervised classifiers performance. We determine if this framework will outperform traditional methods and more common GAN frameworks. Secondly, we propose a framework that allows individuals to test the impact of imbalance ratios on classifier performance. Finally, we investigate the use of clustering and determine if this information can help GANs create better synthetic information. We explore this in the context of commonly used supervised classifiers and ensemble methods

    Piecewise Deterministic Markov Processes for Bayesian Neural Networks

    Full text link
    Inference on modern Bayesian Neural Networks (BNNs) often relies on a variational inference treatment, imposing violated assumptions of independence and the form of the posterior. Traditional MCMC approaches avoid these assumptions at the cost of increased computation due to its incompatibility to subsampling of the likelihood. New Piecewise Deterministic Markov Process (PDMP) samplers permit subsampling, though introduce a model specific inhomogenous Poisson Process (IPPs) which is difficult to sample from. This work introduces a new generic and adaptive thinning scheme for sampling from these IPPs, and demonstrates how this approach can accelerate the application of PDMPs for inference in BNNs. Experimentation illustrates how inference with these methods is computationally feasible, can improve predictive accuracy, MCMC mixing performance, and provide informative uncertainty measurements when compared against other approximate inference schemes.Comment: Includes correction to software and corrigendum not
    • …
    corecore