150,938 research outputs found

    Mathematical models of games of chance: Epistemological taxonomy and potential in problem-gambling research

    Get PDF
    Games of chance are developed in their physical consumer-ready form on the basis of mathematical models, which stand as the premises of their existence and represent their physical processes. There is a prevalence of statistical and probabilistic models in the interest of all parties involved in the study of gambling – researchers, game producers and operators, and players – while functional models are of interest more to math-inclined players than problem-gambling researchers. In this paper I present a structural analysis of the knowledge attached to mathematical models of games of chance and the act of modeling, arguing that such knowledge holds potential in the prevention and cognitive treatment of excessive gambling, and I propose further research in this direction

    Generalisation : graphs and colourings

    Get PDF
    The interaction between practice and theory in mathematics is a central theme. Many mathematical structures and theories result from the formalisation of a real problem. Graph Theory is rich with such examples. The graph structure itself was formalised by Leonard Euler in the quest to solve the problem of the Bridges of Königsberg. Once a structure is formalised, and results are proven, the mathematician seeks to generalise. This can be considered as one of the main praxis in mathematics. The idea of generalisation will be illustrated through graph colouring. This idea also results from a classic problem, in which it was well known by topographers that four colours suffice to colour any map such that no countries sharing a border receive the same colour. The proof of this theorem eluded mathematicians for centuries and was proven in 1976. Generalisation of graphs to hypergraphs, and variations on the colouring theme will be discussed, as well as applications in other disciplines.peer-reviewe

    Towards Machine Wald

    Get PDF
    The past century has seen a steady increase in the need of estimating and predicting complex systems and making (possibly critical) decisions with limited information. Although computers have made possible the numerical evaluation of sophisticated statistical models, these models are still designed \emph{by humans} because there is currently no known recipe or algorithm for dividing the design of a statistical model into a sequence of arithmetic operations. Indeed enabling computers to \emph{think} as \emph{humans} have the ability to do when faced with uncertainty is challenging in several major ways: (1) Finding optimal statistical models remains to be formulated as a well posed problem when information on the system of interest is incomplete and comes in the form of a complex combination of sample data, partial knowledge of constitutive relations and a limited description of the distribution of input random variables. (2) The space of admissible scenarios along with the space of relevant information, assumptions, and/or beliefs, tend to be infinite dimensional, whereas calculus on a computer is necessarily discrete and finite. With this purpose, this paper explores the foundations of a rigorous framework for the scientific computation of optimal statistical estimators/models and reviews their connections with Decision Theory, Machine Learning, Bayesian Inference, Stochastic Optimization, Robust Optimization, Optimal Uncertainty Quantification and Information Based Complexity.Comment: 37 page
    corecore