972 research outputs found

    Particle Gibbs for Bayesian Additive Regression Trees

    Full text link
    Additive regression trees are flexible non-parametric models and popular off-the-shelf tools for real-world non-linear regression. In application domains, such as bioinformatics, where there is also demand for probabilistic predictions with measures of uncertainty, the Bayesian additive regression trees (BART) model, introduced by Chipman et al. (2010), is increasingly popular. As data sets have grown in size, however, the standard Metropolis-Hastings algorithms used to perform inference in BART are proving inadequate. In particular, these Markov chains make local changes to the trees and suffer from slow mixing when the data are high-dimensional or the best fitting trees are more than a few layers deep. We present a novel sampler for BART based on the Particle Gibbs (PG) algorithm (Andrieu et al., 2010) and a top-down particle filtering algorithm for Bayesian decision trees (Lakshminarayanan et al., 2013). Rather than making local changes to individual trees, the PG sampler proposes a complete tree to fit the residual. Experiments show that the PG sampler outperforms existing samplers in many settings

    Estimation of parameters in linear structural relationships: Sensitivity to the choice of the ratio of error variances

    Get PDF
    Maximum likelihood estimation of parameters in linear structural relationships under normality assumptions requires knowledge of one or more of the model parameters if no replication is available. The most common assumption added to the model definition is that the ratio of the error variances of the response and predictor variates is known. The use of asymptotic formulae for variances and mean squared errors as a function of sample size and the assumed value for the error variance ratio is investigated

    The Evolution of Our Preferences: Evidence from Capuchin-Monkey Trading Behavior

    Get PDF
    Behavioral economics has demonstrated systematic decision-making biases in both lab and field data. But are these biases learned or innate? We investigate this question using experiments on a novel set of subjects — capuchin monkeys. By introducing a fiat currency and trade to a capuchin colony, we are able to recover their preferences over a wide range of goods and risky choices. We show that standard price theory does a remarkably good job of describing capuchin purchasing behavior; capuchin monkeys react rationally to both price and wealth shocks. However, when capuchins are faced with more complex choices including risky gambles, they display many of the hallmark biases of human behavior, including reference-dependent choices and loss-aversion. Given that capuchins demonstrate little to no social learning and lack experience with abstract gambles, these results suggest that certain biases such as loss-aversion are an innate function of how our brains code experiences, rather than learned behavior or the result of misapplied heuristics.Prospect theory, Loss aversion, Reference dependence, Evolution, Neuroeconomics, Capuchin monkeys, Monkey business

    Training Big Random Forests with Little Resources

    Full text link
    Without access to large compute clusters, building random forests on large datasets is still a challenging problem. This is, in particular, the case if fully-grown trees are desired. We propose a simple yet effective framework that allows to efficiently construct ensembles of huge trees for hundreds of millions or even billions of training instances using a cheap desktop computer with commodity hardware. The basic idea is to consider a multi-level construction scheme, which builds top trees for small random subsets of the available data and which subsequently distributes all training instances to the top trees' leaves for further processing. While being conceptually simple, the overall efficiency crucially depends on the particular implementation of the different phases. The practical merits of our approach are demonstrated using dense datasets with hundreds of millions of training instances.Comment: 9 pages, 9 Figure

    Repurposing cancer drugs, batimastat and marimastat, to inhibit the activity of a group I metalloprotease from the venom of the Western Diamondback rattlesnake, Crotalus atrox

    Get PDF
    Snakebite envenomation causes over 140,000 deaths every year predominantly in developing countries. As a result, it is one of the most lethal neglected tropical diseases. It is associated with an incredibly complex pathophysiology due to the vast number of unique toxins/proteins found in the venoms of diverse snake species found worldwide. Here, we report the purification and functional characteristics of a group I metalloprotease (CAMP-2) from the venom of the western diamondback rattlesnake, Crotalus atrox. Its sensitivity to matrix metalloprotease inhibitors (batimastat and marimastat) was established using specific in vitro experiments and in silico molecular docking analysis. CAMP-2 shows high sequence homology to atroxase from the venom of Crotalus atrox and exhibits collagenolytic, fibrinogenolytic and mild haemolytic activities. It exerts a mild inhibitory effect on agonist-induced platelet aggregation in the absence of plasma proteins. Its collagenolytic activity was completely inhibited by batimastat and marimastat. Zinc chloride also inhibits the collagenolytic activity of CAMP-2 by around 75% at 50 M, while it is partially potentiated by calcium chloride. Molecular docking studies demonstrate that batimastat and marimastat are able to bind strongly to the active site residues of CAMP-2. This study demonstrates the impact of matrix metalloprotease inhibitors in the modulation of a purified, group I metalloprotease activities in comparison to the whole venom. By improving our understanding of snake venom metalloproteases and their sensitivity to small molecule inhibitors, we can begin to develop novel and improved treatment strategies for snakebites
    corecore