345 research outputs found

    Probabilistic multiple kernel learning

    Get PDF
    The integration of multiple and possibly heterogeneous information sources for an overall decision-making process has been an open and unresolved research direction in computing science since its very beginning. This thesis attempts to address parts of that direction by proposing probabilistic data integration algorithms for multiclass decisions where an observation of interest is assigned to one of many categories based on a plurality of information channels

    Boosted Off-Policy Learning

    Full text link
    We investigate boosted ensemble models for off-policy learning from logged bandit feedback. Toward this goal, we propose a new boosting algorithm that directly optimizes an estimate of the policy's expected reward. We analyze this algorithm and prove that the empirical risk decreases (possibly exponentially fast) with each round of boosting, provided a "weak" learning condition is satisfied. We further show how the base learner reduces to standard supervised learning problems. Experiments indicate that our algorithm can outperform deep off-policy learning and methods that simply regress on the observed rewards, thereby demonstrating the benefits of both boosting and choosing the right learning objective

    Banking the unbanked: the Mzansi intervention in South Africa:

    Get PDF
    Purpose This paper aims to understand household’s latent behaviour decision making in accessing financial services. In this analysis we look at the determinants of the choice of the pre-entry Mzansi account by consumers in South Africa. Design/methodology/approach We use 102 variables, grouped in the following categories: basic literacy, understanding financial terms, targets for financial advice, desired financial education and financial perception. Employing a computationally efficient variable selection algorithm we study which variables can satisfactorily explain the choice of a Mzansi account. Findings The Mzansi intervention is appealing to individuals with basic but insufficient financial education. Aspirations seem to be very influential in revealing the choice of financial services and to this end Mzansi is perceived as a pre-entry account not meeting the aspirations of individuals aiming to climb up the financial services ladder. We find that Mzansi holders view the account mainly as a vehicle for receiving payments, but on the other hand are debt-averse and inclined to save. Hence although there is at present no concrete evidence that the Mzansi intervention increases access to finance via diversification (i.e. by recruiting customers into higher level accounts and services) our analysis shows that this is very likely to be the case. Originality/value The issue of demand side constraints on access to finance have been largely ignored in the theoretical and empirical literature. This paper undertakes some preliminary steps in addressing this gap

    An adaptive ensemble learner function via bagging and rank aggregation with applications to high dimensional data.

    Get PDF
    An ensemble consists of a set of individual predictors whose predictions are combined. Generally, different classification and regression models tend to work well for different types of data and also, it is usually not know which algorithm will be optimal in any given application. In this thesis an ensemble regression function is presented which is adapted from Datta et al. 2010. The ensemble function is constructed by combining bagging and rank aggregation that is capable of changing its performance depending on the type of data that is being used. In the classification approach, the results can be optimized with respect to performance measures such as accuracy, sensitivity, specificity and area under the curve (AUC) whereas in the regression approach, it can be optimized with respect to measures such as mean square error and mean absolute error. The ensemble classifier and ensemble regressor performs at the level of the best individual classifier or regression model. For complex high-dimensional datasets, it may be advisable to combine a number of classification algorithms or regression algorithms rather than using one specific algorithm

    Identifying Sources of a Change in Metrics of a Stack of Servers

    Get PDF
    Generally, the present disclosure is directed to identifying a source of a change in one or more metrics of one or more servers. In particular, in some implementations, the systems and methods of the present disclosure can include or otherwise leverage one or more machine-learned models to predict contribution from a component within a stack of one or more servers based on one or more states of the component

    Stat Methods Med Res

    Get PDF
    This paper investigates different approaches for causal estimation under multiple concurrent medications. Our parameter of interest is the marginal mean counterfactual outcome under different combinations of medications. We explore parametric and non-parametric methods to estimate the generalized propensity score. We then apply three causal estimation approaches (inverse probability of treatment weighting, propensity score adjustment, and targeted maximum likelihood estimation) to estimate the causal parameter of interest. Focusing on the estimation of the expected outcome under the most prevalent regimens, we compare the results obtained using these methods in a simulation study with four potentially concurrent medications. We perform a second simulation study in which some combinations of medications may occur rarely or not occur at all in the dataset. Finally, we apply the methods explored to contrast the probability of patient treatment success for the most prevalent regimens of antimicrobial agents for patients with multidrug-resistant pulmonary tuberculosis.CC999999/Intramural CDC HHS/United StatesK24 AI114444/AI/NIAID NIH HHS/United StatesP30 AI050409/AI/NIAID NIH HHS/United StatesU19 AI111211/AI/NIAID NIH HHS/United States2019-12-01T00:00:00Z30381005PMC6511477696

    Contextual Advertising Based on Content Recognition in a Video

    Get PDF
    Generally, the present disclosure is directed to providing relevant advertisements based on the visual content of a video. In particular, in some implementations, the systems and methods of the present disclosure can include or otherwise leverage one or more machine-learned models to determine a relevant advertisement and/or a relevant time for the relevant advertisement based on image data taken from a video

    Using Imagery to Identify Abandoned Property in Public Spaces

    Get PDF
    Generally, the present disclosure is directed to identifying private property that has been illegally dumped or stored for a prolonged period without use. In particular, in some implementations, the systems and methods of the present disclosure can include or otherwise leverage one or more machine-learned models to predict whether an item has been abandoned in a public right-of-way based on imagery of the item and record of public right-of-way

    Training High Quality Spam-detection Models Using Weak Labels

    Get PDF
    To be effective in detecting spam in online content sharing networks, it is necessary that techniques used to detect spam have good precision, high recall, and the ability to adapt to new types of spam. A bottleneck in developing such machine learning techniques is the lack of availability of high quality labeled training data. Human labeling to obtain high quality labeled data is expensive and not scalable. Current approaches such as unsupervised learning or semi-supervised learning can only produce low quality labels. Generally, the present disclosure is directed to a weak supervision approach to train a machine learning model to detect spam content items. Weak labels are generated for content items in training data using various techniques such as rules that encode domain knowledge and/or anomaly detection techniques such as unsupervised machine learning/ clustering or semi-supervised machine learning. The accuracy of the various techniques is estimated based on observed agreements/ disagreements in the weak labels. The weak labels are combined into a single value (e.g., per content item) that is used as a probabilistic training label to train a machine learning model using supervised learning that is noise aware. In the training, a penalty is applied for deviation from the probabilistic label such that the penalty is higher for a label associated with a higher confidence and lower for a label associated with a lower confidence. The model thus trained can be used to detect spam content
    • …
    corecore