7,802 research outputs found

    A Utility-Theoretic Approach to Privacy in Online Services

    Get PDF
    Online offerings such as web search, news portals, and e-commerce applications face the challenge of providing high-quality service to a large, heterogeneous user base. Recent efforts have highlighted the potential to improve performance by introducing methods to personalize services based on special knowledge about users and their context. For example, a user's demographics, location, and past search and browsing may be useful in enhancing the results offered in response to web search queries. However, reasonable concerns about privacy by both users, providers, and government agencies acting on behalf of citizens, may limit access by services to such information. We introduce and explore an economics of privacy in personalization, where people can opt to share personal information, in a standing or on-demand manner, in return for expected enhancements in the quality of an online service. We focus on the example of web search and formulate realistic objective functions for search efficacy and privacy. We demonstrate how we can find a provably near-optimal optimization of the utility-privacy tradeoff in an efficient manner. We evaluate our methodology on data drawn from a log of the search activity of volunteer participants. We separately assess usersā€™ preferences about privacy and utility via a large-scale survey, aimed at eliciting preferences about peoplesā€™ willingness to trade the sharing of personal data in returns for gains in search efficiency. We show that a significant level of personalization can be achieved using a relatively small amount of information about users

    On Optimal Allocation of Indivisibles Under Uncertainty

    Get PDF
    The optimal use of indivisible resources is often the central issue in the economy and management. One of the main difficulties is the discontinuous nature of the resulting resource allocation problems which may lead to the failure of competitive market allocation mechanisms (unless we agree to "divide" the indivisibles in some indirect way). The problem becomes even more acute when uncertainty of the outcomes of decisions is present. In this paper we formalize the problem as a stochastic optimization problem involving discrete decision variables and uncertainties. By using some concrete examples, we illustrate how some problems of "dividing indivisibles" under uncertainty can be formalized in such terms. Next, we develop a general methodology to solve such problems based on the concept of the branch and bound method. The main idea of the approach is to process large collections of possible solutions and to devote more attention to the most promising groups. By gathering more information to reduce the uncertainty and by specializing the solution the optimal decision can be found

    Likelihood-Free Frequentist Inference: Bridging Classical Statistics and Machine Learning in Simulation and Uncertainty Quantification

    Full text link
    Many areas of science make extensive use of computer simulators that implicitly encode likelihood functions of complex systems. Classical statistical methods are poorly suited for these so-called likelihood-free inference (LFI) settings, outside the asymptotic and low-dimensional regimes. Although new machine learning methods, such as normalizing flows, have revolutionized the sample efficiency and capacity of LFI methods, it remains an open question whether they produce reliable measures of uncertainty. This paper presents a statistical framework for LFI that unifies classical statistics with modern machine learning to: (1) efficiently construct frequentist confidence sets and hypothesis tests with finite-sample guarantees of nominal coverage (type I error control) and power; (2) provide practical diagnostics for assessing empirical coverage over the entire parameter space. We refer to our framework as likelihood-free frequentist inference (LF2I). Any method that estimates a test statistic, like the likelihood ratio, can be plugged into our framework to create valid confidence sets and compute diagnostics, without costly Monte Carlo samples at fixed parameter settings. In this work, we specifically study the power of two test statistics (ACORE and BFF), which, respectively, maximize versus integrate an odds function over the parameter space. Our study offers multifaceted perspectives on the challenges in LF2I.Comment: 49 pages, 12 figures, code available at https://github.com/Mr8ND/ACORE-LF

    Convex Relaxations and Approximations of Chance-Constrained AC-OPF Problems

    Full text link
    This paper deals with the impact of linear approximations for the unknown nonconvex confidence region of chance-constrained AC optimal power flow problems. Such approximations are required for the formulation of tractable chance constraints. In this context, we introduce the first formulation of a chance-constrained second-order cone (SOC) OPF. The proposed formulation provides convergence guarantees due to its convexity, while it demonstrates high computational efficiency. Combined with an AC feasibility recovery, it is able to identify better solutions than chance-constrained nonconvex AC-OPF formulations. To the best of our knowledge, this paper is the first to perform a rigorous analysis of the AC feasibility recovery procedures for robust SOC-OPF problems. We identify the issues that arise from the linear approximations, and by using a reformulation of the quadratic chance constraints, we introduce new parameters able to reshape the approximation of the confidence region. We demonstrate our method on the IEEE 118-bus system
    • ā€¦
    corecore