3,137 research outputs found

    Entropy Concentration and the Empirical Coding Game

    Full text link
    We give a characterization of Maximum Entropy/Minimum Relative Entropy inference by providing two `strong entropy concentration' theorems. These theorems unify and generalize Jaynes' `concentration phenomenon' and Van Campenhout and Cover's `conditional limit theorem'. The theorems characterize exactly in what sense a prior distribution Q conditioned on a given constraint, and the distribution P, minimizing the relative entropy D(P ||Q) over all distributions satisfying the constraint, are `close' to each other. We then apply our theorems to establish the relationship between entropy concentration and a game-theoretic characterization of Maximum Entropy Inference due to Topsoe and others.Comment: A somewhat modified version of this paper was published in Statistica Neerlandica 62(3), pages 374-392, 200

    On derivations with respect to finite sets of smooth functions

    Full text link
    The purpose of this paper is to show that functions that derivate the two-variable product function and one of the exponential, trigonometric or hyperbolic functions are also standard derivations. The more general problem considered is to describe finite sets of differentiable functions such that derivations with respect to this set are automatically standard derivations

    On the equality problem of generalized Bajraktarevi\'c means

    Get PDF
    The purpose of this paper is to investigate the equality problem of generalized Bajraktarevi\'c means, i.e., to solve the functional equation \begin{equation}\label{E0}\tag{*} f^{(-1)}\bigg(\frac{p_1(x_1)f(x_1)+\dots+p_n(x_n)f(x_n)}{p_1(x_1)+\dots+p_n(x_n)}\bigg)=g^{(-1)}\bigg(\frac{q_1(x_1)g(x_1)+\dots+q_n(x_n)g(x_n)}{q_1(x_1)+\dots+q_n(x_n)}\bigg), \end{equation} which holds for all x=(x1,…,xn)∈Inx=(x_1,\dots,x_n)\in I^n, where n≥2n\geq 2, II is a nonempty open real interval, the unknown functions f,g:I→Rf,g:I\to\mathbb{R} are strictly monotone, f(−1)f^{(-1)} and g(−1)g^{(-1)} denote their generalized left inverses, respectively, and p=(p1,…,pn):I→R+np=(p_1,\dots,p_n):I\to\mathbb{R}_{+}^n and q=(q1,…,qn):I→R+nq=(q_1,\dots,q_n):I\to\mathbb{R}_{+}^n are also unknown functions. This equality problem in the symmetric two-variable (i.e., when n=2n=2) case was already investigated and solved under sixth-order regularity assumptions by Losonczi in 1999. In the nonsymmetric two-variable case, assuming three times differentiability of ff, gg and the existence of i∈{1,2}i\in\{1,2\} such that either pip_i is twice continuously differentiable and p3−ip_{3-i} is continuous on II, or pip_i is twice differentiable and p3−ip_{3-i} is once differentiable on II, we prove that \eqref{E0} holds if and only if there exist four constants a,b,c,d∈Ra,b,c,d\in\mathbb{R} with ad≠bcad\neq bc such that \begin{equation*} cf+d>0,\qquad g=\frac{af+b}{cf+d},\qquad\mbox{and}\qquad q_\ell=(cf+d)p_\ell\qquad (\ell\in\{1,\dots,n\}). \end{equation*} In the case n≥3n\geq 3, we obtain the same conclusion with weaker regularity assumptions. Namely, we suppose that ff and gg are three times differentiable, pp is continuous and there exist i,j,k∈{1,…,n}i,j,k\in\{1,\dots,n\} with i≠j≠k≠ii\neq j\neq k\neq i such that pi,pj,pkp_i,p_j,p_k are differentiable

    Inconsistency of Bayesian Inference for Misspecified Linear Models, and a Proposal for Repairing It

    Get PDF
    We empirically show that Bayesian inference can be inconsistent under misspecification in simple linear regression problems, both in a model averaging/selection and in a Bayesian ridge regression setting. We use the standard linear model, which assumes homoskedasticity, whereas the data are heteroskedastic, and observe that the posterior puts its mass on ever more high-dimensional models as the sample size increases. To remedy the problem, we equip the likelihood in Bayes' theorem with an exponent called the learning rate, and we propose the Safe Bayesian method to learn the learning rate from the data. SafeBayes tends to select small learning rates as soon the standard posterior is not `cumulatively concentrated', and its results on our data are quite encouraging.Comment: 70 pages, 20 figure
    • …
    corecore