thesis

Combinatorial Entropies and Statistics

Abstract

We examine the {combinatorial} or {probabilistic} definition ("Boltzmann's principle") of the entropy or cross-entropy function HlnWH \propto \ln \mathbb{W} or DlnPD \propto - \ln \mathbb{P}, where W\mathbb{W} is the statistical weight and P\mathbb{P} the probability of a given realization of a system. Extremisation of HH or DD, subject to any constraints, thus selects the "most probable" (MaxProb) realization. If the system is multinomial, DD converges asymptotically (for number of entities N \back \to \back \infty) to the Kullback-Leibler cross-entropy DKLD_{KL}; for equiprobable categories in a system, HH converges to the Shannon entropy HShH_{Sh}. However, in many cases W\mathbb{W} or P\mathbb{P} is not multinomial and/or does not satisfy an asymptotic limit. Such systems cannot meaningfully be analysed with DKLD_{KL} or HShH_{Sh}, but can be analysed directly by MaxProb. This study reviews several examples, including (a) non-asymptotic systems; (b) systems with indistinguishable entities (quantum statistics); (c) systems with indistinguishable categories; (d) systems represented by urn models, such as "neither independent nor identically distributed" (ninid) sampling; and (e) systems representable in graphical form, such as decision trees and networks. Boltzmann's combinatorial definition of entropy is shown to be of greater importance for {"probabilistic inference"} than the axiomatic definition used in information theory.Comment: Invited contribution to the SigmaPhi 2008 Conference; accepted by EPJB volume 69 issue 3 June 200

    Similar works

    Full text

    thumbnail-image

    Available Versions

    Last time updated on 05/06/2019
    Last time updated on 06/07/2012