106,736 research outputs found
Statistical thermodynamics for choice models on graphs
Formalism based on equilibrium statistical thermodynamics is applied to
communication networks of decision making individuals. It is shown that in
statistical ensembles for choice models, properly defined disutility can play
the same role as energy in statistical mechanics. We demonstrate additivity and
extensivity of disutility and build three types of equilibrium statistical
ensembles: the canonical, the grand canonical and the super-canonical. Using
Boltzmann-like probability measure one reproduce the logit choice model. We
also propose using q-distributions for temperature evolution of moments of
stochastic variables. The formalism is applied to three network topologies of
different degrees of symmetry, for which in many cases analytic results are
obtained and numerical simulations are performed for all of them. Possible
applications of the model to airline networks and its usefulness for practical
support of economic decisions is pointed out.Comment: 17 pages, 13 figure
Estimation of heterogeneous preferences, with an application to demand for internet services
This paper presents a structural econometric framework for discrete and continuous consumer choices in which unobserved intrapersonal and interpersonal preference heterogeneity is modeled explicitly. It outlines a simulation-assisted estimation methodology applicable in this framework. This methodology is illustrated in an application to analyze data from the U.C. Berkeley Internet Demand Experiment
Portfolio selection models: A review and new directions
Modern Portfolio Theory (MPT) is based upon the classical Markowitz model which uses variance as a risk measure. A generalization of this approach leads to mean-risk models, in which a return distribution is characterized by the expected value of return (desired to be large) and a risk value (desired to be kept small). Portfolio choice is made by solving an optimization problem, in which the portfolio risk is minimized and a desired level of expected return is specified as a constraint. The need to penalize different undesirable aspects of the return distribution led to the proposal of alternative risk measures, notably those penalizing only the downside part (adverse) and not the upside (potential). The downside risk considerations constitute the basis of the Post Modern Portfolio Theory (PMPT). Examples of such risk measures are lower partial moments, Value-at-Risk (VaR) and Conditional Value-at-Risk (CVaR). We revisit these risk measures and the resulting mean-risk models. We discuss alternative models for portfolio selection, their choice criteria and the evolution of MPT to PMPT which incorporates: utility maximization and stochastic dominance
Equation of state for agents on graphs
Choice models for populations of agents on graphs are studied in terms of
statistical thermodynamics. Equations of state are derived and discussed for
different connectivity schemes, utility approximations, and temperature and
volume regimes. Analogies to ideal classical and quantum gases are found and
features specific for network systems are discussed.Comment: The Eur. Phys. J. B, in prin
Recommended from our members
Econometrics: A bird's eye view
As a unified discipline, econometrics is still relatively young and has been transforming and expanding very rapidly over the past few decades. Major advances have taken place in the analysis of cross sectional data by means of semi-parametric and non-parametric techniques. Heterogeneity of economic relations across individuals, firms and industries is increasingly acknowledge and attempts have been made to take them into account either by integrating out their effects or by modeling the sources of heterogeneity when suitable panel data exists. The counterfactual considerations that underlie policy analysis and treatment evaluation have been given a more satisfactory foundation. New time series econometric techniques have been developed and employed extensively in the areas of macroeconometrics and finance. Non-linear econometric techniques are used increasingly in the analysis of cross section and time series observations. Applications of Bayesian techniques to econometric problems have been given new impetus largely thanks to advances in computer power and computational techniques. The use of Bayesian techniques have in turn provided the investigators with a unifying framework where the tasks and forecasting, decision making, model evaluation and learning can be considered as parts of the same interactive and iterative process; thus paving the way for establishing the foundation of the "real time econometrics". This paper attempts to provide an overview of some of these developments
Using graphical models and multi-attribute utility theory for probabilistic uncertainty handling in large systems, with application to nuclear emergency management
Although many decision-making problems involve uncertainty, uncertainty handling within large decision support systems (DSSs) is challenging. One domain where uncertainty handling is critical is emergency response management, in particular nuclear emergency response, where decision making takes place in an uncertain, dynamically changing environment. Assimilation and analysis of data can help to reduce these uncertainties, but it is critical to do this in an efficient and defensible way. After briefly introducing the structure of a typical DSS for nuclear emergencies, the paper sets up a theoretical structure that enables a formal Bayesian decision analysis to be performed for environments like this within a DSS architecture. In such probabilistic DSSs many input conditional probability distributions are provided by different sets of experts overseeing different aspects of the emergency. These probabilities are then used by the decision maker (DM) to find her optimal decision. We demonstrate in this paper that unless due care is taken in such a composite framework, coherence and rationality may be compromised in a sense made explicit below. The technology we describe here builds a framework around which Bayesian data updating can be performed in a modular way, ensuring both coherence and efficiency, and provides sufficient unambiguous information to enable the DM to discover her expected utility maximizing policy
- …