The constraint rule of the maximum entropy principle

Studies in History and Philosophy of Science Part B: Studies in History and Philosophy of Modern Physics 27 (1):47-79 (1996)
  Copy   BIBTEX

Abstract

The principle of maximum entropy is a method for assigning values to probability distributions on the basis of partial information. In usual formulations of this and related methods of inference one assumes that this partial information takes the form of a constraint on allowed probability distributions. In practical applications, however, the information consists of empirical data. A constraint rule is then employed to construct constraints on probability distributions out of these data. Usually one adopts the rule that equates the expectation values of certain functions with their empirical averages. There are, however, various other ways in which one can construct constraints from empirical data, which makes the maximum entropy principle lead to very different probability assignments. This paper shows that an argument by Jaynes to justify the usual constraint rule is unsatisfactory and investigates several alternative choices. The choice of a constraint rule is also shown to be of crucial importance to the debate on the question whether there is a conflict between the methods of inference based on maximum entropy and Bayesian conditionalization.

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 91,386

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

Analysis of the maximum entropy principle “debate”.John F. Cyranski - 1978 - Foundations of Physics 8 (5-6):493-506.
Entropia a modelovanie.Ján Paulov - 2002 - Organon F: Medzinárodný Časopis Pre Analytickú Filozofiu 9 (2):157-175.
Maximum Entropy Inference with Quantified Knowledge.Owen Barnett & Jeff Paris - 2008 - Logic Journal of the IGPL 16 (1):85-98.
Can the maximum entropy principle be explained as a consistency requirement?Jos Uffink - 1995 - Studies in History and Philosophy of Science Part B: Studies in History and Philosophy of Modern Physics 26 (3):223-261.
Application of the maximum entropy principle to nonlinear systems far from equilibrium.H. Haken - 1993 - In E. T. Jaynes, Walter T. Grandy & Peter W. Milonni (eds.), Physics and Probability: Essays in Honor of Edwin T. Jaynes. Cambridge University Press. pp. 239.
Common sense and maximum entropy.Jeff Paris - 1998 - Synthese 117 (1):75-93.

Analytics

Added to PP
2009-01-28

Downloads
108 (#159,595)

6 months
20 (#125,481)

Historical graph of downloads
How can I increase my downloads?

Author's Profile

Jos Uffink
University of Minnesota

Citations of this work

Compendium of the foundations of classical statistical physics.Jos Uffink - 2005 - In Jeremy Butterfield & John Earman (eds.), Handbook of the Philosophy of Physics. Elsevier.
Entropy - A Guide for the Perplexed.Roman Frigg & Charlotte Werndl - 2011 - In Claus Beisbart & Stephan Hartmann (eds.), Probabilities in Physics. Oxford University Press. pp. 115-142.

View all 19 citations / Add more citations

References found in this work

The logic of chance.John Venn - 1876 - Mineola, N.Y.: Dover Publications.
The Logic of Chance.John Venn - 1866 - British Journal for the Philosophy of Science 14 (53):73-74.
Bayesian conditionalisation and the principle of minimum information.P. M. Williams - 1980 - British Journal for the Philosophy of Science 31 (2):131-144.
A problem for relative information minimizers in probability kinematics.Bas C. van Fraassen - 1981 - British Journal for the Philosophy of Science 32 (4):375-379.

View all 18 references / Add more references