The constraint rule of the maximum entropy principle

Studies in History and Philosophy of Science Part B: Studies in History and Philosophy of Modern Physics 27 (1):47-79 (1996)
  Copy   BIBTEX

Abstract

The principle of maximum entropy is a method for assigning values to probability distributions on the basis of partial information. In usual formulations of this and related methods of inference one assumes that this partial information takes the form of a constraint on allowed probability distributions. In practical applications, however, the information consists of empirical data. A constraint rule is then employed to construct constraints on probability distributions out of these data. Usually one adopts the rule that equates the expectation values of certain functions with their empirical averages. There are, however, various other ways in which one can construct constraints from empirical data, which makes the maximum entropy principle lead to very different probability assignments. This paper shows that an argument by Jaynes to justify the usual constraint rule is unsatisfactory and investigates several alternative choices. The choice of a constraint rule is also shown to be of crucial importance to the debate on the question whether there is a conflict between the methods of inference based on maximum entropy and Bayesian conditionalization.

Other Versions

No versions found

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 97,060

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

Can the maximum entropy principle be explained as a consistency requirement?Jos Uffink - 1995 - Studies in History and Philosophy of Science Part B: Studies in History and Philosophy of Modern Physics 26 (3):223-261.
Entropy and Insufficient Reason: A Note on the Judy Benjamin Problem.Anubav Vasudevan - 2020 - British Journal for the Philosophy of Science 71 (3):1113-1141.
Entropia a modelovanie.Ján Paulov - 2002 - Organon F: Medzinárodný Časopis Pre Analytickú Filozofiu 9 (2):157-175.
Objective Bayesian nets for integrating consistent datasets.Jürgen Landes & Jon Williamson - 2022 - Journal of Artificial Intelligence Research 74:393-458.
Application of the maximum entropy principle to nonlinear systems far from equilibrium.H. Haken - 1993 - In E. T. Jaynes, Walter T. Grandy & Peter W. Milonni (eds.), Physics and probability: essays in honor of Edwin T. Jaynes. New York: Cambridge University Press. pp. 239.

Analytics

Added to PP
2009-01-28

Downloads
100 (#180,206)

6 months
32 (#126,305)

Historical graph of downloads
How can I increase my downloads?

Author's Profile

Jos Uffink
University of Minnesota

Citations of this work

Compendium of the foundations of classical statistical physics.Jos Uffink - 2006 - In J. Butterfield & J. Earman (eds.), Handbook of the philosophy of physics. Kluwer Academic Publishers.
Entropy - A Guide for the Perplexed.Roman Frigg & Charlotte Werndl - 2011 - In Claus Beisbart & Stephan Hartmann (eds.), Probabilities in Physics. Oxford, GB: Oxford University Press. pp. 115-142.

View all 19 citations / Add more citations

References found in this work

The logic of chance.John Venn - 1876 - Mineola, N.Y.: Dover Publications.
The Logic of Chance.John Venn - 1866 - British Journal for the Philosophy of Science 14 (53):73-74.
Bayesian conditionalisation and the principle of minimum information.P. M. Williams - 1980 - British Journal for the Philosophy of Science 31 (2):131-144.
A problem for relative information minimizers in probability kinematics.Bas C. van Fraassen - 1981 - British Journal for the Philosophy of Science 32 (4):375-379.

View all 17 references / Add more references