Abstract
‘Information transmitted’ is defined as the amount by which added evidence (or ‘message received’) diminishes ‘uncertainty’. The latter is characterized by some properties intuitively suggested by this word and possessed by conditional entropy, a parameter of the posterior probability distribution. However, conditional entropy shares these properties with some other concave symmetric functions on the probability space.
Moreover, a given transmission channel (or, in the context of statistical inference, a given experiment) yields a higher maximum expected benefit than anotherto any user if and only ifall concave functions of the posterior probability vector have higher values for the former channel (or experiment). Hence one information system (channel, experiment) may be preferable to another for a given user although its transmission rate, in entropy terms, is lower.
But only entropy has the economically relevant property of measuring, in the limit, the expected length of efficiently coded messages sent in long sequences. Thus, while irrelevant to the value (maximum expected benefit) of an information system and to the costs of observing, estimating, and deciding, entropy formulas are indeed relevant to the cost of communicating, i.e., of storing, coding and transmitting messages.
Similar content being viewed by others
References
Aczél, J., ‘On Different Characterizations of Entropies’, in M. Beharaet al. (eds.),Probability and Information Theory, 1–11, Springer (1969).
Blackwell, D., ‘Equivalent Comparisons of Experiments’,Ann. Math. Stat. 24(1953) 265–272.
Blackwell, D. and Girshick, A.,Theory of Games and Statistical Decisions, McGraw-Hill (1970).
V. Boehm, Personal Communication.
DeGroot, M. H., ‘Uncertainty, Information and Sequential Experiments’,Ann. Math. Stat. 33 (1962) 404–419.
DeGroot, M. H.,Optimal Statistical Decisions, McGraw-Hill (1970).
Feinstein, A.,Foundations of Information Theory, McGraw-Hill (1958).
Marschak, J., ‘Economics of Information Systems’,J. Amer. Stat. Assn. 66 (1971) 192–219.
Marschak, J., ‘Optimal Systems for Information and Decision’, inTechniques of Optimization, Academic Press (1972).
Marschak, J. and Miyasawa, K., ‘Economic Comparability of Information Systems’,Intern. Econ. Rev. 9 (1968) 137–174.
Savage, L. J.,The Foundations of Statistics, Wiley (1954).
Schroedinger, E.,Statistical Thermodynamics, Cambridge University Press (1948).
Shannon, C., ‘The Mathematical Theory of Communication’,Bell Syst. Tech. J. (1948).
Wolfowitz, T.,Coding Theorems of Information Theory, Springer (1961).
Zangwill, W. I.,Nonlinear Programming. A United Approach. Prentice-Hall, 1969.
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Marschak, J. Limited role of entropy in information economics. Theor Decis 5, 1–7 (1974). https://doi.org/10.1007/BF00140297
Issue Date:
DOI: https://doi.org/10.1007/BF00140297