Algorithmic Information Theory: G. J. Chaitin
Algorithmic Information Theory: G. J. Chaitin
Algorithmic Information Theory: G. J. Chaitin
Chaitin
Abstract: This paper reviews algorithmic information theory, which is an attempt to apply information-theoretic and probabilistic
ideas to recursive function theory. Typical concerns in this approach are, for example, the number of bits of information required to
specify an algorithm, or the probability that a program whose bits are chosenby coin flipping produces a given output. During the past
few years the definitions of algorithmic information theory have been reformulated. The basic features of the new formalism are pre-
sented here and certain resultsof R. M . Solovay are reported.
Historical introduction
To our knowledge, the first publication of the ideas of sal machines, the induced structures cannot be desper-
algorithmicinformation theory was the description of atelydifferent.We appeal to the ‘translationtheorem’
R. J . Solomonoff’s ideas given in 1962 by M. L. Minsky wherebyanarbitrary instructionformula forone ma-
in his paper, “Problemsof formulation for artificial intelli- chine may be converted into an equivalent instruction
gence” [ 11 : formula for the other machine by the addition of a con-
“Consider a slightly different form of inductive infer- stant prefix text. This text instructs the second machine
ence problem. Suppose that we are givenavery long to simulate the behavior of the first machine in operating
‘data’ sequence of symbols; the problem is to makea on the remainder of the input text. Then for datastrings
prediction about the future of the sequence. This is a much larger than this translation text (and its inverse)
problem familiar in discussionconcerning‘inductive thechoicebetweenthetwo machines cannot greatly
probability.’ The problem is refreshed a little, perhaps, affect the induced structure. It would be interesting to
by introducing the modern notion of universal computer see if these intuitive notions could be profitably formal-
and its associated language of instruction formulas. An ized.
instruction sequence will be considered acceptable if it “Even if this theory can be worked out, it is likely that
causesthecomputertoproduce a sequence,perhaps it will present overwhelming computational difficulties in
infinite, that begins with the given finite ‘data’ sequence. application. The recognition problemfor minimal descrip-
Each acceptable instruction sequence thus makes a pre- tions is, in general, unsolvable, and a practical induction
diction, and Occam’s razor would choose the simplest machine will havetouse heuristic methods. [In this
such sequence and advocate its prediction. (More gener- connection it would be interesting to write a program to
ally, one could weight the different predictions by play R. Abbott’s inductive card game [ 21 .] ”
weights associated with the simplicities of the Algorithmic information theory originated in the inde-
instructions.) If the simplicity function is just the length pendent work of Solomonoff (see [ 1, 3 -61 1 , of A. N.
of the instruction, we are then trying to find a minimal Kolmogorov and P. Martin-Lof (see [7 - 141), and of
description, i.e., an optimally efficient encoding of the G. J. Chaitin (see [ 15-261).Whereas Solomonoff
data sequence. weighted together all the programs for a given result into
“Such an induction method could be of interest only if a probability measure, Kolmogorov and Chaitin concen-
one could show some significant invariance with respect trated their attention on the size of the smallest program.
to choice of defining universal machine. There is no such Recently it hasbeenrealized by Chaitin and indepen-
invariance for a fixed pair of data strings. For one could dently by L. A. Levin that if programs are stipulated to
design a machine which would yield the entire first string be self-delimiting, thesetwo differing approaches be-
with a very small input, and the second string only for come essentially equivalent. This paper attempts to cast
some very complex input. On the brighter side, one can into a unified scheme the recent work in this area by
see that in a sense the induced structure on the space of Chaitin [23, 241 and by R. M.Solovay [27, 281. The
data strings has some invariance in an ‘in the large’ or reader may also find it interesting to examine the parallel
350 ‘almost everywhere’ sense.
Giventwo different
univer- efforts of Levin (see [ 29 - 351 ) . There has been a sub-
Acknowledgments
Algorithmic information theory and biology The quotation by M. L.Minskyin the firstsection is
Above we have pointed out a number of open problems. reprintedwith the kind permission of the publisher
In our opinion, however, the most importantchallenge is AmericanMathematical Society
from Mathematical
to see if the ideas of algorithmic information theory can Problems in the BiologicalSciences,Proceedings of
contribute in some form or manner to theoretical mathe- Symposia in AppliedMathematics X I V , pp. 42-43,
matical biology in the style of von Neumann [54], in copyright @ 1962. We are grateful to R. M. Solovay for
which geneticinformation is considered to be an ex- permitting us to include several of his unpublished re-
tremely large and complicated program for constructing sults in the sectionentitled “Moreadvanced results.”
organisms. We alluded briefly to this in a previous paper The quotation by M. Gardner in thesectionon algo-
[21], and discussed it at greater length in a publication rithmic information theory and metamathematics is re-
[ 191 of somewhat limited access. printed with his kind permission, and the quotation by
Von Neumann wished to isolate the basic conceptual B. Russell in that section is reprinted with permission of
problems of biology from the detailed physics and bio- the Johns Hopkins University Press. We are grateful to
chemistry of life as we know it. The gist of his message C . H. Bennett for permitting us to present his notion of
is that it should be possible to formulate mathematically logical depth in print for the first time in the section on
and to answer in a quite general setting such fundamen- algorithmic information theory and biology.
tal questionsas“How is self-reproductionpossible?”,
“What is an organism?”, “What is its degree of organi- References
zation?”,and
“How probable is evolution?’. H e 1. M. L. Minsky, “Problems of Formulation for Artificial In-
achieved this for the first question; he showed that exact telligence,” Mathematical Problems in the Biological Sci-
ences, Proceedings of Symposia in Applied Mathematics
self-reproduction of universal Turing machines is possi- X I V , R. E. Bellman, ed., American Mathematical Society,
ble in a particular deterministic model universe. Providence, RI, 1962, p. 35. 357
359