The mathematical foundation for measuring information content and randomness
Kolmogorov Complexity and Algorithmic Randomness (by Shen, Uspensky, and Vereshchagin) is the definitive textbook on algorithmic information theory—the study of complexity, randomness, and information from a computational perspective.
Core Definition
The Kolmogorov complexity of a string is the length of the shortest program that outputs :
where is a universal Turing machine and is the program length in bits.
Key Properties
Invariance Theorem: Choice of universal machine only affects by a constant:
Upper bound: Every string has complexity at most its length:
Incompressibility: Most strings are incompressible:
Interactive Demo
Explore how different strings have different Kolmogorov complexities:
Kolmogorov Complexity
print("0" * 16)Algorithmic Randomness
A string is algorithmically random if it’s incompressible:
Random strings have no exploitable patterns—they can’t be compressed because any shorter description would be the pattern itself.
Conditional Complexity
Given extra information :
The mutual information between and :
Uncomputability
Fundamental result: is uncomputable.
No algorithm can determine the shortest program for every string. This connects to Gödel’s incompleteness and the halting problem.
Applications to AI
Kolmogorov complexity provides foundations for:
- Minimum Description Length (MDL): Model selection by compression
- Solomonoff Induction: Universal prior for prediction
- AIXI: Theoretical optimal agent framework
- Occam’s Razor: Formalized as preferring low- hypotheses
The Chain Rule
Joint complexity equals one string plus the other given the first.
Sophistication vs Complexity
Kolmogorov complexity: Total information in a string
Sophistication: Structure/pattern complexity (separating regularity from randomness)
A random string has high complexity but low sophistication—it’s complex only because it’s random, not because it has intricate structure.
Why Ilya Included This
This book provides the theoretical foundation for understanding:
- Why simpler models generalize better (MDL principle)
- The nature of learning and compression
- Information-theoretic limits on prediction
These concepts underpin much of modern machine learning theory.
Key Resource
- Kolmogorov Complexity and Algorithmic Randomness — Shen, Uspensky, Vereshchagin
https://www.lirmm.fr/~ashen/kolmbook-eng-scan.pdf