The First Law of Complexodynamics

Why complexity rises then falls while entropy only increases

The First Law of Complexodynamics is Scott Aaronson’s exploration of why physical systems exhibit a characteristic pattern: complexity rises, peaks, then falls—even as entropy monotonically increases.

The Puzzle

Entropy always increases (Second Law of Thermodynamics):

S(t2)S(t1)for t2>t1S(t_2) \geq S(t_1) \quad \text{for } t_2 > t_1

But complexity behaves differently. A freshly shuffled deck isn’t complex—it’s random. An ordered deck isn’t complex—it’s simple. Complexity peaks somewhere in between.

The Coffee Example

Consider cream being poured into coffee:

TimeStateEntropyComplexity
t=0Separated layersLowLow
t=midSwirling patternsMediumHigh
t=∞Uniform mixtureHighLow

The intricate swirls are more “complex” than either extreme.

Defining Complexity

Aaronson proposes complextropy: the length of the shortest efficient program that outputs a distribution from which the observed state appears random.

For a string xx:

Complextropy(x)=minS:xSK(S)\text{Complextropy}(x) = \min_{S: x \in S} K(S)

where K(S)K(S) is the Kolmogorov complexity of set SS, and xx looks random within SS.

Interactive Demo

Watch complexity rise and fall as a system evolves:

Complexodynamics

t = 0
Low Entropy (Ordered)High Entropy (Random)
Entropy
1.000
Complexity
0.102
The Key Insight
Entropy increases monotonically. Complexity peaks at intermediate times— neither fully ordered nor fully random states are complex.

Why Complexity Peaks

At t=0t=0: Simple description (“all black on left, all white on right”)

At t=midt=\text{mid}: Complex description (must specify intricate patterns)

At t=t=\infty: Simple description (“random noise” or “uniform distribution”)

The Sophistication Connection

Kolmogorov’s sophistication formalizes this:

Soph(x)=min{K(S):xS,K(xS)logSO(1)}\text{Soph}(x) = \min \{K(S) : x \in S, K(x|S) \geq \log|S| - O(1)\}

The sophistication of a string is the complexity of the simplest set containing it as a “typical” member.

Why Ilya Included This

This essay connects:

  • Information theory (Kolmogorov complexity)
  • Physics (thermodynamics, entropy)
  • Computation (resource-bounded complexity)

Understanding these connections provides intuition for why certain structures emerge in learning systems and why some patterns are “interesting.”

Implications for AI

Deep learning loss landscapes might follow similar dynamics:

  • Early training: Simple patterns (high loss, low complexity)
  • Mid training: Complex intermediate features
  • Late training: Simplified, generalizable representations

Key Resource

Found an error or want to contribute? Edit on GitHub