Experiment 01 · Youthful Learning

Can a 3,000-Year-Old Algorithm Train AI?

三千年前的算法能否训练人工智能?

Hexagram 4 speaks to the nature of learning itself — the student must come to the teacher, not the reverse. In testing the King Wen sequence on AI training, the question is whether this ancient framework for understanding learning applies to a fundamentally different kind of student.

By Augustin Chan with AI · 2026-03-24

The Question

The King Wen sequence is the traditional ordering of the 64 hexagrams in the I-Ching, attributed to King Wen of Zhou around 1050 BC. For three millennia, this ordering has been studied as a map of change — each hexagram flowing into the next through a logic that resists simple characterization.

In 2025, an analysis of the sequence's mathematical properties revealed something unexpected. The transitions between hexagrams exhibit high variance and zero autocorrelation — meaning each transition is maximally surprising given the previous one. In information theory, this is the signature of an anti-habituation signal: a sequence designed to prevent the observer from falling into predictable patterns.

This raised a provocative question: could these properties, refined over three millennia of human interpretation, improve the training of artificial neural networks? Modern AI training follows carefully designed schedules — learning rates that warm up, hold steady, and decay. What if the King Wen sequence encoded a better schedule?

Experiment 1: The Learning Rate Schedule

The first test was direct. In neural network training, the learning rate controls how aggressively the model updates its understanding with each batch of data. Too fast and the model overshoots; too slow and it stagnates. Conventional schedules use smooth curves — a gradual warmup, a steady period, then a gentle decay.

The King Wen schedule replaced this smooth curve with the sequence's surprise profile: high-surprise transitions produced aggressive learning, low-surprise transitions produced cautious learning. The anti-habituation property would, in theory, prevent the optimizer from settling into unproductive routines.

Six configurations were tested against a standard baseline, varying the strength of the King Wen modulation. The results were unambiguous: every King Wen configuration performed worse than the baseline. The model trained with the standard smooth schedule consistently learned more effectively.

The strongest modulation degraded performance by 2% — modest in absolute terms, but in a controlled experiment where the only variable is the schedule, this is a clear signal. The King Wen sequence's anti-habituation property, far from preventing unproductive routines, was actively disrupting productive ones.

天行,健。君子以自強不息。

易經・乾・象傳

Heaven's movement is ceaseless. The noble one matches this through continuous self-strengthening.

The first hexagram's image commentary describes unceasing, steady effort — precisely the kind of smooth, sustained optimization that gradient descent requires. The King Wen sequence's anti-habituation profile works against this principle: it disrupts continuity rather than sustaining it.

Experiment 2: The Curriculum

The second test took a different approach. Instead of modulating how fast the model learns, it changed what the model learns and when.

In education, curriculum design matters. Teaching arithmetic before calculus is more effective than the reverse. Recent research in AI training has confirmed the same principle: the order in which data is presented affects what the model ultimately learns.

The King Wen sequence was used to reorder training batches by difficulty. Each batch of data was scored by its complexity, then presented in the order dictated by the King Wen surprise profile — harder material at high-surprise positions, easier material at low-surprise positions. Five alternative orderings served as controls: sequential (no reordering), random shuffle, easy-to-hard, hard-to-easy, and a simple sawtooth pattern.

The results were surprising, but not in the way hoped. On one machine, random shuffling produced the best results — better than any structured ordering, including King Wen. King Wen was the worst-performing non-sequential ordering. On a second machine running different software, no ordering produced any measurable effect at all.

蒙:亨。匪我求童蒙,童蒙求我。初筮告,再三瀆,瀆則不告。利貞。

易經・蒙・彖

Youthful Learning: success. It is not I who seek the young learner; the young learner seeks me. At the first inquiry, I give instruction. If asked two or three times, that is importunity — and to the importunate I give no instruction.

Hexagram 4 prescribes a specific pedagogy: the student must come ready. Repetitive, forced exposure degrades learning. Yet the curriculum experiment found that forced ordering — including King Wen's — added nothing beyond what random exposure provided. The ancient text warns against exactly the kind of imposed structure the experiment tried.

What Went Wrong — Or Did It?

Two experiments, two negative results. The King Wen sequence does not improve neural network training, whether applied as a learning rate schedule or a curriculum ordering.

But negative results are not failures. They are boundaries. These experiments established, with rigorous controls and statistical analysis, that the King Wen sequence's properties — high variance and zero autocorrelation — are detrimental to gradient-based optimization. The anti-habituation signal that might benefit a human learner actively destabilizes the mathematical process by which neural networks learn.

This is not a surprising result, in retrospect. Gradient descent is a continuous optimization process that benefits from smooth, predictable updates. Introducing unpredictability into this process is like shaking a table while someone is trying to thread a needle. The hand needs to be steady, not surprising.

The question was never whether the King Wen sequence contains structure — the mathematical analysis confirms it does. The question is whether that structure is useful in a given domain. For continuous optimization, the answer is no. But domains exist where unpredictability is not a liability but a strategic advantage. The subsequent experiments in this series explore that possibility.

地勢,坤。君子以厚德載物。

易經・坤・象傳

Earth's tendency is receptive. The noble one matches this through depth of character, bearing all things.

The second hexagram counsels receptivity — accepting results as they are rather than forcing them to match expectations. A negative result, received with intellectual honesty, strengthens the foundation for future inquiry. The ground must be understood before anything can be built upon it.

Notes

[1]referenceThe information-theoretic analysis is detailed in 'King Wen AGI Framework' (Chan, 2025). The surprise values are computed as absolute differences between consecutive hexagram numbers in the sequence, capturing how much each transition deviates from its predecessor.

[2]technicalAutocorrelation measures how similar a signal is to a delayed copy of itself. Zero autocorrelation means knowing the current value tells you nothing about the next value — the opposite of a predictable pattern like a sine wave.

[3]technicalThe metric used is val_bpb (validation bits per byte) — a measure of how well the model predicts unseen text, independent of vocabulary size or model architecture. Lower is better. The baseline achieved 1.719; King Wen schedules ranged from 1.743 to 1.755. Full results in autoresearch ADR-002.

[4]referenceThe training setup follows Karpathy's autoresearch framework: a small GPT model (3M parameters, 4 layers) trained on the ClimbMix-400B dataset for 5 minutes of wall-clock time on an NVIDIA RTX 2060.

[5]technicalThe two machines were: NVIDIA RTX 2060 (CUDA/PyTorch with torch.compile) and Apple M-series (MLX framework). The CUDA machine showed random shuffle improving by 0.106 bpb over sequential — well above the 0.04 seed noise floor. The MLX machine showed improvements of 0.008–0.037 bpb — within noise. Full cross-platform analysis in autoresearch ADR-006 and ADR-006a.

[6]contextThe difficulty metric was compression ratio: how much a batch of data can be compressed by gzip. Highly compressible data (repetitive tokens) is 'easier' for the model; incompressible data is 'harder.' This follows arXiv:2506.11300, which identified compression ratio as the strongest difficulty signal for curriculum learning.

[7]referenceThe full experimental record spans 8 Architectural Decision Records (ADRs) across two repositories, encompassing 50+ individual training runs on two hardware platforms. All code, data, and results are open source: github.com/augchan42/autoresearch (CUDA) and github.com/augchan42/autoresearch-mlx (Apple Silicon).

This is the first article in a four-part series documenting the King Wen experimental research program. The series follows the hypothesis through five negative results and a domain shift — from training AI to playing ancient strategy games.

Subscribe to receive daily passages from the classical texts that inform this research.