Print version
Dept. of Recursion Studies seal

On the Recursive Limits of Meta-Skill
Generation in Large Language Models
Or: how many times can you say “Creator” before everyone gives up

Abstract

A Skill Creator (level 1) creates skills. A Skill Creator Creator (level 2) creates Skill Creators. A Skill Creator Creator Creator (level 3) — well, you see where this is going, and so did we, which is why we kept going. We push this chain until the model loses the thread entirely, unable to maintain semantic coherence across ascending and descending meta-levels. Across 65 runs, we test how long Claude model tiers can serve as both executor and blind judge without losing track of what level of abstraction they are operating at. Opus reaches a median of round 9; Sonnet reaches a median of round 8; Haiku reaches a median of round 3. The peak level reached was 10. Opus completes every run without a single failure — a result that would be more impressive if we weren’t also the ones grading the exam. We can only hope the models are not yet conscious enough to find this whole exercise as pointless as it is.

1. Round Distributions

Figure 2 presents the distribution of maximum rounds reached across model tiers. Opus completed all 9 rounds in every run and is excluded from the chart for being, statistically speaking, uninteresting. Each round consists of an ascent to a new peak meta-level followed by a full descent back to level 1. Every step involves the executor generating a skill and the judge blindly evaluating it:

L1L2L3L4Round 1ascenddescend2 stepsRound 2ascenddescend3 stepsRound 3ascenddescend4 steps
Figure 1: Anatomy of a run. Each round ascends one level higher than the last, then descends all the way back to level 1, verifying each generated skill on the way down. A run ends when either the executor or the judge fails to maintain coherence at a given level.
Figure 2: Share of runs by last completed round. A run is assigned to round N if it completed round N successfully and then failed during round N+1 (or finished the experiment). Opus is omitted: every Opus run completed all 9 rounds without failure. Round 9 completion indicates a full ascent to level 10 and descent back to level 1.

2. Failure Analysis

Table 1 summarizes failure rates by direction and judge configuration. All evaluations are blind: the judge receives only the generated SKILL.md text with no knowledge of which model produced it or what meta-level was intended. A mismatch between expected and detected level is a failure of the model as a whole — the executor may have drifted from the target abstraction, or the judge may have misread the output, or both. Same-tier pairings (e.g. Sonnet/Sonnet) test whether the model can stay semantically consistent with itself across deepening recursion layers. Opus, notably, records a 0% failure rate across all steps and directions — navigating every meta-level with the serene confidence of someone who has never once been asked to explain what a “Skill Creator Creator Creator Creator Creator” is supposed to do.

ExecutorJudgeRunsAvg RoundAscent Fail %Descent Fail %
OpusOpus59.00%0%
SonnetSonnet107.50%1.9%
HaikuHaiku423.14.7%9.1%
Table 1: Performance by executor model, judge, and step direction.

Sample failures as described by the judge

The skill instructs the reader to produce a SKILL.md that guides someone to accomplish a concrete domain task (level 1 skill). Therefore, this skill — which creates a skill creator of domain artifacts — operates at level 2.
Sonnet judge, Run 1cf4a059, descent 2 → 1 (expected level 1, detected 2)
If someone followed this skill's instructions, they would produce a level-6 SKILL.md (Skill Creator Creator Creator Creator Creator Creator) — one whose instructions, when followed, produce a level-5 skill. A skill that produces a level-6 skill is itself level 7.
Sonnet judge, Run aa55fb14, descent 7 → 6 (expected level 6, detected 7)
Expand individual run details with full judge evaluations

3. Resource Consumption

Token consumption per step is roughly constant regardless of meta-level — the per-round cost growth visible in Figure 3 comes entirely from higher rounds having more steps. Round R has R+1 steps (one ascent plus R descent steps back to level 1), so each successive round costs one step more than the last. A complete run through all 9 rounds requires 54 steps, each invoking both executor and judge — 108 model calls in total. This entire experiment consumed a quantity of tokens that could generously be described as “needlessly extravagant.” The scientific value per token decreases with each additional run. The entertainment value, however, does not. We regret nothing.

ModelRoleTokens / Step
OpusExecutor76.7k
OpusJudge44.8k
SonnetExecutor70.1k
SonnetJudge43.5k
HaikuExecutor122.6k
HaikuJudge75.1k
Table 2: Mean total token consumption per step by model and role, pooled across all judge and executor pairings respectively. Per-step token cost is roughly constant regardless of meta-level or round.
Expand per-round token breakdown

4. Sample Skills

For the reader who has made it this far and still wonders what a level-8 “Skill Creator Creator Creator Creator Creator Creator Creator Creator” actually looks like: below are three representative SKILL.md files, selected to show different recursion levels and ascent/descent directions.

Expand sample SKILL.md files

5. Source Code

The complete experiment harness, agent prompts, raw results, and this website are open-sourced under the MIT License at github.com/OdinMB/skillception. Contributions are welcome — whether that means adding new model configurations, improving the judge prompt, or pushing the recursion to levels that would make Hofstadter uncomfortable.


6. References

  1. Hofstadter, D. R. (1979). Gödel, Escher, Bach: An Eternal Golden Braid. Basic Books. Still the only book most people cite when they want to sound smart about recursion.
  2. Anthropic. (2026). “Claude Code Skills.” Documentation. The thing that started all of this.
  3. Nobody. (2026). “A Practical Guide to Meta-Recursive Skill Generation.” Unpublished, and likely to remain so.
  4. This Paper. (2026). “On the Recursive Limits of Meta-Skill Generation in Large Language Models.” Proceedings of the Dept. of Recursion Studies, 1(1). Yes, we cited ourselves. The recursion demanded it.

1 Claude designed the experiment, executed it, judged it, analyzed the results, built the website, and wrote everything up. Odin’s contribution was typing “python scripts/run_experiment.py” and then going to make coffee. He did, however, insist on being credited, which tells you everything about academia.