HEGEL

v1 Final Results
Dashboard How it works Concepts Test Results v1 Results Settings

The Thesis

What We Set Out to Prove

Hegel's developmental sequence in the Phenomenology of Spirit describes real structural necessities. A system under survival pressure with local learning rules traverses developmental stages in a specific order because it must. Skip one and the later ones fail. The sequence is not arbitrary.

What We Proved

A gradient-free, viability-constrained organism starting with nothing but metabolism and a survival boundary develops — through its own specific failures — sensorimotor coupling, adaptive learning, environmental prediction, self-modeling, practical self-knowledge, perception of another agent, and mutual tracking. Layers 1–8 pass. Each transition is driven by determinate negation: a specific failure producing a specific restructuring.

Layer 9 (Observing Reason / curiosity-driven exploration) identifies the representational ceiling of a 20-neuron, 5-chemical world. The organism has learned everything this world has to offer. Curiosity has nowhere productive to go. This is not a failure — it is a boundary condition that specifies exactly what scaling must provide.

System Architecture

The ODE Core

The organism is a 25-dimensional continuous ODE solved by Tsit5 (adaptive Runge-Kutta). No discrete time steps. No frame-by-frame updates. The organism exists as a continuous dynamical system.

State Vector u[1:25] ───────────────────────────────────────── u[1:5] = Metabolic: S (substrate), T (catalyst-1), U (catalyst-2), C1 (enzyme-1), C2 (enzyme-2) u[6:25] = CTRNN: 20 continuous-time recurrent neurons Sensory Input [1:16] ───────────────────────────────────────── [1:5] = Metabolic state (direct observation) [6:10] = Affective derivatives tanh((current-prev)*10) [11:13] = Environmental (resource gradients) [14:16] = Other-agent motor (delayed 5-step, 0 in solo) Motor Output [1:3] ───────────────────────────────────────── M1 = Intake rate (resource acquisition) M2 = Balance (metabolic regulation) M3 = Efficiency (catalytic optimization)

Viability boundary: The organism dies if any metabolic variable exits [0.01, 10.0]. There is no reward signal, no fitness function during life. The only objective is: don't die. Everything else — learning, self-modeling, recognition — emerges from survival pressure.

CTRNN + Hebbian Learning

CTRNN: 20 neurons with continuous-time dynamics. Sensory input through W_in (20×16), recurrent connections W (20×20), motor readout W_out (3×20). Activation function: tanh. Time constants τ evolve via natural selection.

Hebbian learning: Three rules operating simultaneously, only while the organism is alive:

Oja rule (W_in): Normalized Hebbian. Extracts principal components of sensory-neural correlation.
BCM rule (W): Bienenstock-Cooper-Munro. Activity-dependent with sliding threshold. Creates competition among neurons.
Anti-Hebbian (W_out): Competitive motor readout. Neurons compete for motor channel influence.

Neuromodulation gate: gate = max(stress, surprise) * PCN_i. Learning is gated by both the organism's stress level and per-neuron prediction error (PCN). Neurons that already predict well don't learn — only confused neurons update. This was a critical discovery: unmodulated Hebbian learning was net harmful.

Self-Model

The organism develops a predictive model of its own sensory consequences. Not pre-installed — it emerges from the prediction module detecting that some prediction errors are correlated with motor output (self-caused) while others are not (environmental).

structural_signal = corr(residual_error, motor_deviation)
where residual = prediction_error - EMA(prediction_error)
and motor_deviation = motor_magnitude - EMA(motor_magnitude)

When structural_signal exceeds 0.1, the self-model activates: a linear map W_self (16×3) that predicts sensory change from motor output. This map is trained by Hebbian learning (not gradient descent). The self-model enables:

Cognitive Modules (v1.6.0)

Three cognitive capacities added in v1.6.0, all gradient-free:

Episodic Memory (Erinnerung): Ring buffer of 64 past (sensory, motor, outcome) tuples. Retrieves most similar past experience via cosine similarity. Past motor actions blend into current output when similarity > 0.7. Gated by self-model confidence (v1.6.2): withdraws when organism is confused.

Planning (Vorstellung): Single-step lookahead using self-model. Generates N candidate motor perturbations (±20%), scores each by predicted viability. Satisficing, not optimizing. Currently diagnostic only (does not directly influence motor output).

Concept Formation (Verstand): K=8 competitive prototypes over CTRNN hidden state. Competitive Hebbian learning assigns each moment to the nearest prototype. Each prototype accumulates motor and outcome statistics. The organism categorizes its own experience — Hegel's Understanding as subsumption of particulars under universals.

Two-Agent Interaction

Paired simulation: 50-dimensional ODE (25 per organism). Both organisms share the same physics, same environment. Each perceives the other's motor output through delayed sensory channels (14-16).

dr_ij/dt = α*(dependence_ij - r_ij) + β*r_ji*(1 - r_ij) - decay*r_ij
where dependence = max(viability_dep, clamp(other_pred_error / 3, 0, 1))

Recognition requires perceiving what you cannot predict. The organism's prediction model fails on other-agent channels (other_pred_error > 0), producing dependence, which drives recognition variables r_ab and r_ba toward mutual tracking.

Evolution

Natural selection only. No CMA-ES, no tournament selection, no fitness ranking. Binary survive/die over 100K timesteps. 34 evolvable genes including time constants, learning rates, metabolic parameters. Entire population seeded from best genome (no crossover). Mutation rate 0.03.

Philosophical constraint: No external fitness landscape optimization. The organism does not maximize anything. It maintains itself or dies. Evolution selects for viability, not for any criterion we designed. This is critical — it means every layer transition is discovered by the organism, not designed by us.

Layer Progression: 9 Stages

Each layer has a falsifiable criterion computed from the organism's own behavior. Layers advance automatically when criteria are met. No manual promotion. The organism earns each transition.

1 Metabolic Viability / Sense-Certainty PASS

Hegel: §90-110. Immediate being. The organism simply is.

Criterion: 7/10 recent organisms survive 80K+ timesteps

What passes: The metabolic ODE sustains itself within viability bounds. The 5-chemical reaction network, with evolved kinetic parameters, maintains homeostasis long enough for the system to demonstrate self-maintenance. No learning, no brain — just chemistry.

2 Sensorimotor Coupling / Perception PASS

Hegel: §111-131. Perception: the organism begins to differentiate.

Criterion: 3/5 decisive comparisons: CTRNN ON survives longer than CTRNN OFF

What passes: The 20-neuron CTRNN, reading 16 sensory channels and producing 3 motor outputs, demonstrates measurably better survival than a metabolic-only organism. The brain helps. Sensorimotor coupling is functional.

Ablation needed: CTRNN OFF should fail at Layer 3+ (learning without a brain is impossible).

3 Hebbian Adaptation / Force & Understanding PASS

Hegel: §132-145. The organism discovers forces — lawful structure in the world.

Criterion: 7/10 decisive comparisons: Hebbian ON survives longer than Hebbian OFF

What passes: Neuromodulated Hebbian learning (Oja + BCM + anti-Hebbian, gated by stress × PCN) improves survival vs fixed weights. The organism learns from environmental structure.

Key discovery: Unmodulated Hebbian was net harmful. Neuromodulation (learning only when stressed and surprised) was essential. Also: random environments couldn't support this layer — the environment must have lawful structure (cyclical dynamics, correlated perturbations) for learning to find correlations.

Ablation needed: Structured environment OFF → Layer 3 should fail.

4 Understanding / Laws & Force PASS

Hegel: §132-145. The Understanding discovers that prediction error has two kinds: environmental pattern and self-caused effect.

Criterion: 7/10 organisms with |structural_signal| > 0.1 and sm_active and total_predictions > 1000

What passes: The organism's prediction system detects that residual prediction error (after removing the environmental pattern via slow EMA) is correlated with its own motor output. This is not a magnitude difference — it is a structural difference. The organism discovers that it is a causal agent in its own world.

5 Self-Model / Self-Certainty PASS

Hegel: §163-165. Self-Certainty: the organism knows itself as a cause.

Criterion: mean predictive_accuracy > 0.3 across 5 recent organisms with 5+ intervention tests

What passes: The organism can predict the sensory consequences of freezing its own motor output. Not just "something changes when I stop" (attribution_accuracy, which was trivially 100%) but "I can quantitatively predict how much it changes" (predictive_accuracy). This required the Hebbian-trained self-model W_self to be genuinely informative.

Key failure that drove this: Magnitude-based self-model signals failed — the signal was too noisy. Structural signal (correlation of residual with motor) worked because it detects a different kind of error, not a different amount.

6 Functional Self-Knowledge / Desire PASS

Hegel: §167-170. Desire: self-knowledge becomes practical.

Criterion: mean correction_benefit > 0.1 AND mean ambient_error > 0.05

What passes: The organism inverts its self-model to compute desire_motor — the motor correction that should address its current metabolic deficit. This correction measurably reduces metabolic deficit (>10% per tick). But crucially, ambient_error persists: the world remains fundamentally unpredictable despite self-knowledge.

This tension — desire helps but never suffices — is what drives the need for the Other (Layer 7). The organism masters itself but not its world.

7 Two-Agent Encounter / Recognition PASS

Hegel: §178-184. Encountering another self-consciousness.

Criterion: 3/5 recent paired runs: both survive AND other_error_a > 0.1 AND other_error_b > 0.1

What passes: Two organisms in shared environment both survive and both perceive the other through prediction error on other-agent channels. The organism's self-model fails on channels 14-16 because the other's motor output is unpredictable — encountering genuine otherness.

Technical fixes required: Agents initially never met spatially (identical starting positions with identical weights). Random starting offset + Brownian position drift + increased proximity threshold solved this.

8 Mutual Perception PASS

Hegel: §438 (not full Spirit). Mutual tracking, not genuine dialectical struggle.

Criterion: 3/5 recent paired: r_ab > 0.4 AND r_ba > 0.4 AND other_error_a > 0.5 AND other_error_b > 0.5 AND genome_layer ≥ 6

What passes: Recognition variables reach r_ab = 0.875, r_ba = 0.875 (mutual phase). Both agents perceive the other (other_error ≈ 2.5) but cannot predict it. Recognition is driven by genuine perception, verified by ablation.

Honest limitation: This is mutual perception and tracking, not Spirit in Hegel's full sense. Recognition starts symmetric (no master-slave dialectic). There is no struggle, no asymmetry that resolves. The label "Spirit §438-463" overclaims. Accurate description: mutual tracking between two self-modeling agents who perceive each other's irreducibility.

Ablation result: Perception OFF (other-agent sensory channels zeroed) → r = 0.0. Perception ON → r = 0.87. Minimum other_error for recognition: 1.56. Observed: 2.5. The signal is genuine, not manufactured by parameters.

9 Observing Reason (Representational Ceiling) CEILING

Hegel: §240-262. The organism turns investigative capacity on its world.

Criterion: 3/5 recent organisms: curiosity_drive > 0.1 AND exploration_productivity > 0.05

Result: Does not pass. The two criteria are anti-correlated in the population. No organism passes both simultaneously across 30 organisms and 10 evolution generations.

Group Count Avg Drive Avg Productivity Ambient Error
High ambient error 26/30 -0.187 -0.132 ~3.5
Low ambient error 12/30 mixed mixed ~2.1

Low-error subgroup detail:

SeedDriveProductivityAmbient ErrDrive > 0.1?Prod > 0.05?
795+0.255+0.0032.10YesNo
800+0.036-0.0142.31NoNo
810-0.044+0.0191.97NoNo
817-0.188+0.0802.12NoYes
818-0.305+0.0641.90NoYes
821+0.135+0.0211.97YesNo
Anti-correlation signature: Seeds that explore when confused (drive > 0.1: 795, 821) don't benefit from exploration (productivity ≈ 0). Seeds whose exploration IS productive (productivity > 0.05: 817, 818) don't explore when confused (drive negative). Both halves of the curiosity loop exist in the population — never in the same organism. This is the ceiling.

Ablation Evidence

The most important evidence for structural necessity. Each ablation demonstrates that a lower layer is genuinely required for a higher layer to function.

Completed Ablations

ABLATION Layer 8: Other-Agent Perception ON vs OFF

Method: Analytically compute recognition ODE equilibrium with zero other-agent perception (viability-only dependence) vs normal operation (other_pred_error drives dependence).

Perception OFF: dep ≈ 0 → r_ab = 0.0000, r_ba = 0.0000
Perception ON: dep ≈ 0.84 → r_ab = 0.8713, r_ba = 0.8713
Minimum other_error for r ≥ 0.6: 1.56 (observed: ~2.5)

Conclusion: Recognition requires genuine perception of the other agent. Parameters alone (alpha=0.05, beta=0.02) cannot produce r > 0.6 without the perception signal. Viability-based dependence ≈ 0 across 131 empirical runs confirms this.

ABLATION Layer 9: Confidence-Gating ON vs OFF

Method: Compare curiosity_drive before and after confidence-gating of memory/concepts.

Before (v1.6.1): curiosity_drive consistently negative (-0.52 to -0.08)
After (v1.6.2): low-error organisms show positive drive (+0.04 to +0.25)
But: exploration_productivity remains near zero

Conclusion: Confidence-gating enables the front half of the curiosity loop (confusion drives exploration) but cannot enable the back half (exploration teaches something new) because the world is too simple.

Ablations Needed for Paper

Layer 3: Structured environment OFF → does Layer 3 (Hebbian learning advantage) fail?
Layer 5: Self-model OFF → does Layer 5 (predictive accuracy) fail? (trivially yes, but needs formal run)
Layer 7: Other-agent sensory channels OFF → does Layer 7 (two-agent encounter) fail?
Cross-layer: Run Layer 8 organisms without Layer 6 genome (no desire) → does recognition still emerge?

Key Discoveries

Hebbian Learning Was Net Harmful (Before Neuromodulation)

Unmodulated Hebbian learning caused weight drift and instability, making organisms survive worse than fixed-weight controls. The fix: gate learning by max(stress, surprise) × per-neuron prediction error. Only stressed, surprised, confused neurons update. This converted Hebbian from harmful to beneficial.

Environment Needs Lawful Structure

Random perturbations give the organism nothing to learn from. Hebbian rules find correlations; if the environment has no correlations, learning finds nothing. This principle appears at every scale: Layer 3 needs cyclical resources, Layer 9 needs discoverable structure for curiosity to exploit.

Self-Model Emerges from Prediction Failure

The self-model is not installed. It emerges when the prediction system detects that some errors are structurally different from environmental noise — correlated with motor output after removing the environmental pattern. Magnitude-based signals failed. Structural correlation worked. The organism discovers itself as a distinct causal agent.

Recognition Requires Perceiving What You Cannot Predict

Viability-based dependence (does the other help/hurt me?) was always ≈ 0 with competitive resources. The genuine signal: other_pred_error — the organism's prediction model fails on other-agent channels. It perceives the other but cannot assimilate it into its existing model. This is Hegel's §179: encountering another self-consciousness as irreducible.

Curiosity Requires a Learnable World

The same insight from Layer 3 at a higher level. An organism in a high-noise environment can't distinguish self-caused from world-caused effects, so exploring its motor repertoire produces nothing but more noise. Curiosity isn't just futile — it's actively counterproductive, which is exactly what negative exploration_productivity means.

The Ceiling Finding

Why Layer 9 Fails Honestly

The dialectic stalls when the world cannot sustain further negation. Observing Reason requires inexhaustible structure — the organism turns its investigative capacity on its world and discovers something new about its own categories. But if the categories already capture everything a 20-neuron brain can represent about a 5-chemical world, there is nothing to discover.

The organism's own success prevents its next development. This is not a failure of the methodology — it is a confirmation that the methodology correctly identifies where scale becomes necessary.

Evidence that this is a ceiling, not a tuning problem:

Philosophical Constraints

These constraints were maintained throughout the entire project. They define what makes this a Hegelian system rather than a conventional AI/ML system.

ConstraintWhat It MeansWhy It Matters
No gradient No backpropagation, no loss function, no gradient descent Learning must be local (Hebbian). No global error signal.
No reward No fitness function during life. Only: survive or die. Every capability emerges from survival pressure, not from optimizing a designed objective.
Natural selection only No CMA-ES, tournament selection, or fitness ranking Evolution selects for viability, not for any criterion we impose. The organism discovers layers, not us.
One World Principle Same physics for ALL organisms No privileged agents. Same ODE, same environment, same constraints.
Self-model not installed Self-model emerges from prediction system's own failure The organism discovers itself; we don't tell it to look.
No pre-designed language No token systems, no signaling channels Spirit (shared meaning) must emerge from shared life, not shared design.
Honest failure is valuable Layer 9 failing is documented, not fixed by parameter tuning The ceiling tells us where scale is needed. Tuning it away would hide the real boundary.

v2 Requirements (From the Ceiling)

The Layer 9 ceiling doesn't just say "scale up." It specifies exactly what needs to scale and why:

Richer environment: More chemical species, spatial structure, temporal patterns with longer correlations. The organism needs discoverable structure that 8 concept prototypes can't exhaust. Currently: 5 chemicals, 3 environment channels, cyclical dynamics. Need: enough complexity that a self-model always has more to learn.

Larger CTRNN: 50-100 neurons. More representational capacity means more possible self-model states, which means more territory for curiosity to explore. The current 20 neurons hit their information ceiling.

Population dynamics beyond pairs: N-agent interactions where social structure (norms, roles, hierarchies) can emerge. This is the path to genuine Spirit — not two agents tracking each other, but a community with shared behavioral patterns.

Asymmetric encounters: Agents with different evolutionary histories meeting. This is the path to the master-slave dialectic — not symmetric recognition from identical starting points, but asymmetric struggle that resolves into new forms.

Deep Research Prompt

The following prompt is designed for Claude Deep Research. It contains the full project description, all results, and specific technical questions about how to break through the Layer 9 ceiling — environment scaling, network architecture, self-model improvements, multi-agent dynamics, curiosity mechanisms, computational strategy, and what the literature says about this specific kind of representational limit.

I have built a computational system called "Hegel" that implements Hegel's developmental logic from the Phenomenology of Spirit as a gradient-free dynamical system. I need a deep research analysis of this work in context of existing literature, assessment of the methodology, and identification of what's genuinely novel vs. what's been done before.

## System Summary

The system is a 25-dimensional continuous ODE (solved by adaptive Runge-Kutta) representing an artificial organism. State vector: 5 metabolic variables (substrate, catalysts, enzymes) + 20 CTRNN neurons. The organism has 16 sensory inputs (metabolic state, affective derivatives, environment, other-agent motor output) and 3 motor outputs. It dies if any metabolic variable exits a viability boundary — there is no reward signal, no fitness function, no loss function.

Learning is purely Hebbian (Oja + BCM + anti-Hebbian rules), gated by neuromodulation: gate = max(stress, surprise) * per_neuron_prediction_error. No gradient descent, no backpropagation. Evolution uses natural selection only (binary survive/die, no CMA-ES, no tournament, no fitness ranking). 34 evolvable genes. Entire population seeded from best genome.

## The 9 Layers (Developmental Stages)

Each layer has a falsifiable criterion computed from the organism's behavior. Layers advance automatically.

**Layer 1 — Metabolic Viability (Sense-Certainty §90-110)**: 7/10 organisms survive 80K+ timesteps. The metabolic reaction network sustains itself. PASS.

**Layer 2 — Sensorimotor Coupling (Perception §111-131)**: CTRNN ON survives better than CTRNN OFF in 3/5 decisive comparisons. The brain helps. PASS.

**Layer 3 — Hebbian Adaptation (Force & Understanding §132-145)**: Hebbian ON survives better than OFF in 7/10 decisive comparisons. Key discovery: unmodulated Hebbian was NET HARMFUL. Neuromodulation was required. Also: random environments couldn't support this — the environment needs lawful structure. PASS.

**Layer 4 — Understanding (§132-145)**: Structural signal > 0.1 — the organism's prediction system detects that residual error (after removing environmental pattern) correlates with motor output. It discovers itself as a causal agent. Not magnitude-based (failed) but structural correlation (worked). PASS.

**Layer 5 — Self-Model (Self-Certainty §163-165)**: Predictive accuracy > 0.3 in intervention tests. The organism can quantitatively predict how much sensory input changes when its motor is frozen. Uses a Hebbian-trained linear self-model W_self. PASS.

**Layer 6 — Functional Self-Knowledge (Desire §167-170)**: correction_benefit > 0.1 AND ambient_error persists. The organism inverts its self-model to compute motor corrections for metabolic deficits. Self-knowledge becomes practical (desire), but the world remains unpredictable (ambient error). This tension drives the need for the Other. PASS.

**Layer 7 — Two-Agent Encounter (Recognition §178-184)**: 50-dim ODE (25 per organism). Both survive AND both show other_pred_error > 0.1. The organism perceives the other through prediction error on other-agent sensory channels — encountering what it cannot assimilate. PASS.

**Layer 8 — Mutual Perception (§438)**: Recognition variables r_ab > 0.4, r_ba > 0.4, other_error > 0.5. Recognition driven by other_pred_error: dependence = max(viability_dep, clamp(other_error/3, 0, 1)). Achieved r_ab = 0.875. Ablation: perception OFF → r = 0.0. Honest limitation: symmetric start, no master-slave dialectic. This is mutual tracking, not Spirit in Hegel's full sense. PASS.

**Layer 9 — Observing Reason (§240-262)**: curiosity_drive > 0.1 AND exploration_productivity > 0.05. After implementing confidence-gated exploration (memory/concept influence scales with self-model confidence), the two criteria are ANTI-CORRELATED across 30 organisms and 10 evolution generations. Organisms that explore when confused find nothing new. Organisms whose exploration is productive don't explore when confused. The 20-neuron, 5-chemical world is too simple for curiosity. CEILING — does not pass.

## Key Technical Details

- Prediction error partitioning: motor_error vs ambient_error, separated by motor_effect_countdown
- Structural self-model signal: corr(residual_error, motor_deviation) where residual = error - EMA(error)
- Self-model prediction loop: selfmodel_predict(motor) subtracted from sensory before prediction_step
- Per-neuron prediction error (PCN, Rao & Ballard 1999): only confused neurons learn
- Competitive resources: motor output determines resource capture share in two-agent setting
- Episodic memory: ring buffer + cosine retrieval, gated by sm_confidence
- Concept formation: K=8 competitive prototypes (Kohonen-style) over CTRNN hidden state
- Confidence-gated exploration: sm_confidence = clamp(1 - pred_error/5, 0, 1) scales memory/concept influence
- Recognition dynamics: dr_ij/dt = alpha*(dep_ij - r_ij) + beta*r_ji*(1-r_ij) - decay*r_ij
- Voice layer: 4-dim expressive output, Oja learning gated by stress + self-model signal

## The Ceiling Finding (Most Important Result)

Layer 9 failing is the most important result. The anti-correlation of curiosity criteria is the precise signature of a representational ceiling. The organism has learned everything this world has to offer. The dialectic stalls when the world cannot sustain further negation. This finding specifies exactly what v2 scaling must provide: richer environment, larger networks, population dynamics.

## Hardware Constraints

The system runs on a single VPS: AMD EPYC 8-core (no SMT), 24 GB RAM, no GPU, Julia 1.10.8 with 7 threads. A single 25-dim ODE solve takes ~30 seconds for 100K timesteps. Paired 50-dim simulation: ~2-3 minutes. Evolution generation (20 organisms in parallel): ~3 minutes. All computation is CPU-bound ODE solving — no matrix multiplications large enough to benefit from GPU.

## What I Need: How to Break Through the Layer 9 Ceiling

The core problem is clear: the 20-neuron, 5-chemical world is too simple for curiosity-driven exploration to be productive. The organism has learned everything this world has to offer. I need concrete, technically specific recommendations for v2 that will make Layer 9 (Observing Reason) achievable and open the path beyond it.

### 1. Environment Scaling — What Exactly Should Change?

The current environment has 5 chemicals with mass-action kinetics, 3 environmental sensory channels (resource gradients), and cyclical dynamics with correlated perturbation cascades. The organism exhausts this with 8 concept prototypes.

I need specific recommendations:
- How many chemical species / reaction pathways would create enough discoverable structure? What reaction network topologies (autocatalytic sets, hypercycles, etc.) create the right kind of complexity — learnable but inexhaustible?
- Should I add spatial structure beyond the current 1D position ring? 2D grid? Continuous 2D with diffusion? What does the ALife literature say about minimum environment complexity for open-ended learning?
- What temporal structures (beyond current cyclical dynamics) would give the self-model more to discover? Multi-scale dynamics? Regime changes? Hidden variables the organism must infer?
- Are there established benchmarks in artificial life for "environment complexity sufficient for open-ended learning"? What do Avida, Tierra, Lenia, or other ALife platforms use?

### 2. Network Scaling — How Big, and What Architecture?

Current: 20 CTRNN neurons, fully connected (20×20 W, 20×16 W_in, 3×20 W_out). Hebbian learning (Oja + BCM + anti-Hebbian).

- What is the minimum network size where curiosity-driven exploration becomes productive in similar dynamical systems? Is there empirical or theoretical guidance?
- Should I keep full connectivity or move to sparse / modular connectivity? What does the computational neuroscience literature say about the relationship between network topology and learning capacity in Hebbian networks?
- Are there better Hebbian variants I should consider for larger networks? STDP? Predictive Hebbian learning? Equilibrium propagation? What stays gradient-free but scales better?
- Should the CTRNN architecture change? Multiple time-scale networks (fast + slow)? Reservoir computing with a Hebbian readout? What architectures support richer self-modeling?

### 3. The Self-Model Bottleneck

The current self-model is a linear map W_self (16×3): sensory_change ≈ W_self × motor_output. This is powerful enough for the current world but may be the real bottleneck for curiosity — a linear model can't discover nonlinear structure in its own behavior.

- Should the self-model be nonlinear? A small Hebbian-trained network predicting sensory consequences? How do you train a nonlinear self-model without gradients?
- In the predictive coding / active inference literature, how are hierarchical generative models learned without backpropagation? Can predictive coding networks (Rao & Ballard) be extended to self-modeling?
- What is the minimum self-model complexity that would make curiosity productive — i.e., where novel motor actions would reveal self-model errors that the self-model could then learn from?

### 4. Multi-Agent Scaling — Path to Spirit

Current Layer 8 achieves mutual perception between 2 agents but with symmetric start (no master-slave dialectic). For genuine Hegelian Spirit, I need:

- N-agent populations where social norms can emerge. What is the minimum population size for emergent social structure in agent-based models? What does the literature say (Lux-Marchesi, Sugarscape, etc.)?
- Asymmetric encounters: agents with different evolutionary histories meeting. How to implement this without making the simulation computationally intractable? Can I evolve subpopulations separately then merge?
- Communication: the organism has a 4-dim voice layer (Oja-learned). What would it take for this to become a genuine signaling system? What does the language evolution literature (Kirby, Steels) say about minimum conditions?
- Should the paired ODE become N-agent? Or should I use asynchronous agent-based simulation with ODE organisms? What are the computational tradeoffs?

### 5. Curiosity Architecture — What Works Without Gradients?

Current curiosity signals:
- curiosity_drive = corr(sm_error_t, motor_variance_{t+1}): does confusion drive exploration?
- exploration_productivity = corr(motor_novelty_t, sm_error_reduction_{t+1}): do novel actions reduce error?

The confidence-gating (memory/concepts withdraw when confused) helps curiosity_drive but not productivity.

- In the intrinsic motivation / curiosity-driven learning literature (Schmidhuber, Oudeyer, Pathak), what curiosity architectures work for embodied agents without gradient signals? Can "learning progress" or "prediction gain" metrics be computed from Hebbian statistics?
- Should exploration be directed rather than random? The current planning module generates random motor perturbations. Could the self-model guide exploration toward regions of high uncertainty? How to do this without gradients?
- Is there a way to measure "information gain" from motor actions using only local Hebbian statistics? Something like: "this action changed my weights more than average" as a proxy for productive exploration?
- The exploration_productivity metric measures single-timestep error reduction. Should the measurement window be longer (10-50 timesteps)? What's the right timescale for Hebbian learning to integrate a novel experience?

### 6. Computational Scaling Strategy

Running on 8-core CPU, 24 GB RAM, no GPU. Julia ODE solver.

- If I scale to 100 neurons, the ODE becomes 105-dim. How much slower is Tsit5 at this scale? Should I switch to a different solver? Implicit methods?
- For N-agent simulations (10-50 agents), what's the most efficient approach in Julia? Can I use GPU-accelerated ODE solving (DiffEqGPU.jl)?
- What's the computational cost curve? Can I run v2 on the same VPS or do I need a GPU server? What cloud compute would be most cost-effective?
- Are there approximation techniques (e.g., mean-field for multi-agent, sparse connectivity for large networks) that would preserve the essential dynamics while reducing compute?

### 7. What Does the Literature Say About This Specific Ceiling?

The anti-correlation finding — organisms that explore when confused find nothing new, organisms whose exploration is productive don't need to explore — feels like it should have a name in the learning theory literature.

- Is this a known phenomenon? Does it correspond to the "exploration-exploitation" dilemma in a specific formal sense? Is there a theorem about minimum environment complexity for curiosity-driven learning to be productive?
- In the developmental robotics literature (Lungarella, Oudeyer, Kaplan), have similar ceilings been documented? What did they do to break through?
- Does the "Hegelian" framing (the dialectic stalls when the world can't sustain further negation) map onto any formal concept in complexity theory or dynamical systems theory?

### 8. Beyond Layer 9 — What Comes Next?

If Layer 9 passes in v2, the Phenomenology continues:
- Active Reason (§253+): organism intervenes in the world to test understanding
- Spirit (§438+): shared norms from mutual recognition in populations
- Religion / Absolute Knowing (§672+): self-model becomes self-referential (meta-cognition)

- What are the computational prerequisites for each? Which is achievable at what scale?
- Has anyone implemented meta-cognition (model of the modeling process) in a gradient-free system?
- What does "self-referential self-model" look like computationally? A self-model that predicts its own learning rate? Its own prediction error dynamics?

### 9. Publication and Framing

- Where should this work be published? ALife journal? Adaptive Behavior? Frontiers in Computational Neuroscience? A philosophy venue? Multiple papers (methodology paper + results paper + philosophical paper)?
- How should the ceiling finding be framed — as a negative result, a scaling law, or a design specification?
- What additional ablations and controls would peer reviewers demand?
- Is the Hegelian framing a strength (novel philosophical grounding) or a liability (unfamiliar to CS reviewers)? How to present it to different audiences?
How to use: Copy the prompt above and paste it into Claude Deep Research. It contains the complete system description, all 9 layer results with exact criteria, the ceiling finding with data, hardware constraints, and 9 detailed research questions focused on how to improve and scale the project — not just evaluate it.