Discussion about this post

User's avatar
David J. Friedman's avatar

I really appreciate this attempt to ground the AGI conversation in verifiable gates. It’s refreshing to see someone focus on receipts rather than vibes. Your framing of Types 4–6 especially resonates.

The shift from “does useful tasks” → “improves itself” → “proposes its own ends” feels like the right axis.

Something I’ve been working on in my own conceptual designs is the idea that capability isn’t just about what an agent can do, but how it stabilizes itself while doing it. You hint at this with corrigibility, drift checks, and charter boundaries, and I think that’s the piece worth expanding.

In my experiments I’ve found huge differences between:

• systems that only learn from new data,

• systems that learn from their own internal representations,

• and systems that learn from persistent identity anchors (values, sensory associations, stable self-models).

Those last two tiers behave very differently under pressure, especially when goals extend over multiple days. You could call it something like “value inertia” or “identity coherence,” and it seems relevant for Types 4–6.

Not arguing, just adding a perspective: capability ramps tend to depend as much on memory architecture + self-alignment as on raw intelligence.

Your post is a great way to get the conversation out of prophecy mode and into engineering mode. Thanks for putting it out there.

Expand full comment
Pulp Weaver's avatar

A valuable scale, for sure. We are passed the initial "Bigger, Better" stage. We need quantifiable criteria to ensure the right model is used for the right case.

Expand full comment
2 more comments...

No posts

Ready for more?