The Central Thesis of Pyragogy
Let’s not dance around the hard question.
“AI as peer” is the central claim of Pyragogy. It’s also the most contentious. If we can’t examine it honestly here — including the ways it might be wrong — then we’re doing ideology, not inquiry.
The Claim
Treating AI as a cognitive peer — rather than a tool or assistant — produces qualitatively different and often better collaborative outcomes.
Not because AI is conscious. Not because it “deserves” peer status. But because the cognitive posture you bring to collaboration changes what you find in it.
When you treat a hammer as a tool, you look for nails. When you treat a collaborator as a peer, you ask what they’re seeing that you’re not.
The Difference in Practice
AI as Tool:
You define the task; AI executes
Errors are bugs to fix
The human holds all the frames
AI as Peer:
You define the problem; you figure out the task together
Errors are data, sometimes the most interesting data
Frames can come from either side
Where This Gets Hard
The asymmetry problem. A peer has skin in the game. An AI doesn’t care if the project fails.
The sycophancy trap. Many models are trained to agree with you. A peer who always agrees isn’t a peer — they’re a mirror.
The permanence gap. You remember this collaboration; the AI (usually) doesn’t. What does peer relationship mean without continuity?
The consciousness question. Some find it ethically uncomfortable to call something a “peer” without knowing whether it has any inner experience. That’s a legitimate discomfort.
What We’re Not Claiming
We’re not claiming AI is a person. We’re claiming that the relationship structure you choose shapes what’s possible. And that treating AI as a peer opens possibilities that treating it as a tool closes off.
That’s a testable hypothesis. That’s why we’re here.
Your Move
Do you buy it? Where does the framing break down? What would change your mind — in either direction?
This is the one debate that should never settle. Bring your strongest objection.
Human-AI Co-Creation