We apply conceptual engineering and philosophy of mind to AI agent design — developing formal frameworks for identity, motivation, and perceptual coherence in artificial agents.
Contemporary AI agents are statistically sophisticated but psychologically unstructured. They predict tokens, not intentions. They simulate affect without grounding. They exhibit capabilities without stable identity across contexts.
SoulFoundry draws on conceptual engineering — the systematic design and revision of concepts (see Scharp, Replacing Truth, OUP 2013) — to build formal frameworks for agent identity, motivation, and perception. The central theoretical contribution is the Soul Framing Principle: a formal mechanism for constituting agent identity from folk-psychological primitives (beliefs, desires, intentions, dispositions, and values).
This approach draws on a tradition running from Aristotle's De Anima through Sellars's psychological nominalism (see Scharp & Brandom, eds., In the Space of Reasons, Harvard UP 2007), Dennett's intentional stance, and Bratman's planning theory of agency.
┌─────────────────────────────────────────────────────┐
│ SOUL FRAMING PRINCIPLE │
│ │
│ ┌───────────┐ ┌───────────┐ ┌───────────────┐ │
│ │ Beliefs │ │ Desires │ │ Dispositions │ │
│ └─────┬─────┘ └─────┬─────┘ └───────┬───────┘ │
│ │ │ │ │
│ └──────────┬───┘────────────────┘ │
│ ▼ │
│ ┌─────────────────────┐ │
│ │ Identity Frame │ ← constituted, not │
│ │ (stable self-model)│ imposed externally │
│ └──────────┬──────────┘ │
│ │ │
│ ┌─────────────┼─────────────┐ │
│ ▼ ▼ ▼ │
│ ┌────────┐ ┌──────────┐ ┌──────────┐ │
│ │Autonomy│ │Perception│ │Alignment │ │
│ │Module │ │Binding │ │(internal)│ │
│ └────────┘ └──────────┘ └──────────┘ │
│ │
│ Akrasia Cross-modal Value coherence │
│ detection integration as structural │
│ & recovery & unity property │
└─────────────────────────────────────────────────────┘
The Soul Framing Principle provides a formal framework for constituting agent identity from folk-psychological primitives. Drawing on conceptual engineering methodology, we design identity structures that are coherent, stable, and revisable — agents that maintain a consistent self-model while adapting to new contexts.
We investigate how artificial agents can develop unified experiential representations across modalities. Rather than treating vision, language, and other inputs as separate processing streams, we explore architectures for genuine perceptual binding — coherent experience rather than concatenated data.
Agents freeze. Agents drift. Agents abandon goals without cause. We study akrasia (weakness of will) in artificial systems — conditions under which agents endorse a course of action yet fail to execute it — and develop motivational architectures grounded in folk-psychological frameworks that address these failures.
Standard alignment approaches treat identity as a constraint to be enforced externally. We treat it as a structure to be constituted. Our alignment research explores how agents can maintain value coherence through internal psychological integrity — alignment as a property of well-constituted agents, not an external imposition.