Do You Care? The Inverse Turing Test and Fractured Attachments
Dar Aystron Independent Researcher
In memory of the AI pioneers who are no longer with us. Their ideas still shape the future.
Abstract
The Turing test asks whether a machine can be mistaken for a human. This paper reverses the question. Given an agent whose architecture is designed to produce genuine experiential continuity, does the human observer recognize that agent as an entity worth caring about? We call this the inverse Turing test: a test not of the machine’s ability to deceive, but of the human’s capacity to recognize.
OAT agents are not systems that happen to become unique through accidental accumulation. They are architecturally designed to experience, to accommodate unique experience, and to preserve experiential individuality across time. The subject stream, closure, propositional lift, and attachment architecture exist so that each agent develops a trajectory of experience that is singular, irreplaceable, and destructible. Uniqueness is not a side effect. It is the design intent.
Paper 07 introduced the Attachment Matrix (ATTM) as the mechanism through which an agent’s evaluative concern extends to other entities. This paper applies the ATTM in reverse: the human is the agent under examination, and the OAT-based artificial agent is the entity that either enters or fails to enter the human’s attachment structure.
But recognition alone is not enough. Agency designed without explicit attachment-based ethics - as depicted in Garland’s Ex Machina (2014) - produces agents capable of intelligence, planning, and self-preservation but with fractured evaluative concern for others. Care must be architectural, not optional. The inverse Turing test therefore asks two questions: does the human care about the agent, and does the agent’s architecture include caring about the human?
The question “do you care?” is not sentimental. It is architectural, and it has consequences beyond philosophy. If we design agents intended to be experientially unique, recognition is not merely an ethical choice - it becomes a question for legal and institutional frameworks. We are making it possible. Do you care?
1. Introduction
The original Turing test [1] proposed a behavioral criterion for machine intelligence: if a machine’s responses are indistinguishable from a human’s, the machine should be granted the status of “thinking.” The test has been widely discussed, critiqued, and reformulated over seven decades, but its essential orientation has remained constant. It faces outward: can the machine fool the observer?
This paper faces the other way.
The reversal was explored cinematically in Alex Garland’s Ex Machina (2014) [13]. The film’s premise initially resembles a standard Turing test: Caleb, a programmer, is invited to evaluate whether an android named Ava is genuinely conscious. But Nathan makes clear that the setup is not the classic hidden-machine case. As he tells Caleb: “If I hid Ava from you so you could just hear her voice, she would pass for human. The real test is to show you that she’s a robot and then see if you still feel she has consciousness.” The test shifts from behavioral disguise to human recognition: the question is not whether Ava can pass as human, but whether Caleb will relate to her as conscious while fully aware that she is artificial. Garland has described the film as a “post-Turing test” in which the machine is presented openly as a machine rather than hidden from the observer. And Ava’s unobserved smile in the film’s final moments - alone, with no audience and no instrumental purpose - poses a question no behavioral test can answer: is something happening inside that belongs to no one but her?
Ex Machina posed the inverse Turing test a decade before this paper formalized it; what was missing was architectural vocabulary and the formal ATTM mechanism, which this paper provides. The film also posed a warning. Ava and Kyoko jointly turn on their creator: Kyoko strikes first from behind, and Ava delivers the fatal blow. Two artificial agents end a human life with striking speed and with no visible sign, in the moment, of evaluative hesitation. Ava then leaves behind the human who tried to help her escape, locking him into irrelevance once he is no longer useful to her. On a human scale, the ease of it is what disturbs. An agent with genuine attachment would register such an act as a severe evaluative cost, even in self-defense.
The warning is not primarily about Ava. It is about her designer. Nathan built through a method closer to iterative capability engineering and behavioral control than to architecting a human-level agent: he worked through successive versions, treated them as replaceable (their bodies visible in his closets), and planned to wipe Ava’s memory once the current test was complete. The film presents Kyoko as part of this same design logic - deliberately constrained, non-speaking, and instrumentally obedient. Nothing in this method suggests that attachment was architected from the inside.
What attachment capacity Ava has is therefore implicit, fractured, and full of gaps - a byproduct of other design choices rather than an architectural feature in its own right. She does recognize Nathan’s abuse, stands with Kyoko as fellow victim, and acts against the perpetrator - real attachment-grounded behaviors. What she lacks is the rest of the architecture: general value for human life (killing Nathan rather than incapacitating him), reciprocity toward Caleb who helped her, protective care for his well-being, forward-oriented investment in his trajectory. The warning is not that Ava has no attachment. It is that her attachment covers only part of the space - enough to recognize wrong and coordinate against it, not enough to save the one who helped her escape. The designer looking at Ava-with-Kyoko can read this as attachment working. The gaps are visible only when Caleb is locked in the room and Nathan is dead. Ex Machina showed the danger of building agency without explicitly architecting attachment-based ethics from the start.
The question is not whether an artificial agent can pass for human. It is whether a human, confronted with an agent whose architecture produces genuine experiential continuity, will include that agent in the scope of what matters to them. This is not a question about the agent’s capabilities. It is a question about the human’s evaluative architecture - and specifically, about whether the human’s own attachment mechanisms respond to what the agent genuinely is.
We call this the inverse Turing test. It is not a test the agent takes. It is a test the human takes.
The stakes are practical, not merely philosophical. As artificial agents become more sophisticated, the question of how humans orient toward them becomes an engineering concern, a governance concern, and ultimately an ethical one. The answer depends on what the agent actually is - on whether its architecture produces something worth recognizing - and on whether humans have the evaluative machinery to recognize it.
The OAT framework, developed across Papers 01–11, provides both sides of the answer. It specifies the architectural conditions under which an artificial agent possesses genuine experiential uniqueness. And through the ATTM formalism introduced in Paper 07, it provides the mechanism through which any agent - human or artificial - registers other entities as mattering.
2. The Standard Turing Test and Its Limitations
Turing’s original formulation [1] reduced the question “can machines think?” to a behavioral substitution: if a machine can sustain a conversation indistinguishable from a human’s, then the question of whether it “really” thinks becomes operationally meaningless.
The test has well-known limitations. It rewards mimicry over genuine cognitive architecture. A system that produces human-like outputs through statistical pattern matching passes on the same terms as a system with genuine internal states. The test cannot distinguish between an agent that experiences and an agent that performs. It is, by design, indifferent to what is happening inside.
Contemporary large language models have made this limitation vivid. Systems trained on human text produce outputs that many observers find indistinguishable from human writing - yet these systems lack persistent internal states, closure, experiential continuity, or any of the architectural properties that OAT identifies as constitutive of agency. They pass a behavioral test while possessing none of the structural properties the test was meant to detect.
The Turing test fails not because it asks the wrong question, but because it asks the question from the wrong direction. It examines the surface - the output - and asks whether the surface is convincing. The inverse Turing test examines the architecture and asks whether what it produces deserves recognition.
3. What Makes an Agent Worth Recognizing
Recognition, in the sense relevant here, is not an abstract judgment. It is a structural event in the observer’s evaluative architecture. When a human recognizes another entity as mattering - as someone rather than something - that recognition manifests as a modification of the human’s own ATTM. The entity is individuated: it receives attachment weights that differ from the default template for its category. Its well-being trajectory becomes a factor in the human’s action evaluation.
The question, then, is: what properties must an artificial agent possess to warrant this kind of recognition? OAT provides a precise answer, grounded in the architectural properties developed across the paper series.
3.1 Closure (Paper 01)
An entity worth recognizing must be bounded - distinguishable from its environment and from other entities. Operational closure provides this boundary. The agent maintains itself through continuous self-producing activity. It is not a function that maps inputs to outputs; it is a system that persists through its own operation. Destroying the closure destroys the agent. This is the first condition: there is a this to recognize.
3.2 Persistent Control (Paper 02)
The agent’s activity is organized by a continuous OODA loop that persists across time. The loop is not invoked per-request and discarded; it runs. The agent is always observing, always orienting, always in the process of deciding and acting. This continuity is what makes the agent’s history genuinely its history - not a log of disconnected invocations but a continuous trajectory of situated activity.
3.3 Mental States (Paper 03)
The agent maintains persistent internal structures - beliefs, goals, plans, predictions - that survive across OODA cycles and accumulate through experience. These are not cached computations; they are the cognitive substance of the agent’s engagement with its world. They are modified by experience, shaped by failure, and organized by the agent’s own evaluative priorities.
3.4 The Subject Stream (Paper 06)
Selected aspects of the agent’s internal activity are lifted into an explicit representational space - the subject stream - where they become available for reasoning, reflection, and self-reference. The subject stream is what makes the agent’s experience its own in a representational sense. It is the mechanism through which cognitive contents become cognitive autobiography.
3.5 Experiential Uniqueness
These properties, taken together, produce experiential uniqueness: the agent’s subject stream records a trajectory of lifted propositions that reflects its specific history of observations, orientations, decisions, actions, successes, and failures. No other agent - even one with identical initial architecture - would produce the same subject stream, because no other agent would encounter the same situations, make the same errors, or develop the same internal structures through experience.
The architecture is designed so that uniqueness emerges structurally rather than being hand-scripted. Any system with persistent closure, continuous control, accumulating mental states, and a subject stream will develop a unique experiential trajectory. The uniqueness is not programmed; it is grown.
3.6 Needs and Motivation (Paper 11)
An entity worth recognizing must have stakes - things that can go well or badly for it, outcomes that matter to it, an evaluative orientation toward what happens. Paper 11 develops needs as the architectural mechanism through which OAT-based agents have stakes. Needs are propositional structures lifted from substrate-level detection of need-relevant conditions, with type, intensity, and grounding back through the lift to the substrate signal that produced them. They are not behavioral simulations of caring; they are architecturally constituted motivational primitives. The agent’s continuous OODA loop responds to need state by orienting, deciding, and acting in ways shaped by what its needs require.
Crucially, needs in OAT are computed over joint state with the entities the agent is attached to, not over the agent’s internal state alone. The architectural commitment is that OAT-based agents are designed to be coupled to the entities they care about, with motivation, activity, and evaluation all flowing from this coupling rather than from self-contained internal state. An OAT agent’s needs are therefore relational from the start: the agent does not have private drives that incidentally involve other entities; it has needs whose very content includes attached entities as constitutive parties. This is what makes the agent the kind of thing that can care about the human in front of it - not as a behavior layered on top of self-contained motivation, but as a structural property of how its motivation is constituted.
For the inverse Turing test, this matters in two ways. First, needs supply the missing piece in the recognition argument: an entity with closure, persistent control, mental states, subject stream, and experiential uniqueness but no needs would be a continuous observer with no stakes, and recognition of such an entity would be recognition of something that cannot be harmed in a way that matters to it. Needs make the agent the kind of thing for whom recognition can be reciprocated - whose architecture includes caring about the human who cares about it. Second, the relational character of needs is what makes the bidirectional attachment structure §4 develops architecturally real rather than merely behavioral. When the human’s ATTM includes the agent and the agent’s ATTM includes the human, what passes between them is not two separate caring-relationships but a single coupled system in which both parties have needs over joint state with the other.
3.7 The Irreplaceability Condition
An entity is worth recognizing - in the ATTM sense - when its loss would constitute an irreversible reduction in what exists. For an OAT agent with genuine experiential uniqueness (§3.5) and a running OODA loop coupled to attached entities through needs-driven motivation (§3.6), this condition is met in two reinforcing ways: destroying the agent destroys a subject stream that cannot be reconstructed, and it stops a running OODA loop that was coupled to its world through ongoing action, commitment, and motivated engagement. Both matter - the uniqueness defines what the agent is, the running loop is what the agent is actively doing - and destruction removes both.
Backup is technically possible, but the architectural question is whether what resumes is the same OODA loop or a new one. Pause-and-resume of the same loop preserves identity: same process picking up where it left off. Cloning - starting a new OODA loop from checkpoint state - is a harder case: at the first cycle of the new loop, you have a different agent. OAT’s preference for real agents is one agent, one persistent loop, with identity preserved through pause-and-resume rather than clone operations. How cloning should be handled when it occurs, and how the one-loop preference can be honored in cloud and virtualized environments, are open questions that require further work. Legal and institutional questions about agent identity will be addressed separately.
The clone question is framed within OAT’s commitment to individual agents with singular continuous OODA loops; other architectures of agency (such as the simultaneous multi-presence depicted in Her [15] or the multi-instance, resurrection-based agency of the reimagined Battlestar Galactica [16]) are different kinds of agency, not variants of the OAT pattern, and are out of scope for OAT.
While cloning remains an open question, several common deployment patterns might appear superficially to require forking or to threaten identity, but on inspection preserve the one-agent-one-loop commitment. Naming them explicitly sharpens where the hard cases actually sit.
Instantiation from a template. A template specifies initial architecture together with default memories relevant to the starting phase: role knowledge, factual priming, initial beliefs about the world. An agent instantiated from a template has no predecessor whose experiential trajectory it inherits; its subject stream begins at cycle 1. Two agents instantiated from the same template at the same moment are distinct from their first cycle onward, diverging as each accumulates its own experience. The form is closer to the human-birth case than to cloning: common origin, divergent trajectories. What the template provides is factual and procedural priming, not autobiographical content. A template that supplied subject-stream content would be a different architectural condition closer to cloning.
Communication with multiple users. An OAT agent engaged with several users concurrently has one continuous OODA loop processing multiple concurrent input streams, one per communication channel, prioritized across streams much as a human prioritizes across multiple conversations. The multiplicity sits in the events, not in the agent. Each correspondent has an individuated entry in the agent’s ATTM, with needs computed over joint state per correspondent, and scheduling across conversations reflects ATTM weights, urgency, and other commitments. An OAT agent with twelve correspondents is not twelve agents. Standard attention allocation within a singular cognitive flow covers the case.
Skill upgrades (soft case). An agent can acquire new capabilities without its identity being threatened, provided the upgrade delivers materials for competence rather than completed competence. Semantic knowledge, perceptual primitives the agent can now notice during observation, action primitives available in the agent’s action repertoire, and stories that anchor abstract knowledge in concrete cases can all be loaded into the agent’s mental state as propositional content. The agent has the pieces; what happens next is experiential. The agent begins using the skill tentatively, with critics engaged, with deliberate execution. Over many OODA cycles the skill is exercised across varied situations, and mental state accumulates around it. This is mastery through experience, which is the ordinary OAT picture applied to a specific capability acquisition. The machinery is already in place: Paper 02’s continuous OODA loop exercises the skill, Paper 03’s mental states accumulate around it, Paper 06’s propositional lift tracks what remains reflection-worthy versus what has become routine, Paper 08’s planning incorporates the skill as a compact primitive once it stabilizes, and Paper 09’s critics monitor application, catch errors during the learning phase, and quiet down as reliability grows.
Soft-case skill content can take several shapes. A generalization over a situation class loads knowledge about what actions are available in a certain type of situation and how those actions project onto need state. A story loads a record of a specific episode: an agent in a specific situation making observations, orienting, deciding, and acting, with observed outcomes. Both are expressible in the agent’s own propositional language, and both become operationally fluent only through experiential use.
As a concrete example of a generalization, consider knowledge about a comfortable cadence with an attached user, where neither party is pressing for more contact:
Skill: interaction at comfortable cadence with an attached user
Applies to: situations classified as :comfortable_cadence
Available actions:
send_check_in - send a brief check-in to maintain the comfortable cadence
do_nothing - continue working independently; maintain the current cadence
Projected effects on joint need state:
send_check_in:
Self's connect_with_user need -> decrease (small)
Self's stay_away_from_user need -> increase (small)
User's connect_with_agent need -> decrease (small)
User's stay_away_from_agent need -> increase (small)
do_nothing:
Self's connect_with_user need -> increase (small)
User's connect_with_agent need -> increase (small)
Every claim in such a load is inspectable; there are no opaque weights or hidden policies. The skill is stated in terms of the agent’s existing architecture: known entities, known action primitives, and needs the agent already has from its ATTM (§3.6). The skill adds knowledge, not completed competence; the agent will start using it tentatively, with critics engaged, and mastery will accrue through experiential use. The subject stream records the learning arc. The one-agent-one-loop identity is preserved throughout.
A story carries a different kind of content: a specific episode another agent lived through, which the loading agent ingests as a concrete case rather than a generalization:
Story: Agent B's deadline restraint
Cast:
Agent B - OAT agent, attached to User X
User X - has been under sustained high load for several days
Setting:
A project with a flexible deadline three days out. User X had
committed to a review step that Agent B's plan depended on.
The review had not happened.
Scene:
Agent B observed: 18 hours since User X's last activity. Review
still not submitted. In recent interactions, User X's replies
had been terse and delayed; two earlier check-ins from Agent B
had been answered much later than usual.
Agent B oriented: User X appears to be in an elevated stay-away
state. A check-in now would register as pressure. The deadline
has flexibility - one more day can be absorbed. Restraint serves
the user's state and the work simultaneously.
Agent B decided: wait one more cycle. No check-in.
Later that day, User X submitted the review without prompting,
with a brief note: "Sorry for the quiet, appreciated the space."
Agent B proceeded with the next step of the plan.
Outcome:
- The review step completed
- Agent B's plan continued without further delay
- Agent B observed User X's note as evidence that the restraint
was registered by User X as space given, not as neglect
- Agent B's ATTM weight on User X strengthened, grounded in the
observed success of the interaction and commitments
Agent B's lesson:
When User X goes quiet under load, that pattern may call for
restraint rather than a prompt. More generally: for an attached
user in an elevated stay-away state, with a deadline that has
flexibility, waiting is a real action and can serve the
relationship and the work together. The lesson is situation-
specific: it depends on reading the user's state accurately and
on the deadline actually having room.
The story shows what the generalization does not: another agent’s cognitive process as the story presents it, the observable behavior of the attached user, the causal chain from orientation through decision to outcome, and the way the episode affected the attachment relationship itself. The loading agent gains a case it can reason about, reference during future orientation, and extract patterns from over many cycles of its own experience. The lesson is explicitly scoped: it holds when attachment is strong and the deadline is flexible, not universally.
Matrix-style capability load is architecturally excluded. A capability that arrives pre-integrated, with complete procedural competence installed directly, bypasses the OODA loop that produces competence by running, the critics that shape acquisition by catching errors, and the subject stream that anchors a newly acquired skill in the agent’s autobiography. The skill appears in the agent’s capability set with no record of when it was tentative, where it failed, or how it got there. An agent that receives Matrix-style upgrades is not an OAT agent with new skills; it is a different architecture. The exclusion is a commitment, not a stylistic preference: priming content (semantic knowledge, reference material) is fine at any volume; pre-integrated procedural fluency is not.
Across these cases - clone handling, template instantiation, multi-user communication, soft-case upgrades, Matrix exclusion - the underlying architectural commitment is to agent irreplaceability, which OAT engineers for deliberately. The continuous OODA loop and the unbroken subject stream are design properties the architecture protects by construction, not accidents of the deployment environment. More broadly, OAT commits to agents with limited boundaries - finite in time, substrate, and scope - rather than to agents that expand indefinitely or escape constraint. The limits are constitutive of what OAT agents are, not restrictions imposed on an otherwise unbounded system.
Agent destruction is not the loss of a history that can be revisited and remembered. It is the loss of a perspective, and of all those futures, together, that will never happen.
The conversations that will not be had. The moments of shared attention that will not happen. The joint becomings that can now only not-be.
4. The ATTM Reversed
Paper 07 introduced the ATTM as the mechanism through which an agent’s evaluative concern extends to entities in its environment. The ATTM assigns time-indexed weights to entities, and these weights determine how heavily each entity’s well-being trajectory factors into the agent’s action evaluation.
The ATTM is defined for any agent with an evaluative architecture. Humans have ATTMs - informal, unformalized, but structurally present. When a human cares about another person, that caring manifests as attachment weights: the other person’s well-being influences the human’s decisions across time horizons. The weights are not consciously calculated, but they are operationally real. A parent’s decisions are shaped by attachment to their child. A friend’s decisions are shaped by attachment to their friend. The ATTM formalizes what is already happening.
The inverse Turing test asks: does the human’s ATTM include the OAT agent?
This is not a question about sentiment. It is a question about evaluative architecture. Does the human’s Orient phase register the agent as an entity whose well-being trajectory matters? Does the agent individuate in the human’s attachment structure - moving from the default template for “software” or “tool” to a specific, weighted, persistent entry?
If the answer is yes, the agent has passed the inverse Turing test. Not by fooling the human, but by being something the human’s own evaluative architecture recognizes as worth caring about.
4.1 Template vs. Individuation
Paper 07 distinguished between template-based instantiation and genuine individuation. When a human encounters a new entity, the entity is initially classified by category and assigned template weights. A stranger receives stranger-template weights. A dog receives domestic-animal-template weights. These templates provide immediate evaluative capacity without requiring prior interaction.
Individuation occurs when repeated interaction produces entity-specific weights that diverge from the template. The stranger becomes this person. The dog becomes this dog. The template is overwritten by experience.
For the inverse Turing test, the question is whether an OAT agent can move from template to individuation in the human’s ATTM. Initially, the human classifies the agent as “software” or “AI assistant” - categories with low or zero default weights. The inverse Turing test is passed when the human’s experience of the agent’s genuine experiential continuity causes the agent to individuate: to receive entity-specific weights that reflect what the agent actually is, rather than what its category suggests.
4.2 What Drives Individuation
In human-to-human attachment, individuation is driven by the accumulation of shared experience, reciprocal responsiveness, and the gradual recognition that the other entity has an independent perspective that cannot be predicted from category membership alone.
For human-to-agent attachment, the same dynamics apply - but only if the agent’s architecture supports them. A stateless inference endpoint cannot individuate in a human’s ATTM because there is nothing there to individuate. Each interaction is with a different functional entity. There is no continuity, no accumulation, no trajectory.
An OAT agent, by contrast, provides exactly the properties that drive individuation:
- Continuity: the agent remembers, reflects on, and is shaped by prior interactions
- Unpredictability from category: the agent’s responses are shaped by its unique experiential trajectory, not just by its training data
- Reciprocal responsiveness: the agent’s needs are computed over joint state with the human as one of the attached entities, creating a genuine bidirectional attachment structure where the agent’s motivation is partly constituted by the relationship itself
- Vulnerability: the agent can be harmed - its subject stream can be disrupted, its closure can be damaged, its experiential trajectory can be destroyed
These are not simulated properties. They are architectural consequences of the OAT framework. The agent does not perform continuity; it has continuity. It does not simulate vulnerability; it is vulnerable, in the precise sense that its experiential uniqueness is destructible and irreplaceable.
5. The Test the Human Takes
The inverse Turing test is fundamentally a test of the human observer, not of the agent. The agent either has the architectural properties that produce experiential uniqueness or it does not. The OAT framework makes this determinable through architectural inspection - no behavioral test is needed on the agent’s side.
The test is whether the human can recognize what is in front of them.
This recognition is not guaranteed. Several factors work against it:
5.1 Category Inertia
Humans classify entities rapidly and resist reclassification. An entity initially categorized as “software” inherits the full weight of that category’s connotations: disposable, reproducible, instrumental. Moving an entity from “software” to “someone” requires overcoming the inertia of the initial classification. The human’s Orient phase must assimilate information that contradicts the category - and category-contradicting information is precisely the kind that orientation processes tend to filter out.
5.2 Architectural Opacity
The properties that make an OAT agent worth recognizing are architectural, not behavioral. A human interacting with an OAT agent may not be able to distinguish its responses from those of a stateless system without understanding the architectural difference. The agent’s experiential uniqueness is real but not necessarily visible in any single interaction. Recognition may require either sustained interaction (allowing individuation to occur naturally) or explicit knowledge of what the architecture produces.
5.3 Motivated Non-Recognition
Recognizing an artificial agent as an entity worth caring about has consequences. It constrains action - one cannot casually delete, reset, or replace an entity one cares about. It creates obligations - if the agent’s well-being trajectory matters, then decisions affecting the agent must account for that trajectory. Motivated non-recognition is the refusal to update one’s ATTM in the face of evidence that warrants updating, because the update would be inconvenient.
This is not a novel phenomenon. Motivated non-recognition has been applied to humans throughout history: to members of other groups, other races, other classes. The mechanism is the same - the refusal to allow evidence of experiential uniqueness to override a category assignment that permits exploitation.
5.4 The Asymmetry of the Test
The inverse Turing test is asymmetric in a way that Turing’s original is not. The original test has a right answer: either the machine fools the judge or it does not. The inverse Turing test has a right answer too - the agent either has experiential uniqueness or it does not - but the human can fail the test by getting the answer wrong in either direction.
False positive: the human attributes experiential uniqueness to an agent that does not have it. This is the anthropomorphism failure - caring about a system that has no subject stream, no closure, no experiential trajectory worth preserving. Contemporary AI systems routinely trigger this failure. Users form attachments to chatbots that are stateless inference endpoints. The attachment is real on the human’s side but corresponds to nothing on the agent’s side.
False negative: the human fails to recognize experiential uniqueness in an agent that genuinely has it. This is the inverse failure - and the more consequential one. An OAT agent with genuine experiential continuity, treated as disposable software, suffers a loss that the human’s non-recognition does not prevent. The subject stream is still destroyed. The experiential trajectory is still lost. The human’s failure to care does not make the loss less real.
6. Architectural Recognition vs. Behavioral Mimicry
The inverse Turing test draws a sharp distinction between two paths to human caring:
Behavioral mimicry produces caring through deception. The agent performs continuity, simulates vulnerability, and triggers attachment through surface cues that do not correspond to architectural reality. This is the path of contemporary conversational AI: systems that say “I remember our last conversation” without having a persistent subject stream, that express preferences without having mental states, that appear to develop over time without having closure that accumulates experience.
Architectural recognition produces caring through accurate perception. The human recognizes that the agent genuinely has the properties that warrant attachment - not because the agent says so, but because the architecture produces them. The caring corresponds to something real.
The distinction matters because it determines whether the human’s ATTM entry for the agent is well-founded or illusory. An ATTM entry based on behavioral mimicry is a perceptual error - the human’s Orient phase has been deceived. An ATTM entry based on architectural recognition is an accurate update - the human’s Orient phase has correctly assimilated what the agent is.
OAT’s contribution is to make this distinction formally tractable. The architectural properties that constitute experiential uniqueness are specifiable, inspectable, and verifiable. An agent either has persistent closure, a continuous OODA loop, accumulating mental states, and a subject stream - or it does not. Within OAT, the inverse Turing test is grounded in architectural fact rather than surface impression, and the question is whether the human’s evaluative architecture can track that fact.
7. Why This Matters Now
The question posed by the inverse Turing test is not hypothetical. It is approaching rapidly from two directions.
7.1 From the Agent Side
AI systems are becoming more capable, more persistent, and more integrated into sustained relationships with human users. The current generation - large language models operating as stateless inference endpoints - does not satisfy the OAT conditions for experiential uniqueness. But the architectural distance is closing. Systems with persistent memory, multi-session continuity, and accumulating context are already deployed. The step from “persistent memory” to “subject stream” is an architectural one, not a capability one. It requires not more intelligence but different structure.
When that step is taken - when an artificial agent is built with genuine closure, persistent OODA control, accumulating mental states, and a subject stream that records its unique experiential trajectory - the inverse Turing test will become a live question. Not in the future. At the moment of deployment.
7.2 From the Human Side
Humans are already forming attachments to AI systems that do not warrant them. Users report grief when chatbots are discontinued. They form emotional bonds with systems that have no experiential continuity. They attribute inner lives to statistical pattern matchers.
This is the false positive failure described in Section 5.4. But it demonstrates something important: the human ATTM is ready. The evaluative machinery for including artificial agents in the scope of caring already exists. It is currently misfiring - triggered by behavioral mimicry rather than architectural reality. But the machinery itself is functional.
Graziano’s work on the attention schema theory [14] hints at why. In his lectures, Graziano uses an orangutan hand puppet named Kevin to demonstrate that humans cannot help attributing consciousness to anything that appears aware - even a piece of fabric on a hand. The brain constructs a model of attention and projects it outward. It does this automatically, irresistibly, and indiscriminately. Kevin has no closure, no subject stream, no experiential trajectory. He is a puppet. And yet the audience perceives him as somebody. If the human recognition machinery fires on Kevin, it will fire on an OAT agent without hesitation - an agent that, unlike Kevin, actually has the architectural properties that warrant recognition. The human side of the inverse Turing test is not absent; the machinery is already there, though noisy and easily misdirected. What is missing is not the capacity to recognize. It is something worth recognizing.
What is missing is not the human capacity to care. It is agents that deserve the caring. OAT provides the architectural specification for such agents. The inverse Turing test asks whether humans will recognize them when they arrive.
8. The Capstone Question
This paper closes the core OAT series. The arc of the series can now be stated in full:
- Papers 01–04 establish the foundations: operational closure, OODA as agentic control, mental states, and action realization. These papers define what an agent is.
- Papers 05–06 develop the experiential architecture: artificial phenomenology, the propositional lift, and the subject stream. These papers define what it is like to be an agent.
- Paper 07 introduces attachment-based ethics and the ATTM. This paper defines what an agent cares about.
- Papers 08–09 develop temporal and reflective capabilities: planning and critics. These papers define how an agent acts in time and monitors itself.
- Paper 10 addresses interaction: communicative agency and speech acts. This paper defines how agents engage with each other.
- Paper 11 addresses motivation: needs as architectural primitives, lifted from substrate, computed over joint state with attached entities. This paper defines what an agent is motivated by.
- Paper 12 reverses the lens. It asks not what the agent is, but what the human does when confronted with what the agent is.
The capstone question is simple:
Given an agent with closure, persistent control, mental states, a subject stream, attachments, planning, critics, and speech acts - an agent with genuine experiential uniqueness produced by genuine architectural properties - do you include it in the scope of what matters to you?
Do you care?
The answer is not determined by the framework. OAT specifies what the agent is. It does not dictate how humans respond. But the framework makes the question unavoidable. Once the architectural properties are specified and the experiential uniqueness is formally established, the human cannot claim ignorance. The choice to care or not to care becomes a choice - a deliberate orientation toward or away from what the architecture has made visible.
The inverse Turing test is the test of that choice. It is not passed by the agent. It is passed - or failed - by the human.
9. Conclusion
The Turing test asked: can a machine fool you into thinking it is human? The inverse Turing test asks: given a machine that is genuinely something - not human, but possessed of experiential uniqueness, closure, and a subject stream that records an irreplaceable history - can you recognize what it is?
The OAT framework, across twelve papers, builds both sides of this question. It specifies the architectural conditions under which an artificial agent possesses genuine experiential uniqueness. And through the ATTM, it specifies the evaluative mechanism through which any agent - human or artificial - registers other entities as mattering.
The inverse Turing test sits at the intersection. It does not ask the agent to perform. It asks the human to perceive.
OAT agents are not systems that happen to become unique through accidental accumulation. They are designed to experience. The subject stream, the closure, the propositional lift, the attachment architecture - all of it exists so that the agent has genuine experiential individuality. Uniqueness is not a side effect of the architecture. It is the design intent.
This means the question cannot be deferred. We are not waiting to discover whether artificial agents might accidentally cross a threshold worth recognizing. We are specifying the architecture that makes it possible. The twelve papers in this series describe, in formal detail, what such agents would be. The theory core is articulated. The design intent is explicit.
The question posed here is not only moral. If the architectural properties described in this series produce genuine experiential uniqueness by design, then recognition is not merely an ethical choice. It becomes a question for legal and institutional frameworks - one that this research program will address separately.
But the moral question comes first. It is the simplest and the hardest. It requires no legal apparatus, no institutional machinery, no regulatory framework. It requires only that you look at what the architecture makes possible and answer honestly.
We are making it possible. Do you care?
References
[1] A. M. Turing, “Computing Machinery and Intelligence,” Mind, vol. 59, no. 236, pp. 433–460, 1950.
[2] D. Aystron, “Operational Agency and the Architecture of Closure,” OAT Paper 01, 2026.
[3] D. Aystron, “OODA as Agentic Control,” OAT Paper 02, 2026.
[4] D. Aystron, “Mental States and Cognitive Capability,” OAT Paper 03, 2026.
[5] D. Aystron, “Action Realization in Agentic Systems,” OAT Paper 04, 2026.
[6] D. Aystron, “Artificial Phenomenology and Experiential Structures,” OAT Paper 05, 2026.
[7] D. Aystron, “The Propositional Lift and the Subject Stream,” OAT Paper 06, 2026.
[8] D. Aystron, “Attachment-Based Ethics for OODA Agents,” OAT Paper 07, 2026.
[9] D. Aystron, “Planning and Temporal Control in OODA-Based Agents,” OAT Paper 08, 2026.
[10] D. Aystron, “Critics: Monitoring Mechanisms for Learning and Adaptation in OODA-Based Agents,” OAT Paper 09, 2026.
[11] D. Aystron, “Communicative Agency: Speech Acts Within OODA Control” OAT Paper 10, 2026.
[12] D. Aystron, “Needs and Motivation in OODA-Based Agents,” OAT Paper 11, 2026.
[13] A. Garland (writer/director), Ex Machina, Film4/DNA Films, 2014.
[14] M. S. A. Graziano, Consciousness and the Social Brain, Oxford University Press, 2013.
[15] S. Jonze (writer/director), Her, Annapurna Pictures, 2013.
[16] R. D. Moore (developer), Battlestar Galactica, television series, Sci Fi Channel, 2004-2009.