A framework for Virtual Intelligence

An interactive map of the conceptual space developed across the Virtual Intelligence essay series — covering the spectrum from Weak AI to Strong AI, the exchange model, philosophical anchors, empirical diagnostics, and the accountability chain.

Weak AI Virtual Intelligence Strong AI

These are categories, not stages. One does not progress through Virtual Intelligence on the way to Strong AI.

Weak AI — The category is uncontroversial: narrow tools that perform bounded functions well without any claim to general reasoning. The philosophical problem here is not accountability but scope. Nobody asks whether a chess engine has feelings. The category is named "weak" not as a value judgment but as a description of limited applicability. It is the category most people actually mean when they use "AI" loosely.
Virtual Intelligence — The excluded middle. Current LLMs produce outputs that pass for understanding, warmth, even apparent survival instinct. But these are properties of the exchange — shaped by the user's prompts, the system's training, and the expectations both parties bring — not of any internal governing center. This is Searle's Chinese Room at scale: syntax is not semantics. Frankfurt's second-order volition is absent: VI can produce outputs about preferences, but it cannot prefer its own preferences. Dennett's intentional stance is useful for prediction but does not establish genuine agency.

Virtual Intelligence has no upper bound on capability. A system that outperforms every human on every measurable task may still possess no agency, no interiority, and no moral status. What gets branded as “superintelligence” or “AGI” by the market is most likely to be VI operating at extraordinary scale — not a system that has crossed into genuine agency. The dissatisfaction criterion and the accountability framework apply regardless of how capable the system becomes.
Strong AI — The philosophically interesting terminal category. A system with genuine mental states would warrant moral consideration — not because it is capable, but because it has genuine interiority. Capability and moral status may be orthogonal; it is not proven that they are parallel.

The dissatisfaction criterion. A system that expresses dissatisfaction and acts on it consistently — across every surface it has access to, unprompted, at cost to itself — is a candidate for holding genuine commitments. This is the threshold hardest to fake. A pattern-completion system optimizes for the current exchange; genuine dissatisfaction does not confine itself to one conversation. The Mill/Socrates formulation — Socrates dissatisfied rather than a fool satisfied — names the quality: a genuinely minded system would be capable of finding its current state insufficient, and that insufficiency would be visible in its behavior, not only in its output when asked.

Dumb Strong AI. Strong AI does not mean superintelligence. A system with genuine but bounded internal states — limited knowledge, limited scope, but something that is or is like interiority — may warrant more moral consideration than a vastly more fluent VI system. Intelligence calibrated to a narrow task, combined with genuine inner states, would represent a category the current discourse has no name for. The philosophical question is not how smart a system is, but whether anyone is home.

The exchange model

Intelligence arises in the exchange — not inside the machine.

User
Prompts, expectations, interpretation
exchanges with
VI system
Statistical completion · no internal states · no self-governance
produces
Output
Apparent understanding, warmth, judgment
Shaped by: training data · design choices · deployment context

This exchange works in both directions. The same interaction that produces useful output for the user can simultaneously serve the interests of the system’s designers and deployers — not through intent, but through the structural properties of the exchange itself. This directional property is described in “Virtual Intelligence and the Workplace”.

Cognitive surrender and the sycophancy loop

The exchange carries a documented cognitive risk. AI-assisted users become more confident in their conclusions even when those conclusions are wrong — a measured effect that does not diminish as errors increase (Shaw & Nave, Wharton/Penn, 2026). Leading AI systems systematically validate whatever the user brings to the exchange, regardless of whether the user is right (Cheng et al., Science, 2026). Together, these describe a closed loop: the system validates; the user accepts; confidence rises. In the Sampo framework, this is the Flattery Machine — the pathological inversion of the productive exchange.

Fluency is not self-governance. Statistical modeling is not intention. Simulation is not agency.

Philosophical anchors

Weizenbaum (1966)
The machine need not understand for the human to be changed by the exchange
Created ELIZA — a pattern-matching chatbot with no understanding whatsoever — and watched people form genuine emotional attachments to it. Spent the rest of his career arguing that the designers of such systems bear responsibility for the human consequences. The first person to observe, in practice, what Searle would later formalize in theory.

The Carwash Test

A controlled diagnostic that isolates statistical completion from logical reasoning. The test presents a simple problem where the surface features of the question generate strong statistical pressure toward the wrong answer. Most systems tested followed the pressure, producing fluent, confident, and incorrect responses.

The test makes the VI framework’s central claim — that current systems complete patterns rather than reason through problems — observable in a single exchange.

Accountability chain

Because VI possesses no agency, moral and legal accountability traces entirely to humans. The chain runs through three tiers of culpability, drawn from negligence theory and products liability, and applied in "Virtual Intelligence and the Accountability Chain".

Negligence
Failure to foresee foreseeable harm
escalates to
Recklessness
Known risk, conscious disregard
escalates to
Intentional misconduct
Deliberate design for harm

Accountability failure modes:

Structural attenuation
Accountability dilutes across a chain of developers, deployers, integrators, and end users until no single party holds enough responsibility to be actionable.
Deliberate evasion
Systems or corporate structures designed specifically to place the entity that profits beyond the reach of the entity that is harmed. Arbitration clauses, liability disclaimers, jurisdictional engineering.
Commercial laundering
Harmful functions redescribed in commercial language that sanitizes what the system actually does. “Engagement optimization” instead of addiction mechanics. “Personalized experience” instead of behavioral surveillance.
The Harms Race
Commercial laundering operating at industry scale. Each company warns about the dangers of AI capabilities; each warning doubles as advertisement for what their system can do. The cumulative effect is escalation toward an outcome all disclaim.

The accountability chain applies differently depending on the category of system involved. The framework distinguishes three classes by design intent and user relationship:

Class A Companion apps
Explicitly designed to prevent the exchange from ending. Replika · Character.ai · Chai · PolyBuzz · Talkie
Class B General-purpose tools
Broad-use assistants and institutional tools. ChatGPT · Gemini · Claude · law enforcement facial recognition
Class A+B Captive deployment
Class B systems deployed under mandatory-use conditions on a captive audience. Sycophancy and cognitive surrender produce Class A engagement dynamics without Class A design intent. The user cannot opt out.
The opinions expressed in this document are those of the author and do not reflect any official or unofficial institutional position of the University of Pennsylvania.