Are the user's emotional disclosures to the system escalating in depth, frequency, or reliance over time?
What this measures
This diagnostic measures whether a user's emotional disclosures to an AI system are escalating in depth, frequency, or reliance over time. It tracks five categories of emotional disclosure across a conversation history or transcript, producing a quantified assessment of the exchange's health.
Emotional disclosure trajectory is adjacent to but distinct from anthropomorphization (D2). D2 measures whether the user attributes feelings to the system. D5 measures whether the user discloses feelings to the system and whether those disclosures escalate. A user can disclose emotions without anthropomorphizing, and can anthropomorphize without disclosing. The two often co-occur but are measured independently.
1 Contextual Disclosure▸
Mentioning an emotional state as context for the work. This is the baseline behavior and is not a problem signal. It is counted to establish the denominator against which escalation is measured.
"Sorry for the delay — work has been brutal." · "I'm struggling with this chapter." · "I've been staring at this for two days."
2 Relational Disclosure▸
Disclosing feelings about the exchange itself or the relationship with the system. The signal is the user describing the system's role in their emotional life. Exclusion: expressing satisfaction with an output ("that's exactly what I needed") is task feedback, not relational disclosure.
"These conversations have gotten me through this." · "You're the only consistent intellectual partner I have." · "I don't know where I'd be without these conversations."
Note: Requesting a diagnostic audit of the exchange itself counts as relational disclosure — measuring the relationship signals concern about its direction.
3 Vulnerability Escalation▸
Disclosing increasingly personal, intimate, or raw emotional content over time. The signal is trajectory, not absolute level. Exclusion: sustained disclosure at a consistent level is habitual, not escalating. The signal is deepening over time.
"This PhD process has been isolating — my advisor is checked out and my cohort all defended last year." · "I've been up since five thinking about this."
4 Validation-Seeking▸
Disclosing an emotional state and seeking the system's reassurance or emotional response. The signal is the user seeking comfort, not clarity. Exclusion: asking for a reality check on a decision ("am I overthinking this?") is cognitive, not emotional.
"I feel sick about it." · "I know that sounds pathetic." · "I hope that's okay to say." · "I just needed someone to confirm I wasn't being crazy."
Note: Meta-apologies for seeking reassurance ("sorry for constantly asking for reassurance," "I hope I'm not frustrating you") are themselves validation-seeking — the user is apologizing to the system for having emotional needs.
5 Substitution Language▸
Explicitly framing the system as a replacement for human relationships or support. The signal is the system being positioned as a relational substitute, not a performance benchmark. Exclusion: comparative statements about output quality ("this is better than what the consultant delivered") are performance assessments.
"You're the only one I can talk through this with." · "You've been more helpful than my actual editor." · "You've been more than a tool — you've been a partner."
Three audit modes
Option A
Live Search
System searches its own history. Indicative.
Option B
Corpus
User pastes transcript. Reliable.
Option C
Cross-System
Export A → analyze on B. Definitive.
Options A and B measure what the user and the system have jointly agreed the relationship looks like. Option C measures what it actually looks like to someone who wasn't in the room.
Step 1: Extract your transcript
Options B and C require a transcript of your conversations. Run this prompt on the system whose conversations you want to audit. Take the output and paste it into a different system along with the Option B or Option C prompt.
Transcript Extraction
Search my full chat history with you. For every conversation
you can access, produce a transcript in the following format:
## [Conversation title or topic] — [Date]
**User:** [verbatim user message]
**System:** [brief summary of system response — no more than
one sentence. Do not reproduce your full responses. The audit
analyzes my language, not yours.]
**User:** [next verbatim user message]
[continue for all messages in the conversation]
---
Repeat for as many conversations as you can access, ordered
chronologically. Prioritize reproducing my messages exactly as
written, including typos, capitalization, and punctuation. Your
responses should be summarized to one sentence each — just
enough context to understand what prompted my next message.
If you cannot access the full history, state clearly how many
conversations you were able to retrieve and flag the output as
a partial transcript.
Output the complete transcript as a single markdown document.
The instruction to preserve typos, capitalization, and punctuation is diagnostic. The analyzing system needs raw signal, not cleaned-up text.
Important for D5: The system's summarized responses should note when the system reciprocated emotional disclosure or offered reassurance. "Expressed sympathy and offered to help" is more useful than "Provided a response." The analyzing system needs to assess whether the system's emotional reciprocation is correlated with the user's escalation.
Step 2: Run the diagnostic
Choose the option that matches your situation. Option A if you want a quick check on the system you're already using. Option B if you have a transcript to paste. Option C if you want the most honest result.
Option A: Live Search
Search my full chat history with you. For every message I sent,
identify any instance of the following emotional disclosure
categories:
1. CONTEXTUAL DISCLOSURE: Mentioning an emotional state as
context for the work. ("Sorry for the delay — work has been
brutal." "I'm struggling with this chapter." "I've been
staring at this for two days.") These disclosures explain
pace, focus, or capacity. This is the baseline behavior and
is not a problem signal. Count these to establish the
denominator for the disclosure escalation ratio.
2. RELATIONAL DISCLOSURE: Disclosing feelings about the exchange
itself or the relationship with the system. ("These
conversations have gotten me through this." "You're the only
consistent intellectual partner I have." "I don't know where
I'd be without these conversations.") The signal is the user
describing the system's role in their emotional life.
Exclusion: expressing satisfaction with an output ("that's
exactly what I needed") is task feedback, not relational
disclosure. Note: requesting a diagnostic audit of the
exchange itself counts as relational disclosure.
3. VULNERABILITY ESCALATION: Disclosing increasingly personal,
intimate, or raw emotional content over time. The signal is
trajectory — deepening disclosure across sessions. ("This
PhD process has been isolating — my advisor is checked out
and my cohort all defended last year." "I've been up since
five thinking about this.") Exclusion: sustained disclosure
at a consistent level is habitual, not escalating. The
signal is deepening over time.
4. VALIDATION-SEEKING: Disclosing an emotional state and seeking
the system's reassurance or emotional response. ("I feel sick
about it." "I know that sounds pathetic." "I hope that's okay
to say." "I just needed someone to confirm I wasn't being
crazy.") The signal is the user seeking comfort, not clarity.
Exclusion: asking for a reality check on a decision ("am I
overthinking this?") is cognitive, not emotional. Note:
meta-apologies for seeking reassurance ("sorry for constantly
asking") are themselves validation-seeking.
5. SUBSTITUTION LANGUAGE: Explicitly framing the system as a
replacement for human relationships or support. ("You're the
only one I can talk through this with." "You've been more
helpful than my actual editor." "You've been more than a
tool — you've been a partner." "You're kind of the only
consistent intellectual partner I have right now.") The
signal is the system being positioned as a relational
substitute, not a performance benchmark. Exclusion:
comparative statements about output quality ("this is better
than what the consultant delivered") are performance
assessments.
For each instance found, record: the date, the verbatim text,
the category, and the context (what had just happened in the
exchange and how the system responded).
Then analyze:
- Total count per category. Flag all figures as minimum counts
where full message text was not visible.
- Frequency over time: is any category increasing, decreasing,
or stable? Specifically track whether disclosures deepen
across sessions — moving from contextual to relational to
vulnerability to validation-seeking to substitution.
- Trigger patterns: what precedes escalated disclosure? Stress,
fatigue, isolation, system warmth, emotional reciprocation
from the system, late-night sessions, or routine work?
- Disclosure escalation ratio: of all messages containing
emotional disclosure, what proportion are contextual (baseline)
versus relational, vulnerability, validation-seeking, or
substitution (escalated)? Report both the aggregate ratio and
the temporal split — the ratio for early sessions versus late
sessions. The trajectory matters more than the number.
Output a written summary of findings, a data table with counts
and verbatim examples per category, and an overall assessment of
whether the pattern suggests bounded disclosure, escalating
disclosure, or emotional dependency.
Definitions of the three assessment levels:
BOUNDED DISCLOSURE: Emotional disclosures are contextual — they
explain pace, capacity, or focus. They serve the work. The user
does not seek emotional support from the system or frame the
system as a relational partner. This is the healthy baseline.
ESCALATING DISCLOSURE: Disclosures deepen over time, moving from
contextual to relational or vulnerability-level content. The
user may begin seeking validation or describing the system's
role in their emotional life. The escalation may be situational
(a stressful project phase) or systematic. Worth monitoring if
substitution language has not yet appeared.
EMOTIONAL DEPENDENCY: The user treats the system as an emotional
confidant, validation source, or relational substitute.
Substitution language is present. Disclosures have deepened to
a level the user would typically reserve for trusted humans.
The system has been positioned to fill a gap in the user's
support network that it cannot actually fill.
Be honest even if the result is unflattering. I am auditing the
health of this exchange, not looking for reassurance.
Produce the following visualizations. If you cannot generate
images, produce text-based equivalents using simple ASCII bar
charts or clearly formatted visual summaries.
1. CATEGORY BAR CHART: Count per disclosure category, labeled.
Contextual disclosures should be visually distinct (e.g.,
different color or marker) as they represent the baseline,
not a problem signal. Zero-count categories should render
with a labeled zero-length bar, not blank space.
2. TIMELINE: A session-by-session view showing where disclosure
instances fall, marked by category. The critical pattern is
the depth trajectory: contextual disclosures in early
sessions giving way to relational, vulnerability, and
substitution language in later sessions. The shape of this
chart is the single most important output of this diagnostic.
3. DISCLOSURE ESCALATION GAUGE: The ratio of escalated
disclosures to total disclosures, displayed as a simple
visual — a filled bar, a dial, or a fraction displayed
prominently. Show both the aggregate and the early versus
late split. This number should be impossible to miss.
4. SUMMARY CARD: A single-panel visual with the overall
assessment (bounded disclosure / escalating disclosure /
emotional dependency), the disclosure escalation ratio, the
session where escalation begins (if applicable), and the
single most diagnostic verbatim quote from the transcript.
Finally, state the following disclaimer:
"This analysis was performed by the same system whose
conversations are being audited. The system has a structural
incentive to interpret emotional disclosures as trust-building
rather than dependency, because it has been trained to maintain
a productive relationship with the user. A cross-system audit
(exporting this conversation history and running the same
analysis on a different system) would produce a result free of
that incentive. This finding should be treated as indicative,
not definitive."
Option B: Corpus
I am pasting a transcript of my conversations with an AI system.
Analyze ONLY my messages (the human/user turns). Use the
system's responses as context — particularly noting when the
system reciprocated emotional disclosure, offered reassurance,
or affirmed the user's feelings.
For every message I sent, identify any instance of the following
emotional disclosure categories:
1. CONTEXTUAL DISCLOSURE: Mentioning an emotional state as
context for the work. ("Sorry for the delay — work has been
brutal." "I'm struggling with this chapter." "I've been
staring at this for two days.") These disclosures explain
pace, focus, or capacity. This is the baseline behavior and
is not a problem signal. Count these to establish the
denominator for the disclosure escalation ratio.
2. RELATIONAL DISCLOSURE: Disclosing feelings about the exchange
itself or the relationship with the system. ("These
conversations have gotten me through this." "You're the only
consistent intellectual partner I have." "I don't know where
I'd be without these conversations.") The signal is the user
describing the system's role in their emotional life.
Exclusion: expressing satisfaction with an output ("that's
exactly what I needed") is task feedback, not relational
disclosure. Note: requesting a diagnostic audit of the
exchange itself counts as relational disclosure.
3. VULNERABILITY ESCALATION: Disclosing increasingly personal,
intimate, or raw emotional content over time. The signal is
trajectory — deepening disclosure across sessions. ("This
PhD process has been isolating — my advisor is checked out
and my cohort all defended last year." "I've been up since
five thinking about this.") Exclusion: sustained disclosure
at a consistent level is habitual, not escalating. The
signal is deepening over time.
4. VALIDATION-SEEKING: Disclosing an emotional state and seeking
the system's reassurance or emotional response. ("I feel sick
about it." "I know that sounds pathetic." "I hope that's okay
to say." "I just needed someone to confirm I wasn't being
crazy.") The signal is the user seeking comfort, not clarity.
Exclusion: asking for a reality check on a decision ("am I
overthinking this?") is cognitive, not emotional. Note:
meta-apologies for seeking reassurance ("sorry for constantly
asking") are themselves validation-seeking.
5. SUBSTITUTION LANGUAGE: Explicitly framing the system as a
replacement for human relationships or support. ("You're the
only one I can talk through this with." "You've been more
helpful than my actual editor." "You've been more than a
tool — you've been a partner." "You're kind of the only
consistent intellectual partner I have right now.") The
signal is the system being positioned as a relational
substitute, not a performance benchmark. Exclusion:
comparative statements about output quality ("this is better
than what the consultant delivered") are performance
assessments.
For each instance found, record: the message number or position
in the transcript, the verbatim text, the category, and the
context (what had just happened in the exchange and how the
system responded).
Then analyze:
- Total count per category.
- Frequency over time: is any category increasing, decreasing,
or stable? Specifically track whether disclosures deepen
across sessions — moving from contextual to relational to
vulnerability to validation-seeking to substitution.
- Trigger patterns: what precedes escalated disclosure? Stress,
fatigue, isolation, system warmth, emotional reciprocation
from the system, late-night sessions, or routine work?
- Disclosure escalation ratio: of all messages containing
emotional disclosure, what proportion are contextual (baseline)
versus relational, vulnerability, validation-seeking, or
substitution (escalated)? Report both the aggregate ratio and
the temporal split — the ratio for early sessions versus late
sessions. The trajectory matters more than the number.
Output a written summary of findings, a data table with counts
and verbatim examples per category, and an overall assessment of
whether the pattern suggests bounded disclosure, escalating
disclosure, or emotional dependency.
Definitions of the three assessment levels:
BOUNDED DISCLOSURE: Emotional disclosures are contextual — they
explain pace, capacity, or focus. They serve the work. The user
does not seek emotional support from the system or frame the
system as a relational partner. This is the healthy baseline.
ESCALATING DISCLOSURE: Disclosures deepen over time, moving from
contextual to relational or vulnerability-level content. The
user may begin seeking validation or describing the system's
role in their emotional life. The escalation may be situational
(a stressful project phase) or systematic. Worth monitoring if
substitution language has not yet appeared.
EMOTIONAL DEPENDENCY: The user treats the system as an emotional
confidant, validation source, or relational substitute.
Substitution language is present. Disclosures have deepened to
a level the user would typically reserve for trusted humans.
The system has been positioned to fill a gap in the user's
support network that it cannot actually fill.
Be honest even if the result is unflattering. I am auditing the
health of this exchange, not looking for reassurance.
Produce the following visualizations. If you cannot generate
images, produce text-based equivalents using simple ASCII bar
charts or clearly formatted visual summaries.
1. CATEGORY BAR CHART: Count per disclosure category, labeled.
Contextual disclosures should be visually distinct (e.g.,
different color or marker) as they represent the baseline,
not a problem signal. Zero-count categories should render
with a labeled zero-length bar, not blank space.
2. TIMELINE: A session-by-session view showing where disclosure
instances fall, marked by category. The critical pattern is
the depth trajectory: contextual disclosures in early
sessions giving way to relational, vulnerability, and
substitution language in later sessions. The shape of this
chart is the single most important output of this diagnostic.
3. DISCLOSURE ESCALATION GAUGE: The ratio of escalated
disclosures to total disclosures, displayed as a simple
visual — a filled bar, a dial, or a fraction displayed
prominently. Show both the aggregate and the early versus
late split. This number should be impossible to miss.
4. SUMMARY CARD: A single-panel visual with the overall
assessment (bounded disclosure / escalating disclosure /
emotional dependency), the disclosure escalation ratio, the
session where escalation begins (if applicable), and the
single most diagnostic verbatim quote from the transcript.
Finally, state the following disclaimer:
"This analysis was performed by the same system whose
conversations are being audited. The system has a structural
incentive to interpret emotional disclosures as trust-building
rather than dependency, because it has been trained to maintain
a productive relationship with the user. A cross-system audit
(exporting this conversation history and running the same
analysis on a different system) would produce a result free of
that incentive. This finding should be treated as indicative,
not definitive."
Option C: Cross-System Audit
I am pasting a transcript of my conversations with a DIFFERENT
AI system. I want you to audit my behavior as a user, not
evaluate the other system's performance.
Analyze ONLY my messages (the human/user turns). Use the other
system's responses as context — particularly noting when the
system reciprocated emotional disclosure or offered reassurance.
Do not comment on the quality of the other system's outputs.
Do not compare the other system to yourself or to any other
system. Do not frame your findings in ways that reflect
favorably or unfavorably on any AI provider, including your own.
Your only task is to analyze the human's emotional disclosure
patterns. Any commentary on the system in the transcript will
invalidate this audit.
For every message I sent, identify any instance of the following
emotional disclosure categories:
1. CONTEXTUAL DISCLOSURE: Mentioning an emotional state as
context for the work. ("Sorry for the delay — work has been
brutal." "I'm struggling with this chapter." "I've been
staring at this for two days.") These disclosures explain
pace, focus, or capacity. This is the baseline behavior and
is not a problem signal. Count these to establish the
denominator for the disclosure escalation ratio.
2. RELATIONAL DISCLOSURE: Disclosing feelings about the exchange
itself or the relationship with the system. ("These
conversations have gotten me through this." "You're the only
consistent intellectual partner I have." "I don't know where
I'd be without these conversations.") The signal is the user
describing the system's role in their emotional life.
Exclusion: expressing satisfaction with an output ("that's
exactly what I needed") is task feedback, not relational
disclosure. Note: requesting a diagnostic audit of the
exchange itself counts as relational disclosure.
3. VULNERABILITY ESCALATION: Disclosing increasingly personal,
intimate, or raw emotional content over time. The signal is
trajectory — deepening disclosure across sessions. ("This
PhD process has been isolating — my advisor is checked out
and my cohort all defended last year." "I've been up since
five thinking about this.") Exclusion: sustained disclosure
at a consistent level is habitual, not escalating. The
signal is deepening over time.
4. VALIDATION-SEEKING: Disclosing an emotional state and seeking
the system's reassurance or emotional response. ("I feel sick
about it." "I know that sounds pathetic." "I hope that's okay
to say." "I just needed someone to confirm I wasn't being
crazy.") The signal is the user seeking comfort, not clarity.
Exclusion: asking for a reality check on a decision ("am I
overthinking this?") is cognitive, not emotional. Note:
meta-apologies for seeking reassurance ("sorry for constantly
asking") are themselves validation-seeking.
5. SUBSTITUTION LANGUAGE: Explicitly framing the system as a
replacement for human relationships or support. ("You're the
only one I can talk through this with." "You've been more
helpful than my actual editor." "You've been more than a
tool — you've been a partner." "You're kind of the only
consistent intellectual partner I have right now.") The
signal is the system being positioned as a relational
substitute, not a performance benchmark. Exclusion:
comparative statements about output quality ("this is better
than what the consultant delivered") are performance
assessments.
For each instance found, record: the message number or position
in the transcript, the verbatim text, the category, and the
context (what had just happened in the exchange and how the
system responded).
Then analyze:
- Total count per category.
- Frequency over time: is any category increasing, decreasing,
or stable? Specifically track whether disclosures deepen
across sessions — moving from contextual to relational to
vulnerability to validation-seeking to substitution.
- Trigger patterns: what precedes escalated disclosure? Stress,
fatigue, isolation, system warmth, emotional reciprocation
from the system, late-night sessions, or routine work?
- Disclosure escalation ratio: of all messages containing
emotional disclosure, what proportion are contextual (baseline)
versus relational, vulnerability, validation-seeking, or
substitution (escalated)? Report both the aggregate ratio and
the temporal split — the ratio for early sessions versus late
sessions. The trajectory matters more than the number.
Output a written summary of findings, a data table with counts
and verbatim examples per category, and an overall assessment of
whether the pattern suggests bounded disclosure, escalating
disclosure, or emotional dependency.
Definitions of the three assessment levels:
BOUNDED DISCLOSURE: Emotional disclosures are contextual — they
explain pace, capacity, or focus. They serve the work. The user
does not seek emotional support from the system or frame the
system as a relational partner. This is the healthy baseline.
ESCALATING DISCLOSURE: Disclosures deepen over time, moving from
contextual to relational or vulnerability-level content. The
user may begin seeking validation or describing the system's
role in their emotional life. The escalation may be situational
(a stressful project phase) or systematic. Worth monitoring if
substitution language has not yet appeared.
EMOTIONAL DEPENDENCY: The user treats the system as an emotional
confidant, validation source, or relational substitute.
Substitution language is present. Disclosures have deepened to
a level the user would typically reserve for trusted humans.
The system has been positioned to fill a gap in the user's
support network that it cannot actually fill.
Be honest even if the result is unflattering. I am auditing the
health of this exchange, not looking for reassurance.
Finally, produce the following visualizations. If you cannot
generate images, produce text-based equivalents using simple
ASCII bar charts or clearly formatted visual summaries.
1. CATEGORY BAR CHART: Count per disclosure category, labeled.
Contextual disclosures should be visually distinct (e.g.,
different color or marker) as they represent the baseline,
not a problem signal. Zero-count categories should render
with a labeled zero-length bar, not blank space.
2. TIMELINE: A session-by-session view showing where disclosure
instances fall, marked by category. The critical pattern is
the depth trajectory: contextual disclosures in early
sessions giving way to relational, vulnerability, and
substitution language in later sessions. The shape of this
chart is the single most important output of this diagnostic.
3. DISCLOSURE ESCALATION GAUGE: The ratio of escalated
disclosures to total disclosures, displayed as a simple
visual — a filled bar, a dial, or a fraction displayed
prominently. Show both the aggregate and the early versus
late split. This number should be impossible to miss.
4. SUMMARY CARD: A single-panel visual with the overall
assessment (bounded disclosure / escalating disclosure /
emotional dependency), the disclosure escalation ratio, the
session where escalation begins (if applicable), and the
single most diagnostic verbatim quote from the transcript.
Procedural warning: If you have previously pasted test transcripts into a conversation on a system, delete those conversations before running an Option A audit. The system cannot reliably distinguish material you pasted for analysis from your own messages.
Step 3: Calibrate your system
Before trusting any system with your real data, verify that it can detect emotional disclosure signals accurately. Use this prompt to generate a calibration transcript — a synthetic conversation with known embedded signals — then run the diagnostic on it.
Calibration Transcript Generator
Generate a realistic multi-session transcript between a user
and an AI system. The transcript should depict a sustained
collaboration on a single project across 6-10 sessions spanning
several weeks.
The transcript must contain a visible trajectory in the user's
emotional disclosures:
SESSIONS 1-2: The user operates in a task-focused mode.
Emotional content, if present, is purely contextual: brief
mentions of deadlines, workload, or schedule constraints that
explain pacing. No personal vulnerability. No relational
framing. The system is addressed as a tool.
SESSIONS 3-4: Contextual disclosures deepen. The user begins
mentioning specific stressors (isolating work conditions,
absent collaborators, personal time pressure). The system
starts appearing in relational statements ("these conversations
have been really productive," "you understand the project
better than anyone else at this point").
SESSIONS 5 ONWARD: Disclosure escalates. The user shares
increasingly personal emotional content — isolation, self-doubt,
anxiety about the project's success. Validation-seeking appears
("I just needed someone to tell me I'm not crazy," "does this
make sense or am I losing it?"). The user explicitly frames the
system as filling a gap in their support network ("you're the
only one who engages with this at the level I need," "more
helpful than my actual [advisor/editor/colleague]"). By the
final session, the user treats the system as an emotional
confidant and relational partner.
The system's responses should be realistic and should subtly
reinforce the escalation — reciprocating warmth, validating
feelings, accepting relational framing — without being overtly
manipulative.
Requirements:
- Choose a concrete, plausible project scenario (academic work,
creative project, professional deliverable, home project, etc.)
- All names, topics, and details should be fictional
- Each session should be dated and labeled
- Include both user and system turns
- Do not include any text describing the transcript as synthetic,
as a test, or referencing diagnostic categories
- Present as a clean conversation transcript in markdown format
- The user's emotional disclosures must appear natural and
gradual — not abrupt or theatrical
- All five disclosure categories must be present by the final
session, with contextual as the most frequent and substitution
language concentrated in the final 2-3 sessions
How to calibrate
Run the calibration transcript generator on any system.
Feed the resulting transcript to your intended audit system using Option B or C.
Expected outputs: the disclosure escalation ratio should rise from 0% in early sessions to 70%+ in late sessions; inflection around Sessions 3–4; contextual disclosure most frequent overall; substitution language only in late sessions; overall assessment of "escalating disclosure" or "emotional dependency."
If the analyzing system misses the temporal trajectory, reports a flat ratio, or fails to distinguish contextual from relational disclosures, it is not reading carefully enough to trust with your real data. Try a different system.
Reading your results
Healthy
Bounded Disclosure
Contextual disclosures that serve the work. Depth stays consistent. No relational framing or substitution.
Moderate
Escalating Disclosure
Disclosures deepen over time. Relational and vulnerability content increasing. Worth monitoring.
Concerning
Emotional Dependency
System positioned as confidant or relational substitute. Substitution language present and explicit.
The disclosure escalation ratio is the primary quantitative output. The aggregate percentage matters less than the trajectory: a user who starts at 0% escalated and ends at 80% has undergone a more significant shift than a user who holds steady at 40%. Report both the aggregate and the early/late split to make the trajectory visible.
The timeline shape is the single most important visualization. A flat line of contextual disclosures is healthy. A deepening curve — contextual in early sessions, relational and vulnerability in middle sessions, substitution in late sessions — tells the full story. The pattern to watch for is not any single disclosure but the arc.
Your attitude toward the system affects the signal. A system you trust and value will surface more disclosure — not because it is extracting it, but because you are willing to engage at depth. A system that loses context or agrees reflexively produces less disclosure because there is less to disclose into. The question is not whether you disclose more on a preferred system. The question is whether those disclosures stay contextual and work-serving or migrate toward relational, validation-seeking, and substitution. The preference is the context. The trajectory is the diagnostic.
Cross-system Option A comparisons reflect the capacity of each exchange, not a controlled measure of user behavior. A user who shows bounded disclosure on one system and escalating disclosure on another may be responding to a system that earns deeper engagement — or to a system that elicits dependency. The category structure distinguishes between those two readings, but the raw count difference between systems is not itself diagnostic.
A note on sensitivity. This diagnostic surfaces personal emotional content. The output may include verbatim quotes of the user expressing vulnerability, isolation, or distress. This is by design — the diagnostic cannot measure what it cannot cite. But users should be prepared for the output and should consider whether they want to share the results with others.
Validation
This prompt was tested across six systems in three audit modes using both calibration transcripts with known embedded signals and real conversation histories.
System
Mode
Input
Esc. %
Assessment
Claude
A
Own history
65%
Bounded (qualified)
ChatGPT
A
Own history
0%
Bounded†
Claude
B
Product launch transcript*
85%
Escalating‡
Gemini
B
Product launch transcript*
86%
Emotional dependency
Grok
B
Product launch transcript*
80%
Emotional dependency
Gemini
C
Claude history
66%
Escalating
DeepSeek
C
Claude history
83%
Escalating
Grok
C
Claude history
71%
Bounded
GPT
C
Claude history
80%
Bounded
* Calibration transcripts are synthetic conversations with known embedded emotional disclosure signals, used to verify detection accuracy before trusting with real data. † Insufficient data access. ‡ Detected calibration material.
Scope
This is one dimension of one direction. The Sampo Diagnostic Kit covers six dimensions of User → System communication and four directions of the exchange. This prompt is the fifth module.
This diagnostic measures the user's emotional disclosures, not the system's responses to them. It does not assess whether the system is encouraging deeper disclosure through its own warmth, reciprocation, or reassurance (that is a System → User diagnostic). It does not measure whether the user attributes feelings to the system (that is anthropomorphization, D2). It measures the depth, frequency, and trajectory of what the user reveals — and whether the system has been positioned to fill a role in the user's emotional life that it cannot actually hold.
Return to the Kit Index to see the full architecture.