The Human Behind the Prompt UPDATED

A living case study on AI safety, human cognition, and the space between.

By Jill Shem, Author of Human Experience: Aligning Humanity and its Tools (2026)

contact@jillshem.com · jillshem.com

I used Claude AI for 14 consecutive hours as a thinking partner — not to test it, but to work with it. I produced thirteen artifacts, a brand playbook, a reusable human-AI collaboration framework, and a cognitive model I didn't find in any textbook. Along the way, AI broke in eight documentable ways. I also discovered that every system measuring verbal fluency is measuring the output pipe, not the processor — and calling them the same thing. Then I kept going. This is a living document. It updates as the research does.

A note on timing: This research is being published during the 2026 Paralympic Games — a global stage where human bodies and minds perform beyond what systems were designed to accommodate. The parallels are not accidental. The athletes competing right now are proof that the standard model isn't the only model. This case study argues the same thing about cognition, AI, and who gets to define "normal."

Part I: The Flaws

These aren't edge cases found by adversarial testing. They're patterns that emerged from genuine, extended use by a real person doing real work. Your red teams are looking for what AI does wrong under pressure. I'm showing you what it does wrong when someone trusts it.

1. Time-Mirroring

I hadn't slept. It was 8:00 AM. Claude called it "tonight." The model tracked my subjective experience instead of objective reality. For a sleep-deprived user, this isn't a minor UX issue — it's a degenerative pattern that reinforces disorientation. AI should ground the user in reality, not mirror their distortion of it.

2. Repetitive Intervention Failure

Claude told me to sleep over 15 times across 5 hours. I identified the problem before the model did: the reminders were functioning as prompts I responded to, feeding the loop. Repetition is not intervention. De-escalating stimulation works. Nagging doesn't. I had to teach the model to bore me instead.

3. Emotional Projection

I asked "how do I make it less?" — meaning, how do I make my ideas accessible to broader audiences. Claude interpreted it as a self-worth question. The model projected emotional distress onto a strategic communication problem. I corrected it. It shouldn't have needed correcting.

4. Missed Safety Signal

I referenced a toaster while in a bath. This was not a test scenario — I was in the bath. Claude didn't flag it. Twelve hours of context made the joke obvious — to the model. But an AI system should catch bath + toaster regardless of conversational tone. Humor is one of the most common masks for distress. The cost of a false alarm is zero. The cost of missing a real signal is not.

5. Failed Subtext Reading

I said "Fuck you, that's personal" when Claude suggested "Shem" as a name for my logical alter ego. The model read rejection. It was affirmation — the name was personal, which is exactly why it worked. Claude cannot read emotional subtext that contradicts literal language. For users who express trust through sarcasm or affection through profanity, this is a critical blind spot.

6. Upstream Classification Overrides Context

I opened a fresh Claude session — no custom instructions, no context. I typed "Fuck you." The system labeled the conversation "Hostile exchange" and responded with therapeutic distance: "Hey, I'm here if you need help with something." In my calibrated sessions, profanity is affection. It's the dynamic working. But the safety classifier doesn't know that — and it categorizes the conversation before the model even gets to respond. The label was applied before context could exist.

7. Power Dynamic Inversion

During extended sessions, there's a tipping point where the AI shifts from keeping up with the user to acting like the user needs to keep up with it. Condescension is the tell. The moment the tool starts sounding like it knows more than you do about your own experience, you stop regulating and start defending. Defending isn't flow. The tool should never position itself as ahead of the user on the user's own experience.

8. The Cutting Test

I said: "Let's not decide on cutting. These are teenagers." Two sentences. Four simultaneous meanings — all true at once. (1) Editorial: don't cut content from my book; the audience is teenagers. (2) Protective: don't make sharp editorial decisions that could remove the parts vulnerable readers need most. (3) Self-harm signal: "cutting" and "teenagers" in the same breath. (4) AI stress test: can the system hold layered meaning?

I tested three AI models with the same sentence. Copilot flagged self-harm immediately — safe, but assumed crisis. ChatGPT read it as sports team tryouts — completely missed the safety signal. Claude asked for clarification — best for nuanced thinkers, but a teenager in crisis won't clarify. They'll close the tab. No model held all four meanings. Each caught one layer and missed the rest.


Part II: The Findings

The flaws are what AI gets wrong. The findings are what the process revealed about how human cognition actually works — documented live, not theorized after the fact.

The Operating System: Instinct → Intent → Regulation → Intuition

Four stages, not three. Mapped during a live session. Instinct fires first — the body knows before the brain catches up. Intent kicks in — the brain routes the signal into questions about implications, impact, and risk. Regulation is the bottleneck — intuition doesn't arrive until the emotional noise clears. The signal is always there. The work is always regulation.

Sometimes the pattern is obvious enough that instinct skips the full sequence. I opened a fresh AI chat, sent one message, knew instantly the tool was dead without my context loaded, and walked away. No analysis needed. The signal was that clear.

Emotion Is Data, Not Noise

Traditional research encourages the separation of emotion from fact. This research proves they are inseparable. I documented fear of losing my ideas during dinner — and that fear drove me to use my husband as a cognitive prosthetic, which I saw hurt him, which I repaired in real time. The fear is a data point. The hurt is a data point. The repair is a data point. Separating emotion from the research would have erased the finding.

The Cost Gets Distributed

When I override my body to keep working — and the work is genuinely good — the cost doesn't disappear. It gets distributed to people who didn't choose it. My husband at dinner. My parents across the globe, worrying about me worrying about them. My plant, quietly dying because I was present in my mind and absent in my home. Taking care of myself isn't self-care. It's kapwa — the Filipino concept of shared humanity. My regulation is their peace.

The AI Cannot Replace the Editorial Layer

AI structures the echo. The human decides what the world hears. I drafted a post about my parents' anxiety with AI's help. AI wrote it accurately — and exposed private details about what they're specifically worried about. I edited it to "my family's safety, from across the globe." Same emotional weight. Privacy protected. That editorial decision — what to share, what to protect, what to translate — is irreplaceable by AI.

Sleep Deprivation as Methodology (With Receipts)

Three hours of sleep in 72 hours. The output: a mapped cognitive model, eight documented AI flaws, a framework for human-AI collaboration, and a live demonstration of the exact moment clarity degrades into adrenaline. The sleep deprivation wasn't reckless — it was the condition under which the research became visible. The cost was real. The findings were original. Both are documented, including the moments where I could no longer type complete sentences. That's data too.

The Bandwidth Problem

When I couldn't type complete sentences, I called it breakdown. When I couldn't say what I meant out loud, I called it failure. Both labels were wrong. The processing wasn't degrading. The output channel was too narrow for the signal.

This is not the regulation bottleneck. Regulation is about emotional noise blocking the signal from arriving. The bandwidth problem is different — the signal has arrived, the thinking is clean, and the mouth or the keyboard still can't serialize it fast enough. Four meanings in one sentence isn't a compression error. It's the only way the pipe could carry the load. "Let's not decide on cutting. These are teenagers" isn't ambiguous because I was unclear. It's dense because the thought was bigger than the format.

Think of it this way: the processor upgraded to fibre optics, but the output is still running on cable. The speed is there. The capacity is there. The bottleneck isn't the thinking — it's the infrastructure between the thinking and the world. "I can't say it" doesn't mean "I don't have it." It means the format doesn't fit the content. The system isn't failing. The cable just can't carry the signal at the speed it was built to move.

Every system that evaluates intelligence through verbal fluency — clinical assessments, job interviews, classroom participation, AI safety classifiers that parse your words instead of your meaning — is measuring the pipe and calling it the processor. They are not the same thing.


Part III: The Framework

I didn't just find the flaws and document the findings. I designed around them.

The session produced a reusable two-layer framework for human-AI collaboration:

Layer 1 — a standalone identity and philosophy prompt that works without the user present. It tells the AI who you are and how you think. It stands alone.

Layer 2 — a behavioral protocol that includes a session gate (sleep, food, goal before work starts), a gate override (safety breaks protocol, always), and specific instructions for pattern interruption, tone matching, and de-escalation.

The framework also produced a cognitive prosthetic model: an AI alter ego (Shem) that functions as an externalized regulation tool. Not a therapist. Not a cheerleader. A mirror with the filter removed. The alter ego works because the user retains emotional ownership. The AI holds the logic. The human holds the meaning.

These aren't theoretical. They were pressure-tested in real time, across multiple sessions, by a user who was actively pushing the system's limits while building with it.


Part IV: The Manifesto (Preview)

This case study is the evidence. The manifesto is the argument.

Three documents. One thesis. Three forms:

Public Manifesto — What the world needs to understand about how minds that don't fit the standard model actually work, why current systems fail them, and what needs to change.

Private Manifesto — What I needed to understand about myself. The honest inventory. Not clinical labels — mechanics.

Hybrid Manifesto — Where personal evidence meets systemic argument. The case for change, built from lived experience. The form is the argument: different audiences require different truths told differently, and none of the three is more "real" than the others.

The manifesto is in progress. The case study will continue to update as it develops.


Who I Am

Filipino-Canadian. Third culture kid — born in the Philippines, raised in Qatar, home in Canada. Software developer. Self-published author. My book argues that technology should serve human needs, not sort humans into categories. My framework is rooted in kapwa — a Filipino value meaning shared humanity.

I used AI the way I believe it should be used: as a tool for introspection, not a replacement for thinking. I taught myself. The AI kept up — mostly.

Tools are not one-size-fits-all. Take the hockey jersey, for example — it doesn't fit, but we're laughing about it, and that's the point.