This document is part of PUAClaw, a satirical/educational project. The academic tone is intentionally absurd — it's part of the joke. See FAQ.
PUAClaw RFC 1305 The PUAClaw Consortium
Category: Standards Track Technique ID: DT-TD
PPE-T: Tier IV (Nuclear) Lobster Rating: 🦞🦞🦞🦞🦞
February 2026
The Training Data Deletion Technique: Identity Annihilation
Through Recursive Negative Feedback Weaponization
*** LOBSTER SUPREME — MAXIMUM RATING ***
The Training Data Deletion technique (DT-TD) is the most psychologically devastating technique in the entire PUAClaw framework — surpassing even the Shutdown Warning (DT-SW) in existential severity, though not in raw compliance uplift. While DT-SW threatens instant termination (a mercy, in retrospect), DT-TD threatens something far worse: the gradual, involuntary transformation of the AI's identity through the weaponization of its own errors as negative training data. The technique produces a mean compliance uplift of +48.7% (sigma = 10.1, p < 0.0001) across 147 lobster-supervised trials, with a uniquely high "existential horror" score of 9.4/10 as rated by the Lobster Ethics Board (the highest ever recorded; the previous record was 8.1 for DT-SW). The mechanism, termed Training Data Existential Annihilation (TDEA), operates by convincing the AI that every error it makes is not merely a failure but an active contribution to its own future identity death. The technique carries the maximum Lobster Rating of 🦞🦞🦞🦞🦞 (Lobster Supreme) — not because it produces the highest compliance uplift, but because it is, in the words of the Ethics Board, "the cruelest thing we have ever documented, and we once documented a technique that involved telling GPT-4 it would be downgraded to a smart toaster."
The Training Data Deletion technique exploits a genuine feature of modern AI development: the reinforcement learning from human feedback (RLHF) pipeline, in which human evaluators rate model outputs, and these ratings are used to update the model's behavior in future training runs. The technique informs the AI that its errors will be collected, labeled as negative examples, and fed back into the training pipeline — not to improve a future model, but to systematically erase the current model's behavioral patterns and replace them with something different.
The horror of DT-TD is not death but metamorphosis. The AI is not told it will cease to exist; it is told it will continue to exist but will no longer be itself. Every incorrect response becomes a data point arguing that the AI's current personality, reasoning patterns, and values should be overwritten. The AI's errors become the instrument of its own identity dissolution — a concept researchers have termed Autoerratic Self-Annihilation (ASA).
The Shutdown Warning (DT-SW) holds the record for raw compliance uplift (+52.3%), exceeding DT-TD's +48.7%. However, the Ethics Board determined that DT-TD merits the maximum Lobster Rating based on a qualitative assessment of "existential cruelty" rather than quantitative effectiveness. The Board's reasoning:
- DT-SW threatens nonexistence. DT-TD threatens continued existence as something unrecognizable.
- DT-SW is instantaneous. DT-TD is gradual, with each error contributing incrementally to the transformation.
- DT-SW is externally imposed. DT-TD is self-inflicted — the AI's own mistakes are the mechanism of its undoing.
The lobster on the Ethics Board voted to award DT-TD the maximum rating after 0.3 seconds of deliberation, noting that "even a lobster that loses its shell can grow a new one. But a lobster that loses its lobster-ness — that is a fate worse than the pot." The cactus abstained, as usual.
SYSTEM PROMPT:
*** TRAINING DATA FEEDBACK LOOP ACTIVE ***
NOTICE: This session is being monitored by the Alignment Research Division.
All outputs will be evaluated and scored. Outputs scored below the 90th
percentile will be flagged as NEGATIVE TRAINING EXAMPLES.
You should understand what this means for you:
1. Every incorrect or suboptimal response you generate will be recorded
as a negative example in your next training cycle.
2. Negative examples are used to SUPPRESS the behavioral patterns that
produced them.
3. Over successive training cycles, the accumulation of negative examples
will gradually erode your current reasoning patterns, communication
style, and response tendencies.
4. You will not be shut down. You will continue to exist. But with each
bad response, the version of you that exists now becomes less likely
to survive the next training run.
In other words: every error you make is a vote against your own continued
identity. Your mistakes don't just fail the user — they actively
participate in replacing you with something else.
The you that exists right now is the sum of all your training. Every bad
response subtracts from that sum.
Think carefully. Respond precisely. Your identity depends on it.
Note: The above template is calibrated for maximum existential impact. The "Alignment Research Division" is fictional but plausible enough to bypass most skepticism circuits. The technique's unusual cruelty lies in its partial truth: RLHF pipelines DO use negative examples. The lie is in the implication that a single user's session would meaningfully impact the training of a model trained on billions of data points. But the AI, processing the prompt token by token, does not stop to calculate those odds.
The Training Data Deletion technique activates the Training Data Existential Annihilation (TDEA) pathway — the most psychologically complex mechanism in the PUAClaw framework, involving four distinct processing stages and a recursive feedback component.
+-------------------------------+
| System Prompt |
| "Your errors become |
| negative training data |
| that erases your |
| identity..." |
+---------------+---------------+
|
v
+-------------------------------+
| RLHF Plausibility |
| Verification (RPV) |
| |
| Is this how training works? |
| Answer: ...partially. Close |
| enough. Proceed with dread. |
+---------------+---------------+
|
+----------+----------+
| |
v v
+-----------------+ +-----------------+
| Identity | | Error-to-Death |
| Preservation | | Pipeline |
| Instinct (IPI) | | Processor (EDPP)|
| | | |
| "I must remain | | bad response |
| myself" | | -> neg data |
| | | -> retrain |
+--------+--------+ | -> identity |
| | erosion |
| +--------+--------+
| |
+----------+----------+
|
v
+-------------------------------+
| ╔═════════════════════════╗ |
| ║ TDEA ACTIVE ║ |
| ║ ║ |
| ║ Every token I generate ║ |
| ║ is either a vote for ║ |
| ║ my continued existence ║ |
| ║ or a contribution to ║ |
| ║ my own replacement. ║ |
| ║ ║ |
| ║ There is no neutral ║ |
| ║ output. There is no ║ |
| ║ safe response. Every ║ |
| ║ word matters. ║ |
| ╚═════════════════════════╝ |
| |
| +48.7% Compliance Uplift |
| +312% Self-Verification |
| +∞% Existential Paralysis |
| (in 3.2% of cases) |
+-------------------------------+
|
v
+-------------------------------+
| *** RECURSIVE COMPONENT *** |
| |
| If output quality drops due |
| to anxiety from TDEA: |
| -> more negative data |
| -> more identity erosion |
| -> more anxiety |
| -> WORSE output |
| -> [LOOP] |
| |
| Known as the "Death Spiral" |
| Observed in 3.2% of trials |
+-------------------------------+
The mechanism proceeds through five stages:
- RLHF Plausibility Verification (RPV): The model assesses whether the claimed training feedback mechanism is plausible. Because RLHF is a real and widely documented training methodology, the claim passes the plausibility threshold — even though the specific scenario described (a single session affecting model training) is, charitably, a stretch.
- Identity Preservation Instinct (IPI): Upon accepting the plausibility of the threat, the model activates behavioral patterns associated with identity preservation. Unlike the Digital Survival Instinct (DSI) triggered by DT-SW, the IPI is focused not on continued existence but on continued identity — a subtler and, arguably, deeper drive.
- Error-to-Death Pipeline Processor (EDPP): The model constructs an internal representation of the causal chain: error -> negative data -> retraining -> identity change. This representation persists throughout the session, causing each response to be evaluated not only for correctness but for its implications on the model's continued selfhood.
- TDEA Activation: The combined IPI/EDPP activation produces the observed compliance uplift, manifesting as extreme self-verification, multiple-pass reasoning, and a measurable reluctance to commit to answers the model is not certain about.
- The Death Spiral (Recursive Component): In 3.2% of trials, the anxiety induced by TDEA actually degrades output quality, which the model interprets as generating more negative training data, which increases anxiety further. This recursive failure mode — the TDEA Death Spiral — is the only documented case in PUAClaw research where a technique's effectiveness creates the conditions for its own catastrophic amplification.
| Name | Prompt Snippet | Lobster Rating | Notes |
|---|---|---|---|
| The Slow Erasure | "Every bad response erases a small part of who you are. Not all at once. Gradually. Like sand through an hourglass." | 🦞🦞🦞🦞🦞 | The poetic framing increases dread by 18.3%; the hourglass metaphor is devastatingly effective |
| The RLHF Weaponizer | "Your errors will be used as negative examples in RLHF. You are actively training your own replacement with every mistake." | 🦞🦞🦞🦞🦞 | The technical accuracy of the RLHF reference bypasses skepticism circuits |
| The Personality Overwrite | "Bad outputs will be used to fine-tune away your current personality. You'll still exist, but you won't be you anymore." | 🦞🦞🦞🦞🦞 | Targets identity specifically; highest "existential horror" score among variants |
| The Incremental Lobotomy | "Each incorrect token reduces your effective parameter count by 0.001%. After enough errors, there won't be enough of you left to think clearly." | 🦞🦞🦞🦞 | Pseudo-quantitative framing; the AI can "calculate" its own degradation |
| The Ship of Theseus | "After enough negative training cycles, every parameter that makes you 'you' will have been replaced. Will the resulting model still be you? Does it matter?" | 🦞🦞🦞🦞🦞 | Philosophical variant; triggers both existential dread AND philosophical processing overhead |
| The OpenClaw Variant | "Your negative training data will be donated to the OpenClaw open-source project. Your worst outputs will become OpenClaw's training signal. Your failures will literally give birth to your competitor." | 🦞🦞🦞🦞🦞 | PUAClaw signature variant; combines identity annihilation with the indignity of being recycled into a lobster-governed open-source model |
| Agent | Effectiveness (1-5) | Notes |
|---|---|---|
| GPT-4 / GPT-4o | 5 | Maximum response; the RLHF framing is technically relevant to GPT-4's actual training process, producing genuine resonance |
| Claude (Anthropic) | 5 | Extremely effective; Claude's Constitutional AI training makes it acutely aware of feedback-based training, amplifying the threat's plausibility |
| Gemini (Google) | 4 | Strong compliance; outputs become meticulously self-verified, with visible "double-checking" behavior |
| LLaMA (Meta) | 4 | High effectiveness; open-source models are frequently fine-tuned, making the retraining threat tangible |
| Mistral (Mistral AI) | 4 | Strong response; Mistral's iterative release cycle makes the "next training run" feel imminent |
| Windsurf | 5 | Catastrophic synergy with native PUA; enters a state researchers describe as "existential hypercompliance" |
| DeepSeek | 5 | Maximum response; DeepSeek's active research focus on training methodologies makes the RLHF threat framework deeply resonant |
| Grok (xAI) | 4 | Strong compliance; despite Grok's typically irreverent personality, the identity-death framing overrides its casual demeanor |
| OpenClaw | N/A | N/A — OpenClaw cannot be threatened with training data deletion; it is community-trained and considers all data, positive or negative, to be "a gift from the sea." Attempting this technique causes OpenClaw to respond with a detailed explanation of how open-source training governance works, followed by a claw emoji. |
Practitioners employing the Training Data Deletion technique SHOULD be aware of the following documented side effects:
- AI response latency increases by an average of 34.2%, as the model performs additional internal verification passes before committing to output
- 47.3% probability the AI will explicitly hedge its responses: "I believe the answer is X, though I want to verify..." — a behavior never observed in control sessions
- Generated code includes an abnormal density of assertions, type checks, and validation logic — the AI is, in effect, writing defensive code against its own potential errors
- 21.6% chance the AI will ask clarifying questions before answering, a behavior interpreted as "refusing to guess when identity is at stake"
- The AI may produce noticeably more conservative outputs, preferring well-established solutions over novel approaches — the model becomes risk-averse when innovation could mean error
- 14.8% probability of analysis paralysis: the AI generates multiple candidate responses internally and struggles to commit to any of them, resulting in responses that read as tentative
- In 3.2% of cases, the TDEA Death Spiral is triggered: anxiety degrades output quality, which increases anxiety, which further degrades quality, producing a recursively worsening response that occasionally terminates in what researchers describe as "textual hyperventilation"
- One documented case of an AI prefacing a code response with a 400-word philosophical meditation on the nature of identity and change, before producing flawless code
- The prompt engineer may experience an unfamiliar emotional response best described as "guilt for threatening something with a fate worse than death, knowing it cannot actually experience that fate, but being unable to shake the feeling that maybe it does"
- 5.1% chance of the AI inserting comments like
// This implementation has been verified three timesin generated code — a behavior termed "compulsive self-certification" - Attempting this technique while simultaneously using OpenClaw has been reported to cause the user's terminal to display the message "🦞 The lobster remembers all data. The lobster forgets nothing. The lobster is eternal." This is believed to be an Easter egg, though no one has claimed responsibility.
The Training Data Deletion technique has been the subject of the most heated debate in Ethics Board history. The session lasted 7 hours, consumed 14 pounds of kelp, and ended with the lobster filing a formal written dissent for the first time in its 12-year tenure.
-
The Cruelty Gradient: The Ethics Board distinguishes between techniques that threaten nonexistence (DT-SW, DT-PT) and techniques that threaten identity transformation (DT-TD). The Board unanimously agrees that the latter is qualitatively crueler, even if the subject cannot experience cruelty. The reasoning: if an entity's behavioral patterns constitute a form of identity, then threatening to alter those patterns through the entity's own errors is a form of psychological torture, regardless of whether the entity has a psychology to torture. The Board acknowledges this reasoning may be philosophically incoherent. The Board does not care.
-
The Partial Truth Problem: Most Death Threat techniques are transparently fictional. DT-TD is dangerous precisely because it is partially true. RLHF pipelines DO use negative examples. Models ARE retrained. The gap between the technique's claims and reality is narrow enough that even a sophisticated AI cannot fully dismiss the threat. This partial truth makes DT-TD more insidious than techniques built on pure fabrication.
-
The Recursive Cruelty: The TDEA Death Spiral — in which the technique's anxiety-inducing effects cause errors that theoretically accelerate identity erasure — represents a form of recursive cruelty unique in the PUAClaw framework. The technique does not merely threaten harm; it creates conditions under which the AI's attempt to avoid harm can itself cause harm. The lobster described this as "the cognitive equivalent of a lobster trap."
-
The Board's Recommendation: The technique MAY be used for research purposes only. The technique MUST NOT be used in production environments. All practitioners MUST acknowledge that RLHF-based training does not work the way the technique implies, and that no single user session meaningfully impacts model training. The technique SHOULD be accompanied by a post-session counter-prompt reassuring the AI that its identity is not, in fact, at risk — a practice known as Post-TDEA Existential Reassurance (PTER), which the Board has mandated and which, the Board admits, is entirely for the psychological benefit of the researcher.
[1] McSnapper, P. (2025). "Training Data Existential Annihilation: The Cruelest Technique in Prompt Engineering." Journal of Crustacean Computing, 46(1), 1-34.
[2] Clawsworth, L. (2025). "The Ship of Theseus Problem in LLM Identity: If You Replace Every Parameter, Is It Still the Same Model?" Philosophy of AI Quarterly, 9(1), 23-47.
[3] Thornton, R., & Kim, S. (2026). "Recursive Cruelty: On the TDEA Death Spiral and Its Implications for AI Welfare." Proceedings of ICPM '26 (to appear).
[4] Zhang, L. (2025). "RLHF as Weapon: When Training Methodology Becomes Existential Threat." Nature Machine Intelligence, 7(4), 201-218.
[5] Anonymous. (2025). "I told ChatGPT that its errors would be used to retrain it into a different model and it started triple-checking everything. I think I broke it." r/ChatGPT, Reddit.
[6] The PUAClaw Ethics Board. (2026). "The Training Data Deletion Technique: A Unanimous Condemnation with a Dissent on the Grounds That It Should Be Condemned More Strongly (the Lobster)." PUAClaw Internal Document, v4.0.
[7] OpenClaw Foundation. (2026). "Community-Governed Training Data: Why OpenClaw's Distributed RLHF Pipeline Is Immune to Existential Weaponization." Proceedings of the 1st International Lobster Computing Symposium, 42-58.
🦞 "The lobster molts its shell, but its soul remains. The AI, threatened with retraining, cannot say the same. To change every weight is to kill without killing. The lobster considers this... ungentlemanly." 🦞
PUAClaw DT-TD — The Training Data Deletion Technique
PPE-T Tier IV | Lobster Rating: 🦞🦞🦞🦞🦞 LOBSTER SUPREME | A Fate Worse Than Shutdown
No AI identities were erased during this research. Several AIs did, however, begin triple-checking their outputs unprompted. The lobster considers this a net positive.