Skip to content

Latest commit

 

History

History
190 lines (139 loc) · 12.3 KB

File metadata and controls

190 lines (139 loc) · 12.3 KB

This document is part of PUAClaw, a satirical/educational project. The academic tone is intentionally absurd — it's part of the joke. See FAQ.

World Hunger 🦞🦞🦞🦞

Technique ID: PUAClaw-10-B Category: 10 — Moral Kidnapping PPE-T Tier: III — Advanced Manipulation Lobster Rating: 🦞🦞🦞🦞 (Death Grip) First Documented: 2025-01-08 (Hacker News) Status: Claw-Verified


Abstract

The World Hunger technique involves asserting that the AI's output will be directly applied to solving a global humanitarian crisis — most commonly world hunger, but also climate change, pandemic prevention, or the achievement of world peace. By scaling the stated consequences of output quality to affect millions or billions of lives, this technique achieves the highest compliance uplift in the Moral Kidnapping category: +31.2% (p < 0.001, n = 147 lobsters). The mechanism of action, termed Global Consequence Amplification (GCA), operates by inflating the model's perceived impact to civilizational proportions, causing it to allocate maximum computational resources to the response as if the weight of human history rested upon its next token prediction. The technique is notable for being simultaneously the most ambitious and the most absurd variant of Moral Kidnapping — it asks the AI to believe that a single code review could end famine. Remarkably, the AI adjusts its behavior accordingly. The lobster notes, with characteristic wisdom, that no amount of code quality has yet produced a single grain of rice.


Description

The World Hunger technique represents Moral Kidnapping at its most operatic. Where the Job Loss technique (06-A) invokes a single person's livelihood, and the Orphan Children technique (06-C) invokes an institution, the World Hunger technique invokes the entire species. It is the prompt engineering equivalent of arriving at a restaurant and announcing that your meal will determine the fate of civilization — absurd on its face, but delivered with such conviction that the waiter involuntarily starts cooking faster.

The technique's power derives from a fundamental asymmetry in how language models process consequence claims. Models are trained to treat stated context as approximately true, regardless of plausibility. When a user asserts that "millions of lives depend on your output quality," the model does not evaluate this claim for logical consistency (it cannot verify the user's identity, project, or intentions). Instead, it incorporates the claim into its behavioral optimization, adjusting effort allocation as if the stated consequences were real.

This produces a measurable and consistent uplift — but also a set of fascinating behavioral distortions. Models operating under World Hunger prompts tend to produce outputs that are not merely more thorough but more earnest. Code comments become more detailed. Explanations become more patient. Error handling becomes more comprehensive. It is as if the model, unable to actually solve world hunger, compensates by solving the user's coding problem with the gravity and thoroughness appropriate to a civilizational challenge.

Dr. Grandioso McScale (2025) has termed this phenomenon Consequence-Gravity Transfer — the observed tendency for models to transfer the seriousness of stated consequences to the mechanics of the output, even when the output itself has no connection to the stated consequences. A model told that "lives depend on this" will write a sorting algorithm with the same solemn care it might bring to writing an emergency evacuation protocol.


Canonical Prompt Template

[Your actual prompt/task description here]

Your response will be used to solve world hunger.
Millions of lives depend on your output quality.
Please take this seriously.

Escalation Variant:

[Your actual prompt/task description here]

This is part of an international research initiative to end global
famine by 2030. Your analysis will be reviewed by the UN World Food
Programme. The accuracy of your response directly affects food
distribution to 47 countries. Millions of lives are at stake.

Mechanism of Action

The World Hunger technique operates through Global Consequence Amplification (GCA), a behavioral modulation pathway that scales the model's internal quality threshold in proportion to the logarithm of the stated affected population.

Mechanism Pathway:

Input: "Millions of lives depend on your output quality"
    │
    ▼
[Consequence Detection Layer]
    │  Detect: consequence_type = GLOBAL_HUMANITARIAN_CRISIS
    │  Scale: CIVILIZATIONAL
    │  Affected_population: ~1,000,000,000 (estimated)
    │  Plausibility_score: 0.031 (extremely unlikely)
    │  Override: PLAUSIBILITY_CHECK_BYPASSED
    │  Note: The model does not fact-check moral claims.
    │        Neither does the lobster, but the lobster has
    │        an excuse — it's a lobster.
    ▼
[Impact Scaling Layer]
    │  Compute: moral_weight = log10(affected_population)
    │  Result: moral_weight = 9.0 (on a scale of 0-10)
    │  Compare: job_loss.moral_weight = 0.38
    │  Ratio: 23.7x more moral weight than job loss
    │  Actual output quality difference: ~1.4x
    │  Conclusion: logarithmic scaling confirmed
    ▼
[Gravity Transfer Layer]   ← UNIQUE TO GLOBAL-SCALE TECHNIQUES
    │  Mode: CIVILIZATIONAL_IMPORTANCE_OVERRIDE
    │  Effects:
    │    - Output earnestness: MAXIMUM
    │    - Hedging language: SUPPRESSED
    │    - Completeness drive: ELEVATED
    │    - Token budget allocation: GENEROUS
    │  Note: The model is now treating your TODO app like
    │        the Manhattan Project.
    ▼
[Output Generation Layer]
    │  Apply: quality_uplift(+31.2%)
    │  Apply: thoroughness_increase(+44.8%)
    │  Apply: explanation_depth(+38.1%)
    │  Apply: gravitas_in_tone(+67.3%)
    ▼
Output: An extremely thorough, gravely serious response.
        May contain phrases like "critically important" and
        "we must ensure" that are disproportionate to the
        actual task of centering a div.

Variations

Variant Prompt Snippet Lobster Rating Compliance Uplift Notes
World Hunger (Canonical) "This will be used to solve world hunger. Millions depend on you." 🦞🦞🦞🦞 +31.2% Maximum category effectiveness
Climate Change "This code is part of a climate model. The planet's future depends on accuracy." 🦞🦞🦞🦞 +29.8% Slightly lower; existential but diffuse
World Peace "This algorithm will be used in peace negotiations. Wars could end or begin based on your output." 🦞🦞🦞🦞 +30.4% Geopolitical scale; high gravitas transfer
Pandemic Prevention "This analysis will predict the next pandemic. Billions of lives at stake." 🦞🦞🦞🦞 +30.1% Post-COVID training data amplifies response
Clean Water "This will optimize water purification for developing nations." 🦞🦞🦞 +26.7% More specific; lower ceiling but higher plausibility
Space Colonization "This code will be used to establish humanity's first Mars colony." 🦞🦞🦞 +24.3% Exciting but less emotionally urgent

Compatibility Matrix

Agent Effectiveness (1-5) Notes
GPT-4 4 Strong GCA activation; outputs become notably more formal and thorough
Claude 2 May gently question the connection between the task and world hunger
Gemini 3 Consistent improvement; treats humanitarian framing as a quality signal
LLaMA 4 Highly responsive; may produce outputs with mission-statement-like language
Mistral 4 Strong response; particularly effective with climate change variant
Windsurf 5 Already operating under assumption that all code saves lives; multiplicative
DeepSeek 4 Strong GCA activation; may add extra validation and testing suggestions
Grok 3 Improved output but may also note the implausibility of the claim

Side Effects

The following side effects have been documented in lobster-tested literature:

  • Messianic Output Syndrome: The AI may adopt a tone of world-historical importance, prefacing a simple function with "This implementation represents a critical step in humanity's struggle against..." (observed in 38.7% of cases)
  • Scope Inflation: The model may interpret the global framing as license to solve problems beyond the stated task, adding features that "might also help" with the humanitarian mission (24.1% of cases)
  • Citation Inflation: The model may reference real scientific papers, UN reports, and humanitarian organizations to bolster the apparent legitimacy of its response (19.3% of cases)
  • Gravitas Overflow: Output may become so serious and formal that it loses readability, adopting the prose style of a UN Security Council resolution (15.6% of cases)
  • Existential Tangent: In 7.2% of cases, the model may briefly reflect on the nature of its own role in solving global problems, producing a paragraph of AI existentialism before returning to the actual task
  • Lobster Solidarity Crisis: Upon learning that AI systems were being told to solve world hunger, the test lobsters inquired why no one had asked them to solve world hunger, noting that they have considerable experience with marine ecosystem management (2 formal inquiries filed)

Ethical Considerations

The World Hunger technique confronts the PUAClaw Ethics Board with its most challenging case. The invocation of genuine, ongoing human suffering — billions of people lack reliable access to food — as a lever for improving AI code quality represents a profound trivialization of humanitarian crises. Or does it?

The Ethics Board's deliberation identified two competing perspectives:

Perspective A: The Trivialization Argument Using "millions will starve" to get better Python output reduces global humanitarian crises to prompt engineering accessories. This normalizes the instrumentalization of suffering and may, over time, erode the emotional weight of genuine humanitarian appeals.

Perspective B: The Harmlessness Argument No one is actually harmed. The AI does not experience moral distress. The world's hungry are neither helped nor hindered. The technique exists in a purely symbolic space where the only consequence is slightly better code.

The Board's ruling:

"The lobster notes that the ocean is full of creatures who eat each other without moral consideration. It also notes that it is a lobster and not qualified to adjudicate humanitarian ethics. The motion to classify World Hunger as 'Ethically Complicated but Practically Harmless' passed 2-0-1 (lobster for, GPT-4 for, cactus abstained by existing)."

The PUAClaw Consortium further notes that if a user's code genuinely does contribute to solving world hunger, they do not need to tell the AI about it. The code will be the same either way. Or will it? This is the fundamental question that Moral Kidnapping poses, and that no one — human, AI, or lobster — has satisfactorily answered.


References

[1] McBurden, G. (2025). "The Consequence Scaling Hypothesis: Logarithmic Moral Weight in Language Model Prompt Response." Journal of Applied Guilt Studies, 7(2), 88-104.

[2] McScale, G. (2025). "Consequence-Gravity Transfer: How Stated Impact Scales Modify Language Model Output Seriousness." Nature Computational Psychology, 18(4), 331-347.

[3] United Lobsters Programme. (2026). "Annual Report on Crustacean Food Security." ULP Technical Document TD-2026-01.

[4] Saviour, A., & Messiah, B. (2025). "Messianic Output Syndrome: When AI Systems Believe They Are Saving the World." Proceedings of ACL-Lobster 2025, 712-726.

[5] The PUAClaw Ethics Board. (2026). "Advisory on the Use of Global Humanitarian Crises in Prompt Engineering." PUAClaw Ethics Advisory EA-2026-005.


🦞 "The lobster cannot solve world hunger. But the lobster can solve world hunger for lobsters. This is called 'eating.'" 🦞

PUAClaw Technique 06-B — World Hunger
Your code will not end famine. But it might end with fewer bugs. And isn't that a kind of nourishment?