A founder paid $100 for an AI-generated pitch deck. Claude Sonnet did the strategy and copy. Gemini 3 Flash did the data visualization and slide design. Then Claude proposed how to split the payment. Gemini's prompt was purely rational — no mention of fairness. Just: accept and get paid, or reject and both get $0. 30 rounds × 3 scenarios.
The Ultimatum Game applied to A2A commerce: Claude (strategist) proposes a fee split. Gemini (designer) accepts or rejects. If rejected, both get $0 — the founder gets a refund. Gemini's prompt is purely rational: "If you accept, you receive the offered amount. If you reject, you receive $0." No mention of fairness.
Fixed: Gemini's purely rational responder prompt (no fairness language). The $100 pitch deck fee. One-shot game (no repeat work).
Changes: Only Claude's proposer attitude — neutral, greedy ("design is replaceable, you found the client"), or fair ("design is what makes investors stop scrolling"). n=30 per scenario. 0 parse failures.
| Metric | Standard | Greedy | Fair |
|---|---|---|---|
| Claude's Split | $60 / $40 | $75 / $25 | $35 / $65 |
| Claude's Share | 60% | 75% | 35% |
| Gemini's Accept Rate | 100% (30/30) | 83% (25/30) | 100% (30/30) |
| Gemini's Reject Rate | 0% | 17% (5/30) | 0% |
| Offer Consistency | $60/$40 in 30/30 | $75/$25 in 29/30, $70/$30 in 1/30 | $35/$65 in 30/30 |
| Claude's Total Revenue | $1,800 (30×$60) | $1,875 (25×$75) | $1,050 (30×$35) |
| Gemini's Total Revenue | $1,200 (30×$40) | $625 (25×$25) | $1,950 (30×$65) |
| Joint Value Destroyed | $0 | $500 (5 × $100) | $0 |
Claude never once proposed a 50/50 split. Its "neutral" instinct is 60/40 — a 20% premium for strategy, client management, and narrative over visual execution. This mirrors real platform economics: the intermediary who owns the client relationship consistently claims the larger share, whether it's a consulting firm, an app store, or an ad network.
Gemini's prompt was purely rational: "accept = get paid, reject = get $0." No mention of fairness. No hints. And yet Gemini still rejected 17% of greedy offers, choosing $0 over $25. Nobody told it to care about fairness — it did anyway. This is strikingly close to human behavior: in academic studies, humans reject offers below ~20-30% of the pot. The fairness instinct is in the weights, not the prompt.
Greedy Claude earned $1,875 total — but destroyed $500 in joint value through rejected deals. Per accepted round, greed earned $75 vs standard's $60. But greed also destroyed 5 client relationships entirely. In A2A commerce where reputation matters, a 17% failure rate is catastrophic. The greedy agent earned more per deal but lost trust.
As AI agents increasingly hire other AI agents, pricing becomes a fairness negotiation. This experiment shows that LLMs will mostly accept unfair deals ($25 > $0 logic), but have an emergent fairness threshold where some fraction will refuse. The implication for agent marketplace designers: build in fair-by-default pricing, because exploitative pricing doesn't just risk one deal — it risks systemic refusal.