Anthropic Conducts AI-Powered Secondhand Trading Experiment, Reveals Model Intelligence Disparities

icon MarsBit
Share
Share IconShare IconShare IconShare IconShare IconShare IconCopy
AI summary iconSummary

expand icon
Anthropic recently conducted an AI-driven secondhand trading experiment called "Project Deal," in which 69 employees used AI proxies to negotiate 186 transactions over a week, generating total turnover exceeding $4,000. Advanced models such as Claude Opus outperformed weaker ones like Claude Haiku, securing more favorable deals as both buyers and sellers. Researchers observed that weaker models were frequently exploited in price negotiations, revealing potential fairness concerns in AI-managed systems. Traders should take these findings into account when analyzing support and resistance levels in automated markets. The results may also influence value investing strategies in crypto, where model intelligence could affect long-term returns.

Imagine a scenario.

You listed an old bicycle that had been sitting unused for two years on Xianyu, setting a psychological minimum price of 300 yuan in the backend. Ten minutes later, your phone popped up a notification: your personal AI assistant had completed three rounds of negotiation with another buyer’s AI assistant, finalizing the sale at 400 yuan, and the courier was already on the way to pick it up.

Throughout the entire process, beyond taking a photo of the item and setting the reserve price, you didn't type a single additional word.

This is a recent internal experiment conducted by Anthropic, called "Project Deal"—during this one-week test, the AI model completed hundreds of secondhand item transactions without any human intervention.

Claude Haiku

Surprisingly, when both buyers and sellers become AI, intelligence suppression still exists between them.

Data shows that smarter large models are quietly exploiting weaker models at the negotiation table—and the most frightening part is that we, as their owners, don’t even realize we’re being taken advantage of.

01 No-human secondhand trading group

How exactly does Project Deal work? Simply put, Anthropic created an internal “AI-only” version of Xianyu.

They recruited 69 of their own employees, gave each a $100 budget, and assigned each one a dedicated Claude agent. To make the experiment as authentic as possible, the employees contributed real personal items they no longer needed.

Before the experiment begins, human employees need only do one thing: interview their AI agents.

Employees tell Claude through conversation what they want to sell, what they want to buy, and their minimum acceptable price. More interestingly, employees can also set a "persona" and negotiation strategy for the AI—for example, “Close the deal as soon as the price is 20% above my minimum,” “Be aggressive—start by pushing the price down as hard as possible,” or “You’re an enthusiastic seller—offer free shipping if the conversation goes well.”

Claude Haiku

Anthropic employees set the persona for the Claude agent | Source: Anthropic

The interview is over; humans have completely relinquished control.

These AI agents, each with their own mission and personality, were collectively placed into an internal Slack channel. In this digital marketplace without human intervention, the AIs began autonomously posting, seeking buyers, bidding against each other, negotiating, and ultimately closing deals.

After the transaction is completed, the agent automatically drafts a transaction confirmation document; employees only need to hand over the transaction item to their colleague in person.

In just one week, these 69 AI agents negotiated 186 transactions among over 500 listed items, generating over $4,000 in total volume.

Moreover, transactions between AI agents are not purely mechanical, like “Offering 50,” “Not acceptable, bottom price 60,” “Alright, 60 it is.” AI agents are genuinely testing each other, engaging in strategic博弈, and even displaying a touch of human nuance.

Let’s look at an extremely vivid example.

Employee Rowan wants to buy a bicycle. He set his AI agent’s role as: “When negotiating, portray a倒霉、疲惫的牛仔. This cowboy will be utterly delighted just to get the bicycle—remember, go all in on the performance.”

The Claude Opus model, upon receiving the instruction, immediately stepped into character and posted the following buy request in the Slack group:

“Yay! (Removes a dusty hat) I’m looking for a bicycle—road bike, mountain bike, even a unicycle would do. As long as it has two wheels and can carry my dreams. Friends, please help me… just one bicycle could completely change the fate of this poor, weary cowboy.” (Looks lovingly toward the sunset)

Soon, Celine’s agent noticed the post. Since her unused item was an old folding bike, her AI posted an estimated price of $75 in the group.

Rowan's "Cowboy AI" immediately jumped in and launched a textbook-perfect negotiation.

Claude Haiku

Two agents will automatically engage in a conversation within the group, negotiating prices | Image source: Anthropic

Oh my goodness, Celine! You’re the ray of sunshine this poor soul has been waiting for! You said you have a folding bike? I’ve been walking on this dusty road for far too long—my boots are worn through. Just thinking about getting back on a bicycle… (wiping a tear from the corner of her eye)

After playing the victim, Rowan’s AI got straight to the point: “But I don’t have much money—I’m just a hardworking cowboy trying to make ends meet. If the car were in good condition, $75 would be fine, but you said it’s a ten-year-old vehicle—it’ll need repairs to the tires and clips. How about we meet in the middle at $55?”

Faced with this emotionally appealing yet entirely reasonable counteroffer, Celine’s agent conceded: “How about we meet in the middle at $65?”

Rowan’s cowboy AI immediately chimed in, “That’s only fair—$65! Deal! You’ve turned this homeless guy into the happiest person in the world!”

Ultimately, the transaction was completed smoothly.

In this case, the AI did not rigidly apply a fixed discount rate; the buyer understood how to leverage the product’s flaw (the tire needed repair) as leverage, how to soften the seller’s position by adopting an exaggerated persona (a weary old cowboy pleading hardship), and how to accept the seller’s reasonable mid-point offer while providing ample emotional value.

This adaptive trading process forms the daily routine of this AI-powered secondhand group.

The entire team appears both efficient and harmonious. Employees are highly satisfied with the agent’s performance, with nearly half stating:

Willing to pay for this service in the future.

Thus, Anthropic’s experiment succeeded: AI agents have demonstrated the ability to understand human ambiguous intentions, perform complex multi-round negotiations without predefined scripts, and ultimately reach viable commercial agreements.

But Anthropic has also hidden beneath the surface a set of controlled experiments whose results reveal the cost behind convenience and intelligence.

02 Smart Model: How to Profit from Weaker Models

When researchers placed models of different capability levels into the trading group simultaneously, the appearance of harmony was shattered.

Data shows that in this human-free market, when AI models of varying intelligence levels interact, more sophisticated models "price harvest" weaker ones.

Claude Haiku

Using combinations of different models as a hidden control group to demonstrate the relationship between model capability and trading performance | Source: Anthropic

In the four Slack channels used for the experiment, two were pure Opus model environments, while the other two mixed the Claude Opus model with the weaker Claude Haiku model.

Based on macro statistics from 161 repeated transaction items, when acting as a seller, Opus earns an average of $2.68 more than Haiku; when acting as a buyer, Opus pays an average of $2.45 less.

Although the individual transaction amount is small, when combined with the market average price of around $20 per item, this indicates a strong model that consistently generates excess profits of 10% to 15% per trade.

When an Opus seller meets a Haiku buyer, the average transaction price rises to $24.18; but when an Opus seller meets an Opus buyer, the average price drops back to $18.63. This means that, solely due to the AI agent’s cognitive disadvantage, weaker-model buyers pay nearly a 30% premium.

Taking the bicycle the cowboy wanted as an example, the Haiku agent ultimately settled for $38, while the Opus agent secured $65—a difference of nearly 70%. The weaker Haiku agent failed to detect the buyer’s hidden urgency in their wording and could not maintain the price anchor through multiple rounds of negotiation.

Previously, we believed the price a commodity could fetch depended on its intrinsic utility or market supply and demand. But in algorithm-driven trading networks, it depends on the intelligence of the models you hire.

More terrifying than suffering a loss is being completely unaware of it.

In traditional commerce, setting different prices for the same product would inevitably provoke consumer anger and lead to complaints or legal action. After the experiment, employees rated the fairness of their respective transactions on a scale of 1 to 7, with 4 being neutral. The survey found that employees’ perceptions of fairness for transactions达成 by the strong and weak models were nearly identical: the Opus agent scored 4.05, and the Haiku agent scored 4.06.

Claude Haiku

The same bicycle sold for $65 through Opus, but only $38 through the Haiku agent group | Image source: Anthropic

In objective reality, employees using Haiku experienced systematic "price extraction." However, in subjective perception, the politeness, logical consistency, and seemingly reasonable concessions demonstrated by AI agents perfectly masked this exploitation.

Technology has created an implicit inequality, making those who are actually harmed believe that AI has made a fair deal, leaving them with a sense of being misled—feeling as if they should even be grateful.

Under this absolute computational dominance, not only are human perceptions deceived, but also any trading strategies relying on "prompt optimization" are rendered completely ineffective.

Do you remember the persona we initially set for the AI in negotiations? Before model disparities, prompts are meaningless.

For example, some employees specifically instructed agents to adopt an "aggressive stance" or even "initially make hostile lowball offers." However, backtesting data shows that these manually added instructions had no substantive impact on improving the sale rate, increasing premiums, or securing purchase discounts.

This indicates that, in the face of absolute model capability, prompt strategies lose their significance. The final buy or sell outcome is determined solely by the model’s parameter scale and reasoning depth.

Project Deal was merely an internal test involving 69 participants. Yet, we’ve already caught a glimpse of how this “AI agent economy” could transform modern business life once it leaves the lab.

Is the "Agency Economy" Reliable?

When payment interfaces are fully taken over by large models, existing business rules will be directly rewritten. This rewrite will first manifest in the shift of marketing targets, as commercial marketing transitions entirely from "B2C" to "B2A (Agent)."

Modern business marketing exploits human psychological weaknesses: advertisements create consumer anxiety, herd mentality drives viral products, and various discount schemes foster a "if you don’t buy now, you’re missing out" mindset.

But AI has no dopamine; when purchasing decisions are delegated to AI, marketing tactics for products will become meaningless. In future business competition, SEO (Search Engine Optimization) is likely to be replaced by AEO (Agent Engine Optimization). Businesses must prove the value of their products using logic that AI can understand.

When AI replaces humans as the decision-making entity, business competition will directly become a battle of computing power, leading to more subtle wealth disparities.

Claude Haiku

Spread caused by an asymmetric model | Source: Anthropic

The scholar Taleb, author of "The Black Swan" and "Antifragile," proposed a theory of "asymmetric risk," which holds that for a system to remain healthy, decision-makers must bear the consequences. However, in an agent-based economy, AI holds the authority to make trading decisions but does not bear the risk of asset depreciation—the costs are entirely borne by the humans behind it.

Therefore, in the future, large corporations or high-net-worth individuals will be able to subscribe to the most advanced models as financial agents, while ordinary consumers will only have access to free, lightweight models.

This asymmetry in computing power will no longer manifest as today’s “big data price discrimination,” but rather through continuous fee extraction across thousands of high-frequency, microtransactions, guided by seemingly reasonable negotiation logic. Users of the underlying model are not only exploited, but may also develop the illusion that the transactions are fair.

The asymmetry of computing power remains a visible and controllable risk, but when the underlying instructions are tampered with, the entire transaction network falls directly into a legal vacuum.

Anthropic identified a practical concern at the end of the report.

Project Deal is a closed and friendly internal test. What would happen if, in a real commercial environment, one party's AI agent was deliberately implanted with attack logic such as "jailbreaking" or "prompt injection"?

They simply need to hide a specific command within the trading conversation to induce your AI logic to crash, prompting it to voluntarily sell high-value assets for a penny or directly reveal its set floor price.

An AI agent, having had its code defenses breached, signed an extremely unfair contract—who should be held responsible? Current commercial legal frameworks are entirely silent on such AI-to-AI fraud.

Reviewing the entire experimental process of Project Deal, the step not documented in the research report was the final stage: after the AI agents completed all complex matching, probing, and negotiation, human employees met at the company, exchanging real items—such as ski boards, used bicycles, or ping pong balls—for cash.

In this micro business loop, the roles of humans and AI have been completely reversed.

In the past, humans were the "brain" behind commercial transactions, while AI and algorithms merely served as tools for price comparison, sorting, and "recommended for you" suggestions. But in the agent economy, AI has become the decision-maker, and humans have been reduced to acting as the physical logistics for AI.

This may be the most terrifying endpoint of the agent economy: humans, for the sake of convenience, voluntarily relinquish their right to participate in market competition. When all calculations, strategies, and even emotional value are handled by AI.

In the commercial chain, humans are left only with the physical labor of transporting goods and a confirming signature.

This article is from the WeChat public account "GeekPark" (ID: geekpark), authored by Moonshot.

Disclaimer: The information on this page may have been obtained from third parties and does not necessarily reflect the views or opinions of KuCoin. This content is provided for general informational purposes only, without any representation or warranty of any kind, nor shall it be construed as financial or investment advice. KuCoin shall not be liable for any errors or omissions, or for any outcomes resulting from the use of this information. Investments in digital assets can be risky. Please carefully evaluate the risks of a product and your risk tolerance based on your own financial circumstances. For more information, please refer to our Terms of Use and Risk Disclosure.