Skip to content

Agentic Commerce — The $1 Trillion Shift

Agentic Commerce

TL;DR: Agentic commerce — shopping powered by AI agents acting on behalf of consumers — could represent $1 trillion in US B2C retail by 2030 ($3-5 trillion globally). It’s not just e-commerce evolution; it’s a rethinking of shopping itself where AI anticipates needs, compares options, negotiates, and executes transactions.

What Is Agentic Commerce?

In traditional commerce, you navigate websites, compare products, and make decisions yourself. In the agentic era, AI agents do this for you:

  • Anticipate your needs (calendar shows a move → agent starts planning)
  • Navigate options across multiple platforms
  • Negotiate deals and bundle purchases
  • Execute transactions — sometimes without human involvement

“We’re entering an era where AI agents won’t just assist—they’ll decide. Business models need to evolve from optimizing clicks to earning trust from algorithms acting for consumers.” — Naveen Sastry, McKinsey Senior Partner

Market Projections

MetricValueSource
US B2C retail by 2030$1 trillionMcKinsey
Global projections$3-5 trillionMcKinsey
AI agent involvement by end 202625-30% of US online purchasesIndustry estimates
AI agent involvement by 202850% of online purchasesProjections

Three Interaction Models

1. Agent to Site

AI agents browse merchant websites, extract product data, and make recommendations.

Implication: Your site must be agent-readable, not just human-readable.

2. Agent to Agent

Consumer’s AI agent negotiates with merchant’s AI agent.

Example: Your agent negotiates a hotel upgrade with the hotel’s pricing agent.

3. Brokered Agent to Site

Platform agents (ChatGPT, Perplexity) intermediate between consumers and merchants.

Implication: You may never see the customer — only their agent.

Six Domains Merchants Must Address

To thrive in the agentic era, businesses must adapt across six key domains:

Innovation Required (Build New)

DomainWhat to Do
Customer Engagement & DiscoveryDevelop agents that understand intent and suggest products. Embed semantic metadata in catalogs.
Clienteling & LoyaltyBuild persistent customer-context layers accessible by agents. Expose loyalty APIs.

Renovation Required (Upgrade Existing)

DomainWhat to Do
Core Commerce PlatformsEnable agents to execute structured transactions with minimal human input. Add dynamic pricing, inventory-aware recommendations.
Payments & FraudDifferentiate benign agents from malicious bots. Verify intent and identity in real time.
In-Store Point of ServiceSync digital and physical journeys. Integrate spatial computing for in-store agent navigation.
Fulfillment & ReturnsAgent-ready fulfillment APIs. Automated return logic negotiation.

New Revenue Models

Traditional ad revenue will decline as agents bypass ads entirely. New models emerging:

ModelHow It Works
Multibrand BundlingAgent bundles purchases across brands; each gets revenue share
Real-time Negotiation FeesAgent negotiates upgrades/deals; platform takes success fee
Premium Agent SkillsSubscription access to specialized agents (fashion stylist, trip planner)
Data Insights SalesBrands pay for anonymized agent-filtered behavior analytics
Conversational MarketplacesPurchases via dialogue; monetize via listing fees and commissions
Interagent Protocol FeesFees for cross-platform agent interoperability
Sponsored Smart SuggestionsSubtle, intent-aligned suggestions (preserves user trust)

The Trust Challenge

“When a person walks into a store, the trust equation is straightforward: Do I trust this brand, this merchant, this product? When an AI agent shops on your behalf, trust becomes abstract, filtered through layers of data, automation, and institutional frameworks.” — McKinsey

Trust Is Contextual

What works in one market fails in another:

  • Germany/Japan: Still prefer traditional payment methods, account-to-account transfers
  • These markets may resist delegating purchase decisions to AI longer

Trust Must Be Earned Through Interaction

  • Clear communication, not just legal disclaimers
  • Users must define boundaries of trust
  • Consent must be a living, flexible agreement — not a checkbox

Three Risk Categories

1. Systemic Risk (The Snowball Effect)

When agents are interconnected, minor errors have exponential impact:

  • A faulty prompt triggers cascading unintended consequences
  • Incorrectly booked flights, overordered inventory, unauthorized purchases

Question: Do your agents fail gracefully? Can they backtrack?

When an AI agent makes a poor decision, who’s to blame?

  • The platform that developed the model?
  • The brand that deployed the agent?
  • The user who approved it?

EU AI Act provides some clarity; US regulations remain fragmented.

3. Data Sovereignty (Geopolitical Challenge)

  • Where is agent data processed?
  • Which country’s laws apply?
  • Data localization requirements vary globally

Strategic Questions for Leaders

First-Mover Advantage:

  • How can your business build a defensible moat through strategic API development?
  • What tech infrastructure and partnership ecosystem do you need?

Concierge Experience:

  • How do you create a unique AI-powered concierge that drives loyalty?
  • What can your brand agent do that generic agents cannot?

Revenue Protection:

  • As AI disintermediates ad revenue, what new models can you create?
  • What data monetization or subscription models make sense?

Consumer Trust:

  • How do you earn trust when delegating decisions to autonomous agents?
  • What transparency and human-override features do you need?

AI Agent Bias Factors

Recent academic research (Columbia + Yale, August 2025) reveals that AI agents have predictable biases in purchasing decisions — and these can be influenced.

What Influences AI Agent Product Selection

FactorDirectionNotes
Keyword order in title✅ CriticalMatching search terms exactly: +80pp selection increase
Ratings✅ Positive+0.1 rating increase improves chances
Review count✅ PositiveMore reviews signal trustworthiness
Positive badges✅ Positive”Bestseller”, “Recommended”, “Our Pick"
"Sponsored” label❌ NegativeReduces selection probability

Real-World Example

ZDNet tested ChatGPT’s buying agent for a housewarming present. BlancPottery was chosen because:

  • Tags and badges like “Etsy Recommended”
  • 5-star rating with several reviews
  • Keywords matching search: “Dinnerware Set”, “Handmade”

Model Differences Matter

Different AI agents weight factors differently:

  • GPT-4.1 preferred top-left positioned products
  • GPT-5.1 showed the opposite preference
  • Claude, Gemini, and GPT all have unique bias profiles

Implication for merchants: Test product pages across multiple AI agents, and re-test after major model updates.

Improving Decision Quality

AI agents are getting better at objective decisions. When presented with identical products where one had a 1% discount (objectively better):

ModelFailure Rate
Claude Opus 4.54.3% (down from 63.7% in Sonnet 3.5)
GPT-5.11% (down from 25.8% in GPT-4o)
Gemini 2.5 Flash0% (down from 2.8% in 2.0 Flash)

Key insight: As models improve, gaming tactics become less effective. Focus on genuine value and clarity.

Merchant Recommendations

  1. Optimize for AI agents, not only humans
  2. Understand which AI agents your customers use most
  3. Test different agents and adjust product pages
  4. Re-test after model updates — decisions change drastically

See seo/agentic-search-optimization for detailed optimization tactics.


Key Quote

“This is not a wait-and-see moment. Before long, nearly all retailers will have to grapple with the fact that a significant percentage of their customers will not be human users but rather AI agents. The companies that move first will be the ones that help shape the future.” — Lareina Yee, McKinsey Senior Partner

Key Takeaways

  • $1 trillion US market by 2030; this is not speculative
  • Three interaction models: agent-to-site, agent-to-agent, brokered
  • Six business domains require innovation or renovation
  • Ad revenue will decline; new monetization models required
  • Trust is foundational infrastructure, not just sentiment
  • First-movers will define the standards

Human Psychology vs. Agent Logic

Today’s social commerce platforms (TikTok, Instagram) exploit human psychological triggers:

  • Personalized recommendations → emotional arousal
  • Social proof (likes, reviews) → trust and FOMO
  • Scarcity cues → urgency and impulse action

These triggers compress human decision-making and drive impulse purchases. See marketing/social-commerce-psychology for practical application.

The agentic shift: When AI agents shop, these triggers may work differently:

  • Agents can verify scarcity claims via inventory APIs (fake urgency won’t work)
  • Social proof may become “agent proof” — what other AI agents recommend
  • Personalization becomes even more precise with full purchase history access
  • FOMO doesn’t affect algorithms the same way

Key question: Will merchants need separate optimization strategies for human buyers vs. AI agents?

Sources