Back to Blog

AI in Retail Pricing: Where It Actually Pays Off (and Where It Doesn't)

AI tops 47% of grocery CEO priority lists. But 80% of projects fail. The four places AI actually moves the P&L in retail pricing.
Retailgrid Team
April 27, 2026
10 min read

The 2026 edition of McKinsey's State of Grocery Retail Europe contains a sentence that should make every retail CEO pause: 47% of grocery CEOs now name adopting AI and automation as their top-two priority - up four ranks since 2024. (McKinsey)

That is the loudest signal in the report. It is also, on its own, a dangerous one.

Because here is the other number, from a 2026 RAND analysis of enterprise AI: roughly 80% of AI projects fail to deliver the business value they promised. In retail specifically the failure rate sits at about 74%. Of the global $684B invested in enterprise AI in 2025, more than $547B - around 80% - did not return the value the business cases assumed. (AI Governance Today, citing RAND)

Both numbers can be true. AI is the new CEO priority and most AI projects will quietly fail. The interesting question for any retailer reading this in Q2 is not whether to adopt AI in pricing. It is where AI actually moves the P&L - and where it just produces an expensive demo.

This piece is the version of that conversation we wish more boardrooms were having.

Why most retail AI projects fail

The failure pattern is not mysterious. Two causes show up in almost every postmortem.

The first is data volatility. Retail demand reshapes itself faster than most ML models can keep up. A 2026 industry analysis found that 44% of retail AI projects had their core models invalidated by demand shifts during the project lifecycle. (Folio3 AI) Fashion seasons, weather swings, competitor moves, and category-level resets all change the underlying signal. Models trained on 2024 elasticities behave like a slightly outdated map - directionally fine, dangerously wrong at the corners.

The second is integration. The same analysis found that supply-chain and operational integration was harder than expected in 81% of retail cases. AI doesn't run on slides. It runs on real data piped into real workflows that real category managers use on a real Tuesday afternoon. Integration is where the project budget actually disappears.

Underneath both causes sits one structural truth: AI works when it is bolted onto structured decisions. It struggles when it is asked to invent the structure itself.

A pricing AI that has been told "here is what a KVI is, here are our five product roles, here is the rule for promo cadence" will deliver. A pricing AI that has been asked to figure all of that out from scratch will, in most retailers, produce an expensive blackbox that the category manager overrides 60% of the time. The override rate is the leading indicator of project failure. By the time it shows up in the financials, the business case has already collapsed.

Four places AI actually pays off in retail pricing

There are dozens of demos. Most are noise. After several years of watching mid-market retailers (€10M-€500M revenue, 10k-200k SKUs) try AI in pricing, four use cases stand out as durably profitable. They share one trait: each one assists a structured decision. None of them is asked to make strategy.

1. Setting product roles

Every healthy pricing process starts by classifying SKUs into roles - key value items (KVIs), traffic drivers, margin builders, signature items, and tail. The classification is the foundation of every downstream rule.

Doing this manually at 50,000 SKUs is brutal and inconsistent. Doing it with AI, against historical sales, search frequency, basket-attachment, and competitor presence, is one of the highest-leverage applications you can run. The model proposes a role for each SKU. The category manager confirms or overrides. The next refresh learns from the overrides.

Why this works: classification is bounded. The model is not asked to set strategy - it is asked to apply a framework the team has already chosen. Reviewing 50,000 model proposals with audit trails is far faster than building the segmentation from scratch. And because product roles only meaningfully shift each season, you are not chasing weekly drift.

2. Exception triage

A weekly pricing review at 80,000 SKUs is impossible. A team of three category managers can sanity-check maybe a few hundred prices a week before fatigue makes the work counterproductive.

This is the most under-rated use case for AI in pricing. The model does not set the prices. It triages the queue: "These 230 prices look anomalous against your rules, your competitors, and your own history. The other 79,770 are within tolerance." The team reviews 230, not 80,000. The override rate stays high because the team is making real decisions, not rubber-stamping outputs.

Done well, this collapses the weekly pricing meeting from "panicked" to "productive" without removing humans from the loop. It also turns the override log into the most valuable training data the system has - because every override is a category manager teaching the model where the rule was wrong.

3. Competitor matching

Behind every pricing decision sits an unglamorous data problem: matching your SKUs against the SKUs being sold by competitors. EAN gets you maybe 30% of the assortment. The other 70% is descriptive matching - text, attributes, images, sometimes price clues - across thousands of competitor pages, refreshed weekly.

This used to be a multi-week consultant project. AI now turns it into a rolling refresh. It is not visible to the board. It is not on any keynote slide. But it is the foundation that every "AI pricing" pitch silently depends on. Get it wrong and every downstream model is fed garbage.

A practical test: ask any pricing vendor what their match rate is, what their false-match rate is, and how they catch new competitor SKUs. The answers separate the serious from the demo-only.

4. Margin attribution

Most retailers cannot answer a basic question after a quarter: which price moves drove margin, and which were noise? They have the rolled-up P&L. They do not have the SKU-by-SKU, week-by-week attribution.

AI is unusually good here, because the question is structured, the data is already in your DWH, and the output is a leaderboard, not an instruction. Which categories' margin gain in Q1 was driven by mix vs. price vs. promo timing? Which 200 SKUs were responsible for 80% of the margin movement? Which decisions, on review, were brilliant - and which were lucky?

This is not as flashy as "the AI sets your prices." It is what makes the next quarter's pricing decisions better. And it is the place where AI most often pays for itself in a single cycle.

What AI doesn't do well in retail pricing

The honest section. After two years of running these projects with mid-market retailers, here is what AI is not good at, no matter what the demo shows.

It does not set strategy. It optimizes within a strategy. If your category role definitions are weak, your private label tiering is muddled, or your promo philosophy contradicts your everyday-price philosophy, AI will optimize the contradiction faster - it will not resolve it.

It does not replace the assortment review. Cutting tail SKUs, building out private label, repositioning a category - these are merchandising decisions. AI can rank candidates. It cannot tell you what game you are playing in the category.

It does not override category role decisions. The role of "white bread, 750g" in a discount grocer is not the same as the role of "white bread, 750g" in a premium urban chain. The model can propose; the team owns the decision.

It does not work without clean data. The single most common reason AI projects underperform in retail is not algorithm choice. It is that nobody resolved which version of the SKU master is canonical, what "list price" means in three different systems, or where the cost ledger lives. McKinsey's 2026 number worth remembering: 77% of grocery CEOs name cost and margin pressure as their top focus. The AI investment will only move that needle if the underlying data plumbing is sound. (McKinsey)

The 90-day test for any AI pricing pitch

If you are evaluating AI pricing tools this year, the four questions below separate the 20% of projects that succeed from the 80% that quietly disappear. Every CEO can ask them. None of them require a data science team to answer.

1. Is the model trained on our data, or generic? A model trained only on industry benchmarks treats your assortment as average. Your assortment is not average. Insist on a model that learns from your sales, your competitors, your customers, your seasonality. If the answer is "we use a pre-trained foundation model," ask exactly which decisions are pre-trained and which are tuned to your business.

2. Can the category manager explain a single price move in one sentence? "The model said so" is not an answer. "We moved this SKU up because elasticity in this segment is below 0.6 and our key competitor raised it last week" is. Black-box outputs do not survive contact with category managers, suppliers, or boards.

3. Can the team override without an engineering ticket? Pricing decisions need to move at the speed of the category, not the speed of the vendor's roadmap. A pricing tool that requires a vendor ticket every time you want to add a rule, change a threshold, or freeze a SKU will be dead within a year.

4. Does it improve every week? A model that ships fixed and never re-trains is a 2018 product in 2026 packaging. Ask how often models retrain, what triggers a retrain, and what the override log feeds back into. A model that ignores the override log is throwing away its single most valuable training signal.

Three "yes" answers out of four is the ceiling of the surviving cohort. Two or fewer puts the project squarely in the 80% that fail.

Where mid-market retailers actually have the advantage

There is a quiet truth that does not get said enough: in pricing, mid-market retailers can move faster on AI than the top ten can.

The catalog is smaller, the data is cleaner, the org is flatter, and the category manager who decides on Tuesday is the same one whose number is in the model's slack channel on Friday. The hyperscalers have more data and more headcount. They also have more legacy, more politics, more committees, and more SKU master systems that disagree with each other.

A mid-market retailer with 50,000 SKUs, two pricing analysts, one category lead per major category, and a willingness to treat pricing as a weekly structured decision - not a quarterly committee output - can be running the four use cases above inside a single season. None of it requires a 20-person data science team. It requires structured tools that respect the analyst's judgment and a category lead who can read a margin attribution chart.

That is the disposition the 47% of CEOs who named AI as a top-two priority should be after. Not "deploy AI." Adopt structured decisions, then let AI handle the parts that don't need a human.

What we'd suggest reading next

If this is the conversation you are having with your board this quarter, the four questions above are the right place to start - and the four use cases are the right places to pilot. Pilots that succeed share one trait: they begin with a structured pricing process, then add AI to the parts that are bounded, repetitive, and measurable. The pilots that fail begin by buying an "AI pricing platform" and hoping the structure will appear later.

Retailgrid is built around exactly this disposition - structured, explainable pricing decisions at 10k-200k SKU scale, with AI helping in the four places it actually pays off. If you are mid-pilot or mid-evaluation, tell us how you price today and we will share the diagnostics we use with the retailers who are getting this right.

The 47% will keep climbing. The 80% failure rate will not move much. The retailers who win are the ones who treat AI as plumbing for clear decisions, not magic for unclear ones.

Enjoyed this article?
See how RetailGrid can transform your pricing workflow.