Skip to content
Impact ConversionBook a call

Methodology · April 2026

ICE-L: the prioritisation filter we actually use

Every CRO team has a backlog. Every CRO team runs out of hours before it runs out of backlog. The only question that matters is which test goes next.

Most teams answer that question with ICE. Impact, Confidence, Ease. Score each 1-10, multiply or average, sort descending, ship the top one. It is a perfectly reasonable framework that quietly fails in practice, for three reasons.

Why ICE breaks

The scores lie. When five people score the same idea, you get five numbers between 3 and 9. Nobody is calibrated. The "Confidence" score in particular is almost pure vibes. So the ranking is noise with a number on it.

Ease is the wrong axis. Ease measures the developer’s discomfort, not the company’s cost. A one-line CSS tweak scores a 10 for ease, but if it ships a broken checkout in three regions, it was not easy at all.

There’s no anchor to revenue. ICE tells you which idea is "best" relative to the rest of the list. It does not tell you whether the top idea is worth running at all. You can end up testing enthusiastically on a page that sees 200 visitors a month.

What the L adds

L is for Lifetime. Specifically: does this change move a metric that compounds, or does it move a one-off number?

A 2% lift on a checkout flow that runs 40,000 orders a year compounds for as long as the checkout stays live. A 5% lift on a seasonal campaign banner compounds for six weeks.

When you add L, the prioritisation shifts. Tests on infrastructure (checkout, PDP, pricing page, signup flow) start to beat tests on campaigns, even when the raw ICE score looks similar.

How we actually score

We score each test on four axes, 1 to 5, with written definitions:

  • Impact: estimated revenue or profit if the test wins, given the page’s traffic and the typical effect size for this category of change.
  • Confidence: how much customer research points at this being a real problem. Not a vibe. Did surveys, session recordings, or support tickets surface this?
  • Effort: total engineering + design + QA hours, including rollout. Not just the developer’s optimism.
  • Lifetime: how long does the win stay live? Permanent change scores 5. Seasonal or campaign-specific scores 2.

If a test doesn’t score at least 3 on three of the four axes, it doesn’t get a slot. Slots are precious. A program running 10-15 tests a quarter has maybe 45 slots a year. You want every one of them to compound.

The honest limit

Prioritisation frameworks are a tool, not an oracle. ICE-L will not tell you whether a particular test is the right one. Only research can do that. What it will tell you is which of your research-backed hypotheses goes first.

That is enough to end the meeting. The next meeting is about running the test, not debating whether to run it.


If your backlog is full of 30 ideas and nobody can agree on which one to run, the how we work page shows the loop we use. Or book a 15-minute call and we’ll talk through how to get your list down to the five that actually matter.