Demand Testing in the AI Era: The Only Signals That Matter

Demand Testing in the AI Era: The Only Signals That Matter

Most validation signals are noise. Learn the signal hierarchy that separates genuine demand from wishful thinking, plus a Go/No-Go framework.

Demand Testing in the AI Era: The Only Signals That Matter

Most founders are measuring the wrong things.

They look at email signups and call it traction. They watch visitor numbers and feel validated. They run a landing page test, get 200 email addresses, and start writing a spec.

Then they build. Then they launch. Then almost nobody uses it.

The problem is not the testing. The problem is treating weak signals as strong signals.

In the AI era, running a fake door test or a landing page experiment takes an afternoon. That ease is both a gift and a trap — because the speed of setup says nothing about the quality of the evidence you collect.

This post gives you a signal hierarchy and a Go/No-Go framework for making the call clearly: does genuine demand exist, or are you just measuring interest and calling it intent?

The core mistake: conflating interest with demand

Interest is cheap. It costs a visitor nothing to scroll through a page, nod approvingly, and enter their email.

Demand is different. Demand is demonstrated when someone:

  • exchanges real money,
  • trades meaningful time,
  • or takes an irreversible action.

The gap between "I'd try this" and "I'll pay for this now" is enormous. Most validation frameworks collapse these into the same bucket. That's where the delusion starts.

The goal of demand testing is not to confirm that your idea sounds interesting. It is to find out whether people will actually change their behaviour — and what it takes to make that happen.

The signal hierarchy

Think of validation signals as a ladder. The higher you climb, the stronger the evidence.

Tier 1: Awareness signals (weakest)

  • Page views
  • Social likes or saves
  • Email newsletter opens
  • Podcast listens about your topic

These tell you that people are curious. Nothing more. Awareness signals are useful for sizing an audience, not confirming a market.

Risk: Founders with popular content often mistake audience interest for product demand. Having 5,000 newsletter subscribers tells you nothing about willingness to pay.

Tier 2: Intent signals (moderate)

  • Email sign-ups to a waitlist
  • "Notify me" clicks on a fake door
  • Completing a typeform with qualifying questions
  • Joining a waitlist with a referral incentive

These are better. Someone gave you their email, which is a small commitment. But intent signals can still lie. Email is asymmetric — the visitor knows they can ignore every email that follows.

Use intent signals to identify interested segments, not to confirm a business.

Better version: combine an email sign-up with a qualifying question. Ask: "What's your current solution to this problem?" or "What would you pay for something like this?" Responses are far more diagnostic than sign-ups alone.

Tier 3: Engagement signals (stronger)

  • Attending a live demo or call
  • Using a beta product for more than one session
  • Responding to follow-up emails with specific questions
  • Sharing the product unprompted

Engagement signals indicate that someone invested real time. That raises the bar. A person who attends a 30-minute demo and asks detailed questions is qualitatively different from someone who signed up from a Twitter thread.

Engaged users are your early adopters. They are not your market. Treat them like a focus group, not a launch forecast.

Tier 4: Commitment signals (very strong)

  • Pre-orders (actual payment)
  • Paid pilot agreements
  • Letter of intent from an enterprise buyer
  • Willingness to pay a deposit or setup fee

This is where validation becomes material. When someone hands over money — even a small amount — the dynamic changes entirely. They are now motivated to see the product succeed. Objections become requirements. Feedback becomes honest.

The classic example is Basecamp's early days: Jason Fried and the team sold Basecamp before it was built, by describing what it would do and asking companies to pay to be founding customers. Payment turned vague interest into a real product conversation.

Tier 5: Behaviour change signals (strongest)

  • Switching from a current tool or workflow
  • Cancelling an existing subscription to use yours
  • Bringing colleagues or team members into the product
  • Using the product in a workflow without prompting

These are the hardest to manufacture and the most meaningful. Behaviour change means you disrupted a habit. That is the most reliable indicator of real product-market fit.

AI can help you reach tier 5 faster by automating the early tiers — but it cannot shortcut the evidence. The signals have to be real.

How AI changes demand testing (and what it doesn't change)

AI genuinely accelerates the mechanics:

  • Landing pages can be designed, copywritten, and deployed in hours
  • Ad creative can be generated and tested across multiple audiences simultaneously
  • Interview transcripts can be summarised and clustered for themes in minutes
  • Follow-up sequences can be personalised at scale

What AI cannot do is manufacture genuine demand. You can reach more people faster, but the quality of the signal depends on the behaviour of real humans, not the speed of your tooling.

The practical implication: use AI to increase experiment velocity, not to inflate weak signals into confidence.

If you can run 10 landing page tests in a week instead of one, you should be comparing signals across experiments — which message and which audience generates the highest-quality responses — not treating the first test as definitive.

The Go/No-Go framework

Once you have run your tests, you need a clear decision process. Here is a practical framework.

Score your signals

For each validation experiment, map what you collected to the hierarchy above. Assign a tier to each signal type.

An experiment that generated 400 email sign-ups (tier 2) is weaker evidence than one that generated 15 paid pilots (tier 4). Do not let volume override quality.

Apply the threshold test

Ask these four questions:

  1. Do we have at least one tier 4 or tier 5 signal? If not, you have interest, not demand. Keep testing before building.
  2. Can the paying or committed person articulate the problem clearly? If your early customers cannot explain why they want this, they may be buying a solution looking for a problem.
  3. Is the problem painful enough to justify switching? Status quo bias is real. If your early users are not currently suffering from this problem in a measurable way, churn will be brutal.
  4. Do we understand who didn't convert — and why? This is the most underused question in validation. The people who saw your test and chose not to engage are as important as those who did.

Make the call

  • Go: At least one tier 4 signal, clear problem articulation, and a defined customer segment. Build something small and specific for that segment, not a general product.
  • Pause: Strong tier 3 signals but no tier 4. Run a deeper experiment — a working prototype, a concierge pilot, or a direct sales conversation — before committing to a build.
  • No-go: Tier 1 or 2 signals only. The idea may still be interesting, but you don't have evidence of a market yet. Return to discovery before building.

What a good validation sprint looks like

A two-week validation sprint, run well, should produce at minimum:

  • 3–5 direct conversations with people in the target segment
  • One fake door or landing page experiment with 50–100 qualified visitors
  • At least 10 completed intent signals (with qualifying questions answered)
  • An attempt to collect at least one tier 4 signal

If you can't get a tier 4 signal in a two-week sprint, that is itself information — and more valuable than six months of building.


The AI era has removed the friction from running experiments. That's genuinely useful. But it has also made it easier to run lots of experiments badly, collect weak signals at speed, and mistake activity for evidence. The discipline is in the hierarchy: know what you're measuring, know what it means, and know when the evidence is strong enough to act on. Build on demand, not on interest.