Your Users Are Already Telling You What To A/B Test

Behavioral data and user signals for better A/B test hypotheses.

The most valuable A/B test ideas don’t usually come from brainstorming sessions. They’re already embedded in how users interact with your site or app.

Users rarely tell you what’s wrong directly. Instead, they signal it through their behavior—clicking an image multiple times, expecting it to expand, or searching for a product they couldn’t find in the navigation. The insight is already there. The difference is whether you’re paying attention. 

And sometimes, those signals are so clear, they practically write the hypothesis for you.

When Rage Clicks Write The Hypothesis

On one of our clients’ sites, our behavioral analytics flagged a cluster of rage clicks on article cards. Users were repeatedly clicking directly on the card images, expecting them to function as links.

At a glance, nothing looked wrong. Standard click-through reporting showed the title and “read more” links were tagged correctly and tracking engagement. By that measure, the cards were working.

But session replays and heatmaps told a different story. The images weren’t clickable, yet users consistently treated them as if they were. That expectation isn’t random; it reflects a common web pattern where the entire card acts as a link, not just the text.

The hypothesis was straightforward: If the entire article card is made clickable, then click-through rate will increase, because it aligns with user expectations of how cards should function.

This is a simple example, but it highlights something important. The most effective experiments aren’t always the most creative; they’re the ones that remove friction users are already signaling. Behavioral cues such as rage clicks and drop-offs are among the strongest inputs for hypothesis generation because they reflect real user intent rather than internal assumptions. 

But user behavior alone doesn’t always tell the full story. 

Pair Behavior With The User’s Own Words 

Behavioral analytics data tells you that something is off. It doesn’t always tell you why.

When patterns like rage clicks emerge, a small, targeted survey can add that missing layer. If a user struggles with an interaction, completes it, and moves on, that’s a natural moment to ask a simple question: 

  1. “Was anything confusing about what you just tried to do?”
  2. “Did this page work the way you expected?”

Even a small number of responses can validate, or challenge, what the behavioral data suggests. Replays show you where friction exists; surveys give you the language behind it: what users were trying to do, and where the experience fell short. That language often becomes the most precise input you’ll have when writing the hypothesis. 

And sometimes, users don’t just describe the problem—they search for it.

What Users Search For Tells You What’s Missing

On-site search is one of the most underused signals in experimentation planning. When users repeatedly search for the same thing, it usually means one thing: they couldn’t find it where they expected to.

That’s a direct signal worth acting on. It can point to several types of experiments:

  1. Surface content earlier in the journey on the homepage, category pages, or relevant entry points.
  2. Refine navigation and labels so users recognize what they’re looking for without needing to search.
  3. Reorganize or create content so high-demand topics aren’t buried multiple clicks deep.

The goal isn’t to come up with a clever variation. It’s to let search behavior highlight where users are struggling, then test solutions that directly address that friction.

The signal quality here is high. Users who engage with on-site search tend to convert at significantly higher rates, making search query data especially valuable—you’re learning from your highest-intent visitors.

Across all of these examples, the pattern is the same—users are constantly signaling where the experience breaks down.

Don’t Stop At The Brainstorm

Brainstorming still has a place. It helps generate options, build alignment, and surface ideas worth exploring.

The problem is when it becomes the endpoint. Too often, teams rely on instinct alone, moving forward with solutions without validating which option actually works best. That’s part of why so many experiments end up inconclusive or fail to produce a clear winner: the hypothesis wasn’t grounded in real user behavior to begin with.

The goal isn’t to brainstorm less. It’s to anchor the brainstorm in real signals first.

A more effective loop looks like this:

  1. Identify friction through behavioral signals such as rage clicks, heatmaps, session replays, and search queries.
  2. Brainstorm solutions for that specific problem, narrowing to a few clear options.
  3. Test and learn through experiments, surveys, or targeted research.

This is what makes experimentation valuable. Each test answers a question rooted in actual user behavior. Each result sharpens the next hypothesis.

Over time, this approach leads to clearer learnings and more confident decisions because your testing stays anchored in what users are actually showing you, not just what you assume.

Your users are already showing you what to test. The question is whether you’re looking.

FAQs on User Signals and A/B Test Hypotheses

What are rage clicks, and how can they improve A/B test hypotheses?

Rage clicks, repeated clicks on a non-interactive element, are a direct signal that users expect something to behave differently than it does. When a behavioral analytics tool flags a cluster of rage clicks, that friction point can form the basis of a well-grounded test hypothesis.

For example, on one site Drumline worked on, users repeatedly clicked article card images expecting them to function as links. Standard click-through reporting showed nothing wrong, but session replays and heat maps revealed the real issue. The fix: wrapping the entire card in a link was a hypothesis that the user behavior effectively wrote itself.

How can on-site search data be used to identify experimentation opportunities?

When users repeatedly search for the same term, it typically signals that they couldn’t find what they were looking for in the existing navigation or page structure. That’s a high-quality, actionable input for experiment planning.

On-site search queries can point toward several types of tests: surfacing content earlier in the user journey, testing navigation labels so users can find what they need without resorting to search, or reorganizing content so high-demand topics aren’t buried. Research cited in the blog also notes that users who engage with on-site search convert at significantly higher rates than those who don’t, making this data a particularly valuable signal from your highest-intent visitors.

Why do most A/B tests fail to reach statistical significance?

Research on A/B testing quality suggests that the majority of tests don’t reach statistical significance, and a key contributing factor is hypotheses that aren’t grounded in observed user behavior. Tests built on internal assumptions or brainstorming alone tend to miss the specific friction points users are actually experiencing.

The solution isn’t to run more tests. It’s to anchor hypotheses in real behavioral signals first, such as rage clicks, heat maps, session replays, and search queries, before designing the experiment.

Should you use surveys alongside behavioral analytics when planning experiments?

Yes. Behavioral data tells you that something is off; targeted surveys can reveal why. When a user struggles with an interaction and completes it anyway, that moment is a reasonable trigger to surface one or two short questions, such as “Was anything confusing about what you just tried to do?”

Even a small response rate can confirm or challenge what behavioral data appeared to show. Surveys give you the user’s own language for what they were trying to accomplish, and that language often becomes the most precise input available when writing a test hypothesis.

What does a structured, signal-led experimentation loop look like?

A signal-led experimentation loop has three stages: spot real friction using behavioral data such as rage clicks, heat maps, session replays, and search queries; brainstorm two or three candidate solutions grounded in that specific problem; then test and learn through structured experiments, targeted surveys, or UX research.

This approach transforms experimentation from a series of disconnected bets into a compounding capability. Each test answers a question rooted in user behavior, and each result sharpens the next hypothesis. Over time, teams that follow this habit tend to run fewer but better tests. They also trust their results more because those results connect directly to something users signaled first.

Have questions or comments?

Visit our Contact Us page for all the details on how to get in touch with us. Whether you have a question, feedback, or just want to say hello, we’d love to hear from you!