When Segmentation Is Useful (and When It Can Hurt You)
Published on 30 de set. de 2025
by Christophe Perrin
One of the most powerful features of modern experimentation platforms is the ability to explore many metrics across any number of dimensions like country, browser, device, environment, language, etc. With just a few clicks, users can segment results deeply and discover hidden differences.
This abundance of data is a gold mine but it does not come without risks when not used carefully.
In this post, we’ll draw a clear line between data and metrics that inform reliable launch decisions, and those that should be used as possible red flags and for exploration purposes. We’ll look at the statistical fallacies and biases that can happen during post-hoc analysis, and explain how to extract real value from exploratory slicing without misleading yourself or your team.

Pre-registered Metrics vs. Exploratory Metrics
One of the most important questions in any experiment is: What metrics are you basing your decisions on?
Those are your pre-registered metrics ideally defined before you start. With ABsmartly primary, secondary and guardrail metrics all fall into that category. These are what determine success or failure.
All other metrics (or segments) are exploratory. They are useful, even critical, but they should not drive launch decisions without further validation.
While pre-registered metrics should answer:
Did this experiment succeed or fail according to our plan?
Exploratory metrics can help answer:
Why did this result happen?
What went wrong?
What should we try next?
The Texas Sharpshooter Fallacy
You have all heard about this before but imagine a cowboy shooting bullets at a barn and then drawing a target around the tightest cluster of holes. That is what you're doing when you find a "win" after slicing your experiment data 50 different ways and highlighting the best-performing segment as evidence of success.
It might feel like insight, but it’s likely nothing more than noise.
This is compounded by confirmation bias where humans have a tendency to interpret data in ways that support what they want to believe. One thing is sure, if you spend enough time slicing and dicing the data, you will always find some "evidence" supporting your idea. That's especially true if you are willing to ignore the rest.
The danger is real, you end up shipping products or features based on results which won't replicate. You have wasted time and resources and more importantly you accumulated learning which is simply not true, this hurts your understanding of your users and might mislead your future product direction. This, eventually, will affect trust that your peers might have in experimentation.
If a metric or segment was not part of your original hypothesis and pre-defined decision-making criteria then you simply can’t use it to make informed decisions about the impact of your change.
More Slices = Less Power
Every time you slice your data into segments (e.g., by country, browser, device type), you reduce the sample size you have in each bucket.
Lower sample size means:
Wider confidence intervals
Lower power
Greater variability
Higher chance of misleading effects
This is basic math. A segment that looks like it has a +20% lift might just be noise if it only has 500 participants and a huge variance.
If you're segmenting without considering power and without adjusting for multiple tests, you are increasing the risk of false positives and of making unreliable decisions.
Be sure to always base your launch decisions on pre-registered and, more importantly, on sufficiently powered metrics.
When Segmentation Is Useful
Despite the dangers we highlighted above, segmentation can still play a critical role in experimentation. Used responsibly, it can be a powerful tool for:
Monitoring & Debugging
You may see a flat result in the aggregate, but slicing reveals:
A specific language where the variant broke the layout
An app version with high app crashes
A browser with long load times
This isn't about decision-making but it's about quality assurance and issue detection. Having data processed in real-time, as it is on the ABsmartly platform, allows experimenters to start monitoring as soon as your experiment starts so they can abort early if they observe some strong signal that something might be broken for a segment of visitors. This process is a key part of running a safe experimentation program.
Generating new hypotheses
Segmentation is great for forming questions like:
“Why does the UK show a stronger lift than France?”
“Is this feature more effective on mobile?”
“Do power users respond differently to my treatment?”
These are hypotheses, not conclusions.
If a pattern emerges from exploratory slicing, it deserves its own experiment with proper power and, of course, pre-registered decision-making criteria.
Key takeaways
Segmenting experimental data is a diagnostic tool, not a decision tool.
Use segments to explain, not justify, your result.
Stick to your pre-defined decision-making criteria when evaluating an experiment.
Don’t retrofit a win from post-hoc slicing.
Remember that the more you slice, the more power you lose.
Do explore to generate new ideas but always validate them in follow-up experiments.