Data Analysis Journal

Data Analysis Journal

A/B testing

10 Experiments Every Data Team Should Run - Issue 288

Top must-try experiments every data scientist should know - lessons from StatSig Summit.

Olga Berezovsky's avatar
Olga Berezovsky
Oct 29, 2025
∙ Paid

Welcome to the Data Analysis Journal, a weekly newsletter about data science and analytics.


Today I want to share a recap and summary of a recent talk by Liz Obermaie, a former Meta data scientist now at StatSig, at the summit in San Francisco

I’ll admit, I was skeptical first. The title felt a bit over the top, but after watching the recap, I was impressed. The content is practical and well-structured. Nothing I haven’t covered in my newsletter before, but Liz explained it so clearly that I had to save it - and share it with you.

Below is my summary and commentary of her talk — the 10 experiments every data team should run. Some things I don’t agree with, but overall, it’s a very decent step into data science 2.0.

All Models Are Wrong (But Some Are Useful)

Quoting George Box, “All models are wrong, but some are useful”, is actually a great way to introduce experimentation. It’s controversial because people focus on the “wrong” part. But what it really means is that every model involves uncertainty and assumptions. There’s always a tradeoff between the assumptions you make and how confident you can be in the results.

That tradeoff sits at the core of experimentation. As data scientists, we tend to overcomplicate this. We want to explain every assumption in detail, while business stakeholders just want to know: What should we do?

Why randomized controlled trials matter

There are many types of experiments (behavioral studies, quasi-experiments, difference-in-differences, lots of user research qualitative techniques), but this presentation was mostly about randomized controlled trials (RCTs), where randomization does the heavy lifting - exactly what happens during an A/B test.

Randomization is very important - it balances test confounders, both known and unknown, so you don’t have to model every possible bias. That frees up your mental bandwidth to focus on the real work: edge cases, communication, and stakeholder alignment, rather than defending your assumptions. In other words, if you’re unable to randomize your traffic, you still can run experiments, but analyzing and interpreting them right will take 10x more effort and time - you would need to model the lift separately for every possible attribute.

Keep in mind that randomization can happen at different units - sometimes it’s not users, but sessions, queries, or even servers. The right choice depends on the assumptions you can safely make.

Part 1: Basic Experiments

These are simple but foundational. Most organizations run them early and often.

1. Standard A/B Growth Test

This is the classic marketing experiment. You test variations of a message, call-to-action, or landing page.

Keep reading with a 7-day free trial

Subscribe to Data Analysis Journal to keep reading this post and get 7 days of free access to the full post archives.

Already a paid subscriber? Sign in
© 2025 Olga Berezovsky · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture