<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Data Analysis Journal: A/B testing]]></title><description><![CDATA[Test all the things]]></description><link>https://dataanalysis.substack.com/s/ab-testing</link><image><url>https://substackcdn.com/image/fetch/$s_!WdsI!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd7029b3-f274-4215-ac43-d275f496ecf8_200x200.png</url><title>Data Analysis Journal: A/B testing</title><link>https://dataanalysis.substack.com/s/ab-testing</link></image><generator>Substack</generator><lastBuildDate>Fri, 17 Apr 2026 16:18:59 GMT</lastBuildDate><atom:link href="https://dataanalysis.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Olga Berezovsky]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[dataanalysis@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[dataanalysis@substack.com]]></itunes:email><itunes:name><![CDATA[Olga Berezovsky]]></itunes:name></itunes:owner><itunes:author><![CDATA[Olga Berezovsky]]></itunes:author><googleplay:owner><![CDATA[dataanalysis@substack.com]]></googleplay:owner><googleplay:email><![CDATA[dataanalysis@substack.com]]></googleplay:email><googleplay:author><![CDATA[Olga Berezovsky]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[10 Experiments Every Data Team Should Run - Issue 288]]></title><description><![CDATA[Top must-try experiments every data scientist should know - lessons from StatSig Summit.]]></description><link>https://dataanalysis.substack.com/p/10-tests-every-data-team-should-test</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/10-tests-every-data-team-should-test</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 29 Oct 2025 12:03:12 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!elXI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf9a130f-570a-4dd5-bd5b-ccdde80c622f_1022x716.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to the Data Analysis Journal, a weekly newsletter about data science and analytics.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://dataanalysis.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://dataanalysis.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>Today I want to share a recap and summary of a recent talk by <a href="https://www.linkedin.com/in/lizobermaier/">Liz Obermaie</a>, a former Meta data scientist now at StatSig, at the summit in San Francisco</p><p>I&#8217;ll admit, I was skeptical first. The title felt a bit over the top, but after watching the recap, I was impressed. The content is practical and well-structured. Nothing I haven&#8217;t <a href="https://dataanalysis.substack.com/s/ab-testing">covered in my newsletter before</a>, but Liz explained it so clearly that I had to save it - and share it with you.</p><p>Below is my summary and commentary of her talk &#8212; the 10 experiments every data team should run. Some things I don&#8217;t agree with, but overall, it&#8217;s a very decent step into <strong>data science 2.0</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bCjk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ab60f-495a-47c9-a1fc-7bf0a5b2fcdb_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bCjk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ab60f-495a-47c9-a1fc-7bf0a5b2fcdb_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!bCjk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ab60f-495a-47c9-a1fc-7bf0a5b2fcdb_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!bCjk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ab60f-495a-47c9-a1fc-7bf0a5b2fcdb_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!bCjk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ab60f-495a-47c9-a1fc-7bf0a5b2fcdb_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bCjk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ab60f-495a-47c9-a1fc-7bf0a5b2fcdb_200x200.png" width="136" height="136" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a21ab60f-495a-47c9-a1fc-7bf0a5b2fcdb_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:136,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bCjk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ab60f-495a-47c9-a1fc-7bf0a5b2fcdb_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!bCjk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ab60f-495a-47c9-a1fc-7bf0a5b2fcdb_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!bCjk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ab60f-495a-47c9-a1fc-7bf0a5b2fcdb_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!bCjk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ab60f-495a-47c9-a1fc-7bf0a5b2fcdb_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><h2>All Models Are Wrong (But Some Are Useful)</h2><p>Quoting George Box, <em>&#8220;<a href="https://en.wikipedia.org/wiki/All_models_are_wrong">All models are wrong, but some are useful</a>&#8221;</em>, is actually a great way to introduce experimentation. It&#8217;s controversial because people focus on the &#8220;wrong&#8221; part. But what it really means is that every model involves uncertainty and assumptions. There&#8217;s always a tradeoff between the assumptions you make and how confident you can be in the results.</p><p>That tradeoff sits at the core of experimentation. As data scientists, we tend to overcomplicate this. We want to explain every assumption in detail, while business stakeholders just want to know: <em>What should we do?</em></p><h3>Why randomized controlled trials matter</h3><p>There are many types of experiments (behavioral studies, quasi-experiments, difference-in-differences, lots of user research qualitative techniques), but this presentation was mostly about <strong>randomized controlled trials (RCTs)</strong>, where randomization does the heavy lifting - exactly what happens during an A/B test.</p><p>Randomization is very important - it balances test confounders, both known and unknown, so you don&#8217;t have to model every possible bias. That frees up your mental bandwidth to focus on the real work: edge cases, communication, and stakeholder alignment, rather than defending your assumptions. In other words, if you&#8217;re unable to randomize your traffic, you still can run experiments, but analyzing and interpreting them right will take 10x more effort and time - you would need to model the lift separately for every possible attribute.</p><p>Keep in mind that randomization can happen at different units - sometimes it&#8217;s not users, but sessions, queries, or even servers. The right choice depends on the assumptions you can safely make.</p><h2><strong>Part 1: Basic Experiments</strong></h2><p>These are simple but foundational. Most organizations run them early and often.</p><h3><strong>1. Standard A/B Growth Test</strong></h3><p>This is the classic marketing experiment. You test variations of a message, call-to-action, or landing page.</p>
      <p>
          <a href="https://dataanalysis.substack.com/p/10-tests-every-data-team-should-test">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[How Notion Scaled Experimentation - Issue 286]]></title><description><![CDATA[How the team grew and scaled A/B testing without losing rigor - lessons from StatSig Summit.]]></description><link>https://dataanalysis.substack.com/p/how-notion-scaled-experimentation</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/how-notion-scaled-experimentation</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 15 Oct 2025 10:01:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9S4m!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56b41ce9-9353-4011-9c61-3d2faf6a69ce_1340x774.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Today I want to share a recap of my favorite talk from the <a href="https://www.notion.com/">Notion</a> data science team at the recent <a href="https://www.statsig.com/sigsum/">StatSig Summit</a> in San Francisco.</p><p>It&#8217;s a great one for anyone leading data or analytics teams, or trying to scale experimentation from a few tests to dozens running at once. The Notion team shared what it really takes to build a strong experimentation setup: the right infrastructure, project process, analytics framework, and, most importantly, the right mindset.</p><p>I liked this talk the most because they didn&#8217;t chase flashy milestones like &#8220;<em>50 active tests at a time</em>&#8221;. Instead, they focused on quality over speed - how to build trust in results and make every test count. After all, if you can&#8217;t interpret one test correctly, running 50 more won&#8217;t help much.</p>
      <p>
          <a href="https://dataanalysis.substack.com/p/how-notion-scaled-experimentation">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Rethinking A/B Testing for B2B and SaaS - Issue 274]]></title><description><![CDATA[Lessons from StatSig on best practices of designing and running experiments in B2B]]></description><link>https://dataanalysis.substack.com/p/rethinking-ab-testing-for-b2b-and</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/rethinking-ab-testing-for-b2b-and</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 13 Aug 2025 11:03:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!6-p1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6861e364-f214-4a10-8180-edaf708e7db0_1356x758.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to the Data Analysis Journal, a weekly newsletter about data science and analytics.</em></p><div><hr></div><p>When we talk about experimentation, we usually mean it in the context of B2C products. It&#8217;s rare to find examples or case studies on A/B testing for B2B or SaaS, where onboarding is sales-driven, user samples are small, there&#8217;s little variance, and every customer has a dedicated account manager providing white-glove support. What&#8217;s there to test?</p><p>I used to think that if there was no self-service option for customers, there wasn&#8217;t much to experiment with.</p><p>I was so wrong! There&#8217;s a lot of experimentation happening in B2B, it&#8217;s just different.</p><p>I haven&#8217;t run A/B tests for B2B or SaaS yet, so when I saw <a href="https://statsig.com/">Statsig</a>&#8217;s talk at this year&#8217;s <a href="https://www.datacouncil.ai/bay-2025">Data Council</a> on experimentation in B2B, I was intrigued. It ended up being my second-favorite presentation (after <a href="https://dataanalysis.substack.com/p/causal-inference-methods-for-bridging">Roblox&#8217;s session on causal inference</a>).</p><p>Today, I want to share my takeaways and learnings from that talk. While the principles of A/B testing never change, running experiments in the B2B world comes with unique challenges and some surprising advantages. Let&#8217;s break them down.</p>
      <p>
          <a href="https://dataanalysis.substack.com/p/rethinking-ab-testing-for-b2b-and">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Causal Inference Methods for Bridging Experiments and Strategic Impact - Issue 267]]></title><description><![CDATA[How to connect A/B test results to real-world business decisions - lessons from Roblox]]></description><link>https://dataanalysis.substack.com/p/causal-inference-methods-for-bridging</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/causal-inference-methods-for-bridging</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 16 Jul 2025 12:02:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!rx-E!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddf1b257-5ba4-4e50-84a1-7329971311e3_1440x798.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to the Data Analysis Journal, a weekly newsletter about data science and analytics.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://dataanalysis.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://dataanalysis.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>Today I want to introduce you to someone great - a fellow data scientist, <a href="https://www.linkedin.com/in/wenjing-zheng/">Wenjing Zheng</a>, Senior Data Science Manager at <a href="https://www.roblox.com/">Roblox</a>. I met Wenjing in May at <a href="https://www.datacouncil.ai/bay-2025">Data Council</a>, and today I want to share her insights on connecting A/B test results to real-world business decisions at Roblox.</p><p>This is what data scientists spend most of their time doing - teasing apart the true impact of A/B tests from holidays and seasonality, from external factors driving change, and from other campaigns running in parallel. It&#8217;s hard, like explaining why a true +10% lift in transactions shows up as only a 0.01% ARR increase. I&#8217;ve shared some of <a href="https://dataanalysis.substack.com/p/how-to-find-optimal-proxy-metrics">my methods of estimating such impact</a> before, and today, I want to show how Roblox is doing it.</p><p>Read below a recap of Wenjing&#8217;s talk on causal inference methods that help bridge the gap between clean experimental results and messy strategic decisions, how to attribute business growth to product launches, or generalize experiment outcomes to broader populations.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_YMB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb6d05d7-feaa-40c5-b87e-fb980535f874_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_YMB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb6d05d7-feaa-40c5-b87e-fb980535f874_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!_YMB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb6d05d7-feaa-40c5-b87e-fb980535f874_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!_YMB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb6d05d7-feaa-40c5-b87e-fb980535f874_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!_YMB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb6d05d7-feaa-40c5-b87e-fb980535f874_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_YMB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb6d05d7-feaa-40c5-b87e-fb980535f874_200x200.png" width="174" height="174" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fb6d05d7-feaa-40c5-b87e-fb980535f874_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:174,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_YMB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb6d05d7-feaa-40c5-b87e-fb980535f874_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!_YMB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb6d05d7-feaa-40c5-b87e-fb980535f874_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!_YMB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb6d05d7-feaa-40c5-b87e-fb980535f874_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!_YMB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb6d05d7-feaa-40c5-b87e-fb980535f874_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p><a href="https://www.roblox.com/">Roblox</a> is an online platform where users can both play and create their own games, called "experiences". Wenjing leads a data science team responsible for the experimentation platform, which hosts hundreds of tests and manages time-series tooling for forecasting, business monitoring, anomaly detection, and root cause analysis. The team uses multiple causal inference methods to enable and support data science partners.</p><p>Below, I share slides and takeaways on how Wenjing&#8217;s team approaches estimations for A/B tests' impact.</p><h2><strong>How to segment out local vs. global impact?</strong></h2><p>Individual teams at Roblox run experiments on surfaces like the homepage, notifications, marketplace, etc. Each experiment has a <em>local lift</em>, like +1% increase in time spent from a new notification. But local lift &#8800; global impact: some surfaces get lots of traffic (e.g., homepage), while others have niche audiences.</p><p>When leadership tries to prioritize based on impact, for example, &#8220;Team A improved time spent by 1%, Team B by 0.1%&#8221;, it ignores:</p><ul><li><p>Reach: How many users are exposed.</p></li><li><p>Baseline: Was it already optimized or easy to move?</p></li></ul><p>Typically, in this case, teams use qualitative intuition (&#8220;our surface has less reach,&#8221; &#8220;this was a hard problem,&#8221; etc.), leading to inconsistent prioritization. We need better ways to quantify it.</p>
      <p>
          <a href="https://dataanalysis.substack.com/p/causal-inference-methods-for-bridging">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Significance Level vs. Statistical Power in A/B testing - Issue 266]]></title><description><![CDATA[Statistics 101: Why analysts confuse significance with confidence, power with precision - and how to get it right.]]></description><link>https://dataanalysis.substack.com/p/significance-level-vs-statistical</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/significance-level-vs-statistical</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 09 Jul 2025 12:01:14 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Kva0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc72cc879-f091-4fc0-b7fe-f6eff3fd8724_1538x1032.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to the Data Analysis Journal, a weekly newsletter about data science and analytics.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://dataanalysis.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://dataanalysis.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>A very nice reader pointed out a typo in my <em><a href="https://dataanalysis.substack.com/p/ab-test-checklist-issue-233">A/B test Checklist</a></em>, and it reminded me of a moment during a meetup (many years ago) when I was arguing with a data scientist about whether we should adopt <em>Type I and Type II</em> errors or <em>Alpha and Beta</em> in our statistical documentation.</p><p>My point was that if you studied statistics in China, Europe, or the USSR, most of the literature refers to Alpha and Beta, and referring to Type I or Type II errors is less common (at least it used to be). But in the U.S., I&#8217;ve noticed that most documentation, especially around A/B testing (and especially newer sources), mostly uses Type I and Type II errors. That used to confuse me - I could never remember which one was the false positive and which was the false negative.</p><p>So instead, I trained myself to just stick with <em>Statistical Power</em> and <em>Confidence Level</em>.</p><p>But even then, I kept noticing that during interviews and at work, people often refer to <em>Power</em> as <em>Sensitivity</em>, to <em>p-value</em> as if it were <em>Power</em> (which it&#8217;s not), or simply to <em>Significance</em>, which, depending on context, can also mean <em>Confidence Interval</em> (which some sources <a href="https://www.surveymonkey.com/mp/ab-testing-significance-calculator/">call </a><em><a href="https://www.surveymonkey.com/mp/ab-testing-significance-calculator/">Significance Level</a></em> &#129760;&#128514;).</p><p>The reality is: there are only 4 key statistical terms we deal with in A/B testing, and yet we constantly confuse them or flip their definitions depending on context.</p><p>So, I decided to publish a quick refresher focused only on the core concepts: <strong>Significance Level, Confidence Level, Statistical Power,</strong> and <strong>Confidence Interval</strong>, and how to tell them apart, understand what they actually mean, and share their expected values.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!a3bM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd69ac194-7ec0-4c9a-923a-29bc5acd88f3_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!a3bM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd69ac194-7ec0-4c9a-923a-29bc5acd88f3_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!a3bM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd69ac194-7ec0-4c9a-923a-29bc5acd88f3_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!a3bM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd69ac194-7ec0-4c9a-923a-29bc5acd88f3_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!a3bM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd69ac194-7ec0-4c9a-923a-29bc5acd88f3_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!a3bM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd69ac194-7ec0-4c9a-923a-29bc5acd88f3_200x200.png" width="164" height="164" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d69ac194-7ec0-4c9a-923a-29bc5acd88f3_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:164,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!a3bM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd69ac194-7ec0-4c9a-923a-29bc5acd88f3_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!a3bM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd69ac194-7ec0-4c9a-923a-29bc5acd88f3_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!a3bM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd69ac194-7ec0-4c9a-923a-29bc5acd88f3_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!a3bM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd69ac194-7ec0-4c9a-923a-29bc5acd88f3_200x200.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div>
      <p>
          <a href="https://dataanalysis.substack.com/p/significance-level-vs-statistical">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Spring Recap: 100 Winning Tests, No Growth]]></title><description><![CDATA[Reflections from data summits, industry shifts, and what we&#8217;re still getting wrong in analytics.]]></description><link>https://dataanalysis.substack.com/p/spring-recap-100-winning-tests-no-growth</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/spring-recap-100-winning-tests-no-growth</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 18 Jun 2025 12:02:27 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to the Data Analysis Journal, a weekly newsletter about data science and analytics.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://dataanalysis.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://dataanalysis.substack.com/subscribe?"><span>Subscribe now</span></a></p><p>I waited to send out the May recap so I could include highlights from the Snowflake and Databricks summits, which just wrapped up last week. Today, I&#8217;m sharing the latest news from the data world - what&#8217;s changing, what it means for us, and what we can do to adapt and grow, even when things in the industry feel upside down.</p><p>Apologies for the long and somewhat heavy newsletter.</p><p>Below, I&#8217;ll share updates on the key tools we use every day, new benchmark reports we measure against, new good tutorials, interesting publications and research papers that caught my eye, and other important developments impacting our work in data analytics.</p><p>But first, 2 quick personal updates:</p><ol><li><p>I am heading to Europe this summer - Barcelona in June, Paris in July, and London in early August. I&#8217;ll be meeting with a few founders, analysts, and friends along the way. If you&#8217;re nearby, I&#8217;d love to connect and chat all things analytics.</p></li><li><p>I am currently looking for a full-time, on-site <em>Product Analyst</em> in San Francisco (40 hours/week). If you are interested in working with me or know someone great, please get in touch!</p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Qnep!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657754ed-431b-4278-bbdb-1d3d7d09de17_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Qnep!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657754ed-431b-4278-bbdb-1d3d7d09de17_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!Qnep!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657754ed-431b-4278-bbdb-1d3d7d09de17_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!Qnep!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657754ed-431b-4278-bbdb-1d3d7d09de17_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!Qnep!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657754ed-431b-4278-bbdb-1d3d7d09de17_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Qnep!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657754ed-431b-4278-bbdb-1d3d7d09de17_200x200.png" width="156" height="156" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/657754ed-431b-4278-bbdb-1d3d7d09de17_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:156,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Qnep!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657754ed-431b-4278-bbdb-1d3d7d09de17_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!Qnep!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657754ed-431b-4278-bbdb-1d3d7d09de17_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!Qnep!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657754ed-431b-4278-bbdb-1d3d7d09de17_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!Qnep!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657754ed-431b-4278-bbdb-1d3d7d09de17_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><h1>&#128266; Advocating for analytics</h1><p>Last month, I attended <a href="https://mauvegas.com/">MAU</a> in Vegas, the largest mobile app AdTech and MarTech conference, with over 2,000 apps. It was awesome to reconnect with colleagues, friends, and discover so many new apps. But I walked away with some mixed feelings about the speakers, talks, and content.</p><p>At an AdTech or MarTech event, of course you'd expect a focus on marketing, advertising, and store optimization. But what&#8217;s becoming concerning is how deeply marketing is starting to shape product&#8212;specifically how activation, retention, and product growth are being treated as extensions of marketing, rather than the other way around. Here&#8217;s what stood out:</p><h3><strong>1. Marketers still miss the point on Activation.</strong></h3><p>Many see activation as just the final step in onboarding, or &#8220;a high-value action,&#8221; without digging deeper. The few sessions that touched on user activation missed the mark. Activation is not measured against Trials or TTPs. It&#8217;s measured against long-term retention. And most importantly, it has to be <em><strong>predictive</strong></em> of retention. That means modeling behavior across time windows and activity patterns. Learn more - <a href="https://dataanalysis.substack.com/p/why-your-activation-analysis-is-wrong">Why Your Activation Analysis Is Wrong - And How to Fix It</a>.</p><h3><strong>2. Measurement is often overcomplicated or misused.</strong></h3><p>Another thing that frustrated me was that in every other talk, I saw a slide where an A/B test was measured against ARPU. Why? Are there no other (proper) measurements available?</p><p>I understand when founders or engineers run quick tests and check ARPU in RevenueCat or A/B Tasty to make a fast decision. But when someone with a &#8220;Growth Consultant&#8221; title uses ARPU as the main metric, it&#8217;s worrying.</p><p>Here&#8217;s the thing: unless you&#8217;re changing the price, a higher ARPU might mean fewer customers. ARPU = Revenue / Customers, right? So, <strong>if your customer count goes down, ARPU can go up</strong>. So, if your A/B test shows Variant with +25% increase in ARPU, very likely, <em>that&#8217;s not a win</em>:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WEsn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60585916-58f3-4561-a867-89f96e5b3c3b_1600x560.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WEsn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60585916-58f3-4561-a867-89f96e5b3c3b_1600x560.png 424w, https://substackcdn.com/image/fetch/$s_!WEsn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60585916-58f3-4561-a867-89f96e5b3c3b_1600x560.png 848w, https://substackcdn.com/image/fetch/$s_!WEsn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60585916-58f3-4561-a867-89f96e5b3c3b_1600x560.png 1272w, https://substackcdn.com/image/fetch/$s_!WEsn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60585916-58f3-4561-a867-89f96e5b3c3b_1600x560.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WEsn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60585916-58f3-4561-a867-89f96e5b3c3b_1600x560.png" width="1456" height="510" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/60585916-58f3-4561-a867-89f96e5b3c3b_1600x560.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:510,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WEsn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60585916-58f3-4561-a867-89f96e5b3c3b_1600x560.png 424w, https://substackcdn.com/image/fetch/$s_!WEsn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60585916-58f3-4561-a867-89f96e5b3c3b_1600x560.png 848w, https://substackcdn.com/image/fetch/$s_!WEsn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60585916-58f3-4561-a867-89f96e5b3c3b_1600x560.png 1272w, https://substackcdn.com/image/fetch/$s_!WEsn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60585916-58f3-4561-a867-89f96e5b3c3b_1600x560.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://dataanalysis.substack.com/p/when-simple-becomes-averages">When Simple Becomes Tricky: Making Sense of Averages in Reporting</a></figcaption></figure></div><p>ARPU is a highly misleading measurement for paywall tests (again, unless you change subscription price, which is a different story).</p><p>If you are using LTV, it&#8217;s even more riskier. ARPU has 2 variables (revenue and customers), but LTV includes more (let&#8217;s break it down: (1) revenue, user lifetime, which is (2) recent renewal minus (3) signup, and minus (4) CAC in some cases), making it very hard to know what&#8217;s really changed. Without clear variable breakdowns, the results likely to lead you in the wrong direction. <em><strong><a href="https://www.linkedin.com/posts/olgaberezovsky_teams-often-mistakenly-evaluate-ab-tests-activity-7186334083071782912-Zqtm?utm_source=share&amp;utm_medium=member_desktop&amp;rcm=ACoAAAVar7IBF3om5ghiuk3gTuZo6H8eArR8Yiw">Please, don&#8217;t use LTV or ARPU to measure an A/B test</a></strong></em>:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!D6hn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!D6hn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg 424w, https://substackcdn.com/image/fetch/$s_!D6hn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg 848w, https://substackcdn.com/image/fetch/$s_!D6hn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!D6hn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!D6hn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg" width="950" height="468" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:468,&quot;width&quot;:950,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:58617,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://dataanalysis.substack.com/i/166219453?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!D6hn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg 424w, https://substackcdn.com/image/fetch/$s_!D6hn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg 848w, https://substackcdn.com/image/fetch/$s_!D6hn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!D6hn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe33ddbf9-d119-4dbc-af2e-80e289088832_950x468.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">How To Find Optimal Proxy Metrics</figcaption></figure></div><p>My MAU frustration didn&#8217;t stop there. Sessions on retention, pricing, and monetization had often relied on flawed math or shaky logic. And then we wonder - why after 100 &#8220;winning&#8221; tests, the MRR growth rate doesn&#8217;t improve M/M? Please, invest in analytics and due diligence. It really does matter.</p><h1>&#128293; Last month&#8217;s highlights</h1><h3><strong>Snowflake Summit</strong></h3><p><a href="https://www.snowflake.com/en/summit/save-the-date/">Snowflake Summit</a> was fun. Who isn&#8217;t excited about 200 vendors offering slightly different versions of the same ETL tool?</p><p>This year, Snowflake focused heavily on data ingestion, catalogs, and agents. They also announced support for Postgres. But the highlight for me was <a href="https://www.snowflake.com/en/blog/intelligence-snowflake-summit-2025/">Snowflake Intelligence</a> - a new platform that lets you query any of your data using natural language and receive human-like responses and charts without writing SQL. I don&#8217;t see how this is different from <a href="https://www.databricks.com/product/business-intelligence/ai-bi-genie">Databricks&#8217; Genie</a>, announced a year ago... In any case, it requires (a) all data to be available in the data warehouse, (b) a very solid metadata layer to be built, and (c ) the right questions to be asked, which brings us right back to <a href="https://dataanalysis.substack.com/p/the-hidden-costs-and-pitfalls-of-cdp">where we started</a>.</p><h3><strong>Databricks Summit</strong></h3><p><a href="https://www.databricks.com/dataaisummit">Databricks Summit</a> felt pretty similar - same crowd, same vendors with more Spark references. They announced full support for Apache Iceberg and Delta Lake, MLflow 3.0 for end-to-end GenAI pipelines, direct integration with SAP systems, a partnership with Informatica, and more. For me, the most surprising update was that Sigma, a tool <a href="https://www.snowflake.com/en/blog/snowflake-ventures-expands-investment-in-sigma-deepening-commitment-to-bringing-world-class-bi-directly-into-the-ai-data-cloud/">built to support BI for Snowflake</a>, is now partnering with Databricks.</p><p>Personally, I find Sigma to be a questionable choice for BI.</p><p>I&#8217;m currently working on a deeper dive into Sigma - talking to its customers and reviewing use cases, and so far, I don&#8217;t understand why teams would use it. So, the Databricks + Sigma partnership was unexpeated and raises more questions.</p><h3><strong>Amplitude AI Agent Launch</strong></h3><p>Last week, I had the pleasure of joiningt the <a href="https://amplitude.com/blog/ai-agents">Amplitude AI Agent Launch</a>:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!04ag!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c0f2c28-8e1e-4d92-a435-b24d17c92da1_1600x871.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!04ag!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c0f2c28-8e1e-4d92-a435-b24d17c92da1_1600x871.png 424w, https://substackcdn.com/image/fetch/$s_!04ag!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c0f2c28-8e1e-4d92-a435-b24d17c92da1_1600x871.png 848w, https://substackcdn.com/image/fetch/$s_!04ag!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c0f2c28-8e1e-4d92-a435-b24d17c92da1_1600x871.png 1272w, https://substackcdn.com/image/fetch/$s_!04ag!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c0f2c28-8e1e-4d92-a435-b24d17c92da1_1600x871.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!04ag!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c0f2c28-8e1e-4d92-a435-b24d17c92da1_1600x871.png" width="1456" height="793" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1c0f2c28-8e1e-4d92-a435-b24d17c92da1_1600x871.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:793,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!04ag!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c0f2c28-8e1e-4d92-a435-b24d17c92da1_1600x871.png 424w, https://substackcdn.com/image/fetch/$s_!04ag!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c0f2c28-8e1e-4d92-a435-b24d17c92da1_1600x871.png 848w, https://substackcdn.com/image/fetch/$s_!04ag!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c0f2c28-8e1e-4d92-a435-b24d17c92da1_1600x871.png 1272w, https://substackcdn.com/image/fetch/$s_!04ag!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c0f2c28-8e1e-4d92-a435-b24d17c92da1_1600x871.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><blockquote><p>&#8220;Instead of acting on a vague alert like &#8220;conversion is down,&#8221; you now have a system that thinks like a proactive teammate&#8212;one that not only spots issues but digs into the why, proposes next steps, and runs experiments. While you focus on strategic priorities, your Agents are working in the background, testing multiple hypotheses, analyzing impact, and surfacing the highest-leverage insights.&#8221;</p></blockquote><p>I haven&#8217;t had a chance to test the agent myself yet, so I can&#8217;t comment on how much value it delivers to teams. I&#8217;d say, though, that <em>you can&#8217;t fix what you can&#8217;t see</em>, so the success of any AI agent will depend heavily on how well your tracking and analytics are set up.</p><p>Also, the demo focused on a fairly simple use case: identifying web dead clicks. Is this really a thing? Real-life use cases for hypothesis testing and analysis are way more complicated, but the overall promise and product direction looked impressive.</p><p>If you&#8217;re using Amplitude, you can sign up<a href="https://amplitude.com/ai-agents-beta"> for the beta</a> and try it out (and I&#8217;d love to hear your thoughts if you do!)</p><h3><strong>dbt is upgrading. Meet the <a href="https://docs.getdbt.com/blog/dbt-fusion-engine">dbt Fusion engine</a>.</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!h6H5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f4aba9-ff8e-4232-8f4a-7b380fad1f2f_1312x694.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!h6H5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f4aba9-ff8e-4232-8f4a-7b380fad1f2f_1312x694.png 424w, https://substackcdn.com/image/fetch/$s_!h6H5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f4aba9-ff8e-4232-8f4a-7b380fad1f2f_1312x694.png 848w, https://substackcdn.com/image/fetch/$s_!h6H5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f4aba9-ff8e-4232-8f4a-7b380fad1f2f_1312x694.png 1272w, https://substackcdn.com/image/fetch/$s_!h6H5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f4aba9-ff8e-4232-8f4a-7b380fad1f2f_1312x694.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!h6H5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f4aba9-ff8e-4232-8f4a-7b380fad1f2f_1312x694.png" width="1312" height="694" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/60f4aba9-ff8e-4232-8f4a-7b380fad1f2f_1312x694.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:694,&quot;width&quot;:1312,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!h6H5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f4aba9-ff8e-4232-8f4a-7b380fad1f2f_1312x694.png 424w, https://substackcdn.com/image/fetch/$s_!h6H5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f4aba9-ff8e-4232-8f4a-7b380fad1f2f_1312x694.png 848w, https://substackcdn.com/image/fetch/$s_!h6H5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f4aba9-ff8e-4232-8f4a-7b380fad1f2f_1312x694.png 1272w, https://substackcdn.com/image/fetch/$s_!h6H5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f4aba9-ff8e-4232-8f4a-7b380fad1f2f_1312x694.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://docs.getdbt.com/blog/dbt-fusion-engine">Meet the dbt Fusion Engine</a></figcaption></figure></div><blockquote><p>&#8220;It&#8217;s the same dbt you know and love, but better and faster&#8221;.</p></blockquote><p>dbt recently announced the Fusion Engine, which is currently in beta. It promises a lot of new features for teams, including a VS Code extension, real-time error detection, faster parsing, and improved project execution.</p><p>I haven&#8217;t figured out yet whether Fusion needs to be installed alongside Core or if it fully replaces it. From what I can tell, you&#8217;ll likely have to make a choice: stick with dbt Core, or get your team ready to &#8220;invest&#8221; and upgrade to Fusion.</p><h1>&#128200; New industry reports and benchmarks</h1><ul><li><p><a href="https://chartmogul.com/reports/saas-go-to-market-report/">The SaaS Go-To-Market Report</a> from ChartMogul</p></li><li><p><a href="https://www.bondcap.com/reports/tai">Mary Meeker's 2025 AI Trends Report: An In-Depth Product Analysis</a></p></li><li><p><a href="https://speero.com/experimentation-maturity-program-reports-2025">Experimentation Maturity Program Reports 2025</a> from CXL</p></li></ul><h1>&#9881;&#65039; Know your craft</h1><ol><li><p>Not sure which chart to use? <a href="https://www.metabase.com/learn/cheat-sheets/which-chart-to-use">This cheat sheet from Metabase will help you choose the right chart for your data.</a></p></li><li><p><a href="https://vexlio.com/blog/five-simple-things-that-will-immediately-improve-your-diagrams/">Five simple things that will immediately improve your diagrams</a> from <a href="https://vexlio.com/">Vexlio</a></p></li><li><p><a href="https://www.metabase.com/blog/what-is-embedded-analytics">What is embedded analytics?</a> from Metabase.</p></li><li><p><a href="https://mixpanel.com/blog/what-is-a-user-journeys-for-product-teams/">The essential guide to user journeys for product teams</a> - a must-read guide for product analysts from Mixpanel.</p></li><li><p><a href="https://www.mostlymetrics.com/p/how-public-companies-define-arr">How Public Companies Define ARR</a> from <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Mostly metrics&quot;,&quot;id&quot;:230760,&quot;type&quot;:&quot;pub&quot;,&quot;url&quot;:&quot;https://open.substack.com/pub/cjgustafson&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0447e0e8-ffb1-405c-9062-876b453548e1_1072x1072.png&quot;,&quot;uuid&quot;:&quot;31d23323-817a-4e6b-84db-df285292072d&quot;}" data-component-name="MentionToDOM"></span>. </p></li><li><p><a href="https://vyasenov.github.io/mind-map.html">Visualizing the World of Causal Inference</a> from <a href="https://www.linkedin.com/in/vasil-yasenov/">Vasco Yasenov</a>, a Staff Data Scientist at Adobe.</p></li><li><p><a href="https://github.com/gimseng/99-ML-Learning-Projects">99-ML-Learning-Projects</a> - A list of 99 ML projects for anyone interested to learn ML from coding and building projects.</p></li><li><p><a href="https://borkar.substack.com/p/the-8020-guide-to-r-you-wish-you">The 80/20 Guide to R You Wish You Read Years Ago</a> from <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Amol Borkar&quot;,&quot;id&quot;:165361390,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/67b97fbf-1f7c-4ac7-8698-2e9b24cf116c_1165x1167.jpeg&quot;,&quot;uuid&quot;:&quot;27e76749-84a3-4c5f-ad2a-ccfbba9c1c5b&quot;}" data-component-name="MentionToDOM"></span> </p></li></ol><h1>&#129299; Analysis and case studies</h1><ul><li><p>Last month, I published <em><a href="https://dataanalysis.substack.com/p/from-analytics-to-data-science-forecasting">Introduction to Forecasting</a></em>, followed by Part 2 - <em><a href="https://dataanalysis.substack.com/p/forecasting-in-analytics">Forecasting in Analytics: Choosing the Right Approach</a></em>, followed by an <a href="https://www.linkedin.com/feed/update/urn:li:activity:7329852764195344386/">interesting discussion on LinkedIn</a> where people asked me to break down my manual projection approach. So, now I&#8217;m working on <em>Forecasting Part 3</em>, coming (hopefully) soon. It&#8217;s not easy to consolidate, because forecasting revenue is very different from forecasting subscriptions or MAU. And completely different from projecting churn or growth rates. I&#8217;ll likely need to break it down by use case and product type to make it usable.</p></li><li><p><a href="https://a16z.com/ai-market-research/">Faster, Smarter, Cheaper: AI Is Reinventing Market Research</a> - from a16z</p></li><li><p><a href="https://benn.substack.com/p/no-really-everything-becomes-bi">No, really, everything becomes BI</a> from <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Benn Stancil&quot;,&quot;id&quot;:5667744,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/a317e60a-9bd1-4c75-bb54-66d517f735dc_1100x1100.jpeg&quot;,&quot;uuid&quot;:&quot;209182bf-fc30-4a54-a2b9-9501496aa448&quot;}" data-component-name="MentionToDOM"></span> </p></li><li><p><a href="https://filwd.substack.com/p/the-illusion-of-causality-in-charts">The Illusion of Causality in Charts</a> from <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Enrico Bertini&quot;,&quot;id&quot;:5453110,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9d4b9e31-2658-421a-b1d8-76218bc8c674_2316x2316.jpeg&quot;,&quot;uuid&quot;:&quot;71d6e9a9-fe2d-4dca-b211-fc247b0af939&quot;}" data-component-name="MentionToDOM"></span></p></li></ul><h1>&#127891; Tutorials</h1><p>Not a tutorial, but I wanted to share that <a href="https://rowzero.io/">RowZero</a>, a fast-growing spreadsheet for big data, has launched <a href="https://rowzero.io/edu">a free plan for students and recent graduates</a> who want to learn analytics and get comfortable with big spreadsheets, functions, pivot tables, charts, etc. The platform lets you explore large public datasets, collaborate with others, and publish your research.</p><p>Thanks for reading, everyone!</p><h3><strong>Previous Recaps:</strong></h3><ul><li><p><a href="https://dataanalysis.substack.com/p/april-recap-everyones-ready-for-ai">April Recap: Everyone&#8217;s Ready for AI. Except Your Data.</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/february-recap-the-rising-demand">February 2025 Digest: The Rising Demand for Analysts</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/january-digest-stop-ab-testing">January 2025 Digest: Stop A/B Testing&#8230; No, Do More!</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/november-digest-build-vs-buy">November 2024 Digest: Build vs. Buy Your A/B Testing Platform?</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/october-digest-is-product-analytics-behind">October 2024 Digest: Is Product Analytics Falling Behind?</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/september-digest-from-dashboards-to-spreadsheets">September 2024 Digest: From Dashboards to AI Spreadsheets</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/spring-digest-building-data-culture">Spring 2024 Digest: Building Data Culture With Metric Trees</a></p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://dataanalysis.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Data Analysis Journal is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Refresher on A/B Testing - Issue 258]]></title><description><![CDATA[Stats, tools, and trusted resources for doing experimentation right]]></description><link>https://dataanalysis.substack.com/p/refresher-on-ab-testing</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/refresher-on-ab-testing</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 14 May 2025 12:02:31 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!L9e4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4b7fabeb-3014-4241-918b-3a905300e999_1456x894.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to the Data Analysis Journal, a weekly newsletter about data science and analytics.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://dataanalysis.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://dataanalysis.substack.com/subscribe?"><span>Subscribe now</span></a></p><p>Yesterday, I ran a workshop on experimentation for a wonderful team of analysts who are transforming the creator economy. It hit me - I&#8217;ve never actually shared a proper introduction to A/B testing. Apparently, I just assume people are born knowing what statistical power or a p-value is.</p><p>So, today I decided to consolidate free tutorials and my bookmarked materials to help you get started with A/B testing and experimentation, including tools, key concepts to understand, free classes, and more.</p><div><hr></div><p>Before we jump in, spreading the word about a couple of upcoming events next week:</p><p>&#128467;&#65039; <em><a href="https://lu.ma/l80iwd4x?tk=CXSyHm">Data Science Salon returns to San Francisco</a>:</em></p><p>Join <a href="https://www.datascience.salon/">DSS</a> at the Amazon office next Wednesday for a meetup featuring speakers from Adobe and NextData. Teams will be talking all things data prep and management for AI and automation. Come hang out!</p><p>&#128467;&#65039; <em>I&#8217;ll be in Vegas next week,</em> joining 2,000 marketers at MAU Vegas 2025 - <a href="https://mauvegas.com/">mobile app AdTech and MarTech conference</a>. If you work in mobile app analytics, come say hi! Use this link to get a discount: <a href="https://i.snoball.it/p/jzK5X/l">MAU Vegas 2025</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!btac!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcce54040-2460-4873-bb2c-11afa67892f4_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!btac!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcce54040-2460-4873-bb2c-11afa67892f4_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!btac!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcce54040-2460-4873-bb2c-11afa67892f4_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!btac!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcce54040-2460-4873-bb2c-11afa67892f4_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!btac!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcce54040-2460-4873-bb2c-11afa67892f4_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!btac!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcce54040-2460-4873-bb2c-11afa67892f4_200x200.png" width="176" height="176" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cce54040-2460-4873-bb2c-11afa67892f4_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:176,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!btac!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcce54040-2460-4873-bb2c-11afa67892f4_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!btac!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcce54040-2460-4873-bb2c-11afa67892f4_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!btac!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcce54040-2460-4873-bb2c-11afa67892f4_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!btac!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcce54040-2460-4873-bb2c-11afa67892f4_200x200.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I have a dedicated <a href="https://dataanalysis.substack.com/s/ab-testing">experimentation section</a> in my newsletter, but I realized most of my publications are geared toward experienced analysts, diving into tricky cases like accelerating experimentation, testing on low traffic, or running tests in marketplaces. What I don&#8217;t have is a simple introduction to A/B testing.</p><p>That&#8217;s partly because you can&#8217;t really start with experimentation without first understanding statistics. And I don&#8217;t think it&#8217;s helpful to talk about types of A/B tests with someone who doesn&#8217;t yet understand p-values or probability. (Explaining p-values isn&#8217;t easy either.)</p><p>So, I don&#8217;t believe there&#8217;s such a thing as a gentle introduction to A/B testing. And it&#8217;s definitely not something you just "pick up" on the job. Even basic familiarity with testing requires a solid foundation in probabilities. (<em>Maybe that&#8217;s why it&#8217;s so hard to find analysts experienced with experimentation?</em>).</p><p>There are lots of paid courses, online academies, and YouTube videos out there on experimentation. And Maven products. Below, I&#8217;m sharing my vetted, trusted, and personally bookmarked free sources on A/B testing. If you go through them step by step, in a phased approach, I guarantee you&#8217;ll be well-equipped to work with A/B tests.</p>
      <p>
          <a href="https://dataanalysis.substack.com/p/refresher-on-ab-testing">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[A/B Test Checklist - Issue 233]]></title><description><![CDATA[A short guide to product experimentation steps and process.]]></description><link>https://dataanalysis.substack.com/p/ab-test-checklist-issue-233</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/ab-test-checklist-issue-233</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 20 Nov 2024 13:02:54 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/92f2818a-8828-4017-a509-3e4b14a059cc_1550x1102.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi everyone,&nbsp;I just realized it&#8217;s been 2 years since I shared my <a href="https://dataanalysis.substack.com/p/ab-test-checklist">A/B Test Checklist</a>. It&#8217;s one of the shortest articles in my newsletter but also one of the most popular. I designed it as a quick, handy guide and a refresher on terminology, steps, and key calculations.</p><p>This checklist is meant for product and marketing leaders who may not necessarily have a background in statistics but have to work hands-on on experimentation.&nbsp;</p><p>I&#8217;ve updated the guide to make it even clearer and more actionable. I hope you find it helpful!</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KP42!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6145d35-55dd-4bc3-bef5-ac72710e294b_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KP42!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6145d35-55dd-4bc3-bef5-ac72710e294b_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!KP42!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6145d35-55dd-4bc3-bef5-ac72710e294b_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!KP42!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6145d35-55dd-4bc3-bef5-ac72710e294b_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!KP42!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6145d35-55dd-4bc3-bef5-ac72710e294b_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KP42!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6145d35-55dd-4bc3-bef5-ac72710e294b_200x200.png" width="170" height="170" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c6145d35-55dd-4bc3-bef5-ac72710e294b_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:170,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!KP42!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6145d35-55dd-4bc3-bef5-ac72710e294b_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!KP42!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6145d35-55dd-4bc3-bef5-ac72710e294b_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!KP42!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6145d35-55dd-4bc3-bef5-ac72710e294b_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!KP42!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6145d35-55dd-4bc3-bef5-ac72710e294b_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><h1>&#9997;&#65039; Steps for conducting a product experiment:&nbsp;</h1><h3>1. Can you test it?&nbsp;</h3><p>Not everything can be A/B tested. New experiences or major product releases may not fit the typical A/B testing framework (read <em><a href="https://dataanalysis.substack.com/p/how-to-measure-new-feature-adoption">How To Measure New Feature Adoption</a></em>). Consider potential bias, such as <a href="https://productds.com/wp-content/uploads/Novelty_Effect.html">novelty effect</a> or change aversion.</p><h3>2. Should you test it?</h3><p><strong>Do you have enough users or events? </strong>Without a sample size of at least 50,000 users, A/B test results may not be statistically reliable.</p><p><strong>Set your expected rate</strong> &#8211; this is your <a href="https://splitmetrics.com/resources/minimum-detectable-effect-mde/">Minimum Detectable Effect</a> (MDE). The MDE represents the smallest acceptable difference between the Control and Variant. If the Variant is 0.02% better than the Control, would you still want to run the test? Is it worth the cost and time?</p><p><strong>Is it a good time to run the test?</strong> If the tested dashboard will eventually be sunsetted, the user flow replaced, or removed, there is no point in testing features that will be removed or deprecated.</p><p>Also, consider factors such as seasonality, upcoming version releases, open bugs, etc.</p><h3>3. Formulate A Hypothesis</h3><p><em>The &#8220;Newly tested user navigation flow will improve user retention&#8221;</em> is too vague to serve as a hypothesis. Instead, consider more measurable:</p><ul><li><p>New navigation flow will increase the frequency of user app opens per day by at least 10%.</p></li><li><p>New navigation flow will improve View-to-Paid CVR by at least 5%.</p></li><li><p>New navigation flow will drive at least 10% more DAU to the dashboard.</p></li></ul><p>A well-defined hypothesis should be focused on one or two success metrics.</p><h3>4. Finalize your set of metrics.</h3><p>For A/B analysis, I use a set of 3 metrics:</p>
      <p>
          <a href="https://dataanalysis.substack.com/p/ab-test-checklist-issue-233">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[How Long Until We Reach Significance? - Issue 231]]></title><description><![CDATA[Methods, steps, and tools for estimating A/B test duration]]></description><link>https://dataanalysis.substack.com/p/how-long-until-we-reach-significance</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/how-long-until-we-reach-significance</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 06 Nov 2024 13:00:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vv8J!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff40021d4-27e9-4bcc-b3eb-2913160ab9fc_1540x1032.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>With all eyes on today&#8217;s election results (thankfully, I don&#8217;t know them yet since this post was scheduled a few days ago), I know many of you might feel tired and drained. But here we are on Wednesday, and I&#8217;m using any opportunity to advocate for analytics done right. So, welcome to my <a href="https://dataanalysis.substack.com/">Data Analysis Journal</a> - a weekly newsletter about data science and analytics.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://dataanalysis.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://dataanalysis.substack.com/subscribe?"><span>Subscribe now</span></a></p><p>One of the most popular interview questions, and the first step when preparing for a test launch, is estimating how long the A/B test will take to run.</p><p>Figuring out the duration of a test really comes down to estimating how long it&#8217;ll take to reach significance.</p><p>And to estimate that timeline, you need to calculate the sample size needed for the A/B test.</p><p>Today, let&#8217;s dive into some easy and challenging ways, along with calculators and methods, to help you predict how many days, weeks, or even months it might take for a test to reach significance.</p>
      <p>
          <a href="https://dataanalysis.substack.com/p/how-long-until-we-reach-significance">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Frequentist vs. Bayesian: Which Method Should You Choose for Your A/B Testing? - Issue 224]]></title><description><![CDATA[Learn the pros, cons, and use cases for Frequentist and Bayesian approaches to ensure reliable and trusted A/B test results.]]></description><link>https://dataanalysis.substack.com/p/frequentist-vs-bayesian-which-method</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/frequentist-vs-bayesian-which-method</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Sun, 29 Sep 2024 12:03:11 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!wj-T!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97b509e9-8df0-48ef-a6f1-71b0cfb731d0_1600x661.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I don&#8217;t usually clutter your inboxes on a Sunday, but I have one more quick topic to share with you this month.</p><p>Last week, a colleague asked me which type of A/B test he should run -&nbsp; Bayesian or Frequentist. I searched for an article that explains the use cases but couldn&#8217;t find anything that doesn't dive into Hypothesis Testing Theory over 10 pages while still being helpful. So, I decided to write my own today.</p><p>Let&#8217;s talk about the types of A/B testing and which method to choose for different use cases. I know it&#8217;s Sunday, so I&#8217;ll keep it short and sweet.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tSM1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25905f2b-eafe-49bf-979f-e95a2ccd2383_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tSM1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25905f2b-eafe-49bf-979f-e95a2ccd2383_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!tSM1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25905f2b-eafe-49bf-979f-e95a2ccd2383_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!tSM1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25905f2b-eafe-49bf-979f-e95a2ccd2383_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!tSM1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25905f2b-eafe-49bf-979f-e95a2ccd2383_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tSM1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25905f2b-eafe-49bf-979f-e95a2ccd2383_200x200.png" width="200" height="200" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/25905f2b-eafe-49bf-979f-e95a2ccd2383_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tSM1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25905f2b-eafe-49bf-979f-e95a2ccd2383_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!tSM1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25905f2b-eafe-49bf-979f-e95a2ccd2383_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!tSM1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25905f2b-eafe-49bf-979f-e95a2ccd2383_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!tSM1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25905f2b-eafe-49bf-979f-e95a2ccd2383_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><h1>A quick recap on hypothesis test statistics:</h1><p>An A/B test is a form of hypothesis testing. There are many types of A/B tests, and at work, we typically apply A/A, A/B, Split, or Multivariate testing.</p><p>Frequentist and Bayesian approaches are different methods of interpreting A/B test results. These are NOT types of A/B tests you can choose but rather <strong>approaches to interpreting the test results</strong>.</p><p>While most articles refer to Frequentist and Bayesian methods as &#8220;ways to read test data,&#8221; I prefer to think of them as different schools or, better yet, <em>movements</em> in statistics to handle experimentation.</p><h2>Frequentist statistics&nbsp;</h2><p>If you graduated from any statistics class, you were likely taught the Frequentist method for interpreting A/B test data. This is primarily what academia has taught us. The Frequentist approach makes inferences about the test results using <strong>only data from the current experiment</strong>, without any prior context about user behavior. All parameters (mean, median, variance, etc.) and distributions used for A/B test analysis are considered fixed.</p>
      <p>
          <a href="https://dataanalysis.substack.com/p/frequentist-vs-bayesian-which-method">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The Database of Winning A/B Tests - Issue 218]]></title><description><![CDATA[A/B tests success rates, curated collection of proven A/B tests, and the importance of transparent experimentation.]]></description><link>https://dataanalysis.substack.com/p/the-database-of-winning-ab-tests</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/the-database-of-winning-ab-tests</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 21 Aug 2024 12:01:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Ev72!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99afb30c-bcf5-42bd-87cc-4d7f9d99433a_1600x913.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Before we start, I want to share an exciting opportunity in New York: join Peloton's Product Analytics team as a <a href="https://careers.onepeloton.com/en/all-jobs/6154493/data-scientist-product-analytics/?gh_jid=6154493">Data Scientist</a>! &#128202;&#128200;</p><div><hr></div><p>Have you ever wondered why companies spend time on A/B testing when thousands of other companies have already tested the same and know which shape, color, and type of CTA performs better?</p><p>Wouldn&#8217;t it be amazing to have a universal database or repository of all completed A/B tests, with documented learnings on what color, banner size, and paywall format performed best?</p><p>Well, there is one! Not one, many. And they are growing.</p><p>Today, let&#8217;s talk about:&nbsp;</p><ul><li><p>Where you can access these learnings from other companies and where you can share your own (please do).&nbsp;</p></li><li><p>Why we keep testing and re-testing exactly the same things. Over and over again.</p></li><li><p>What the test success rate is at other companies.&nbsp;</p></li><li><p>What the common winning stereotypes are (dark-screen mode, rounded buttons, sticky banners, etc.), and how accurate and trusted they are.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!b8Cg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8cf08fb-88a2-41bb-83ed-5e76f6f1ee69_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!b8Cg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8cf08fb-88a2-41bb-83ed-5e76f6f1ee69_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!b8Cg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8cf08fb-88a2-41bb-83ed-5e76f6f1ee69_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!b8Cg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8cf08fb-88a2-41bb-83ed-5e76f6f1ee69_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!b8Cg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8cf08fb-88a2-41bb-83ed-5e76f6f1ee69_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!b8Cg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8cf08fb-88a2-41bb-83ed-5e76f6f1ee69_200x200.png" width="164" height="164" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a8cf08fb-88a2-41bb-83ed-5e76f6f1ee69_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:164,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!b8Cg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8cf08fb-88a2-41bb-83ed-5e76f6f1ee69_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!b8Cg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8cf08fb-88a2-41bb-83ed-5e76f6f1ee69_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!b8Cg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8cf08fb-88a2-41bb-83ed-5e76f6f1ee69_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!b8Cg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8cf08fb-88a2-41bb-83ed-5e76f6f1ee69_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><h1>A/B tests repositories</h1>
      <p>
          <a href="https://dataanalysis.substack.com/p/the-database-of-winning-ab-tests">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Why it’s wrong to run A/B test for too long - Issue 213]]></title><description><![CDATA[The risks of extending the A/B test timeline or why duration matters.]]></description><link>https://dataanalysis.substack.com/p/why-its-wrong-to-run-ab-test-for-long</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/why-its-wrong-to-run-ab-test-for-long</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 17 Jul 2024 12:02:10 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oYbG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7466ed35-a341-400b-ab95-6c95e21faaad_1598x924.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I&#8217;ve talked a lot about why you shouldn&#8217;t stop A/B tests early and what the caveats are to disregarding the test significance.&nbsp;</p><p>However, one topic I haven't yet addressed in my newsletter is <strong>why it's also wrong to run an A/B test for too long</strong>.&nbsp;</p><p>In other words:</p><ul><li><p>if the test proved the hypothesis and</p></li><li><p>the winner is truly making a difference and&nbsp;</p></li><li><p>the observed lift didn&#8217;t occur by chance,&nbsp;</p></li></ul><p>Why does it matter if the test runs for 2 weeks or 6 months? If the winner is indeed true, shouldn't the longer timeline confirm it?</p><p>Not quite. Let's dive into this topic today.</p>
      <p>
          <a href="https://dataanalysis.substack.com/p/why-its-wrong-to-run-ab-test-for-long">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Methods To Accelerate A/B Testing - Issue 202]]></title><description><![CDATA[Strategies and statistical methods to increase test velocity.]]></description><link>https://dataanalysis.substack.com/p/methods-to-accelerate-ab-testing</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/methods-to-accelerate-ab-testing</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 15 May 2024 12:01:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29d8dad6-4384-4b3c-99f8-98ce46c57a88_1042x536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Last year, I published <a href="https://dataanalysis.substack.com/p/embracing-the-new-era-of-accelerated">Embracing the New Era of Accelerated Testing</a>, which was both somewhat controversial and emotional to write.&nbsp;</p><p>As a statistician trained to adapt academic principles to the fast-paced tech environment where <a href="https://dataanalysis.substack.com/p/inside-product-analytics-decoding">nothing is trusted</a>, I had to acknowledge that the concepts we were taught at school have become outdated and no longer serve us well.&nbsp;&nbsp;</p><p>The new generation of tools has accelerated the speed of product delivery. Every aspect of mobile/web development, including design, QA, and research, now runs twice as fast as it did a few years ago. Tools like Split, Superwall, Adapty, and even native Apple solutions offer incredible capabilities. Today, we have the technology to <a href="https://www.split.io/blog/dynamic-configurations-run-more-experiments-without-changing-code/">iterate "on the fly"</a> by continuously shipping and optimizing features.&nbsp;</p><p><strong>However, A/B testing practices and frameworks have remained the same, creating a gap between how fast the team is ready to move, how much trust we put in the data we receive, and how quickly we decipher its signals.</strong></p><p>Teams want to run more tests - faster and more efficiently. </p><p>Let&#8217;s discuss today what you can do to increase test velocity. What statistical methods and solutions are available for you to leverage to speed up testing, and how can you strike a balance between trust and speed?</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-Gud!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f88ca67-322c-4f9b-9914-24da5b405590_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-Gud!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f88ca67-322c-4f9b-9914-24da5b405590_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!-Gud!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f88ca67-322c-4f9b-9914-24da5b405590_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!-Gud!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f88ca67-322c-4f9b-9914-24da5b405590_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!-Gud!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f88ca67-322c-4f9b-9914-24da5b405590_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-Gud!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f88ca67-322c-4f9b-9914-24da5b405590_200x200.png" width="170" height="170" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9f88ca67-322c-4f9b-9914-24da5b405590_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:170,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-Gud!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f88ca67-322c-4f9b-9914-24da5b405590_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!-Gud!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f88ca67-322c-4f9b-9914-24da5b405590_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!-Gud!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f88ca67-322c-4f9b-9914-24da5b405590_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!-Gud!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9f88ca67-322c-4f9b-9914-24da5b405590_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><h2>Analytics is accelerating. Gear up.</h2><p>Duolingo is running <a href="https://blog.duolingo.com/improving-duolingo-one-experiment-at-a-time/">"a few hundred experiments simultaneously</a>.&#8221; At <a href="https://medium.com/pinterest-engineering/building-pinterests-a-b-testing-platform-ab4934ace9f4">Pinterest</a> and <a href="https://www.uber.com/blog/xp/">Uber</a>, over 1,000 experiments are active at any given time.</p>
      <p>
          <a href="https://dataanalysis.substack.com/p/methods-to-accelerate-ab-testing">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[How To Find Optimal Proxy Metrics - Issue 197]]></title><description><![CDATA[Using proxy metrics for measuring and quantifying the impact of a product rollout or a new feature.]]></description><link>https://dataanalysis.substack.com/p/how-to-find-optimal-proxy-metrics</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/how-to-find-optimal-proxy-metrics</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 17 Apr 2024 12:01:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!qLRR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71790b6a-ae07-438d-a499-8bb0cb72be0c_950x468.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to the <a href="https://dataanalysis.substack.com/">Data Analysis Journal</a>, a weekly newsletter about data science and analytics.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://dataanalysis.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://dataanalysis.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>Teams often mistakenly evaluate A/B tests against North Star metrics or business KPIs, such as user retention, customer churn, revenue, or LTV. <strong>This approach is flawed</strong> for 2 reasons:&nbsp;</p><ul><li><p>First, business metrics are not sensitive, meaning any lift created by an A/B test should not be reflected in KPIs (<a href="https://dataanalysis.substack.com/p/ab-test-checklist">unless your MDE is over 50%</a>).&nbsp;</p></li><li><p>Second, business metrics are designed to resist short-term changes from A/B tests and to reflect only long-term impacts.</p></li></ul><p>Despite this, using business KPIs to measure the effectiveness of A/B tests remains a common practice. To address this issue, researchers from Google, Stanford, and the Department of Statistical Science at Duke University collaborated on a study to explain why you should use <em>sensitive proxy metrics for A/B tests instead of the North Star or business KPIs</em>.&nbsp;</p><p>They introduced the concept of <a href="https://arxiv.org/pdf/2307.01000.pdf">Pareto Optimal Proxy Metrics</a>, which significantly optimize the accuracy and sensitivity of lift predictions. Let&#8217;s dive into why business core KPIs aren&#8217;t suitable for A/B tests and how to select the appropriate proxy metric for measuring product rollouts or feature optimization.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CbEO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F042e04a5-aec6-4314-aa18-38e8aa93ff27_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CbEO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F042e04a5-aec6-4314-aa18-38e8aa93ff27_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!CbEO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F042e04a5-aec6-4314-aa18-38e8aa93ff27_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!CbEO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F042e04a5-aec6-4314-aa18-38e8aa93ff27_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!CbEO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F042e04a5-aec6-4314-aa18-38e8aa93ff27_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CbEO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F042e04a5-aec6-4314-aa18-38e8aa93ff27_200x200.png" width="162" height="162" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/042e04a5-aec6-4314-aa18-38e8aa93ff27_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:162,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CbEO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F042e04a5-aec6-4314-aa18-38e8aa93ff27_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!CbEO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F042e04a5-aec6-4314-aa18-38e8aa93ff27_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!CbEO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F042e04a5-aec6-4314-aa18-38e8aa93ff27_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!CbEO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F042e04a5-aec6-4314-aa18-38e8aa93ff27_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><h1>Product is a subset of the business.</h1><p>If you have been reading my newsletter, you're already familiar with how I classify all metrics into three layers in product analytics:</p><h3>1. Success metrics&nbsp;</h3><p>Success metrics are 3-4 data points that directly illustrate the change in a user behavior affected by a product launch or a test. It can be the total or the average number of user actions, number of sales or transactions, click-through rate, monthly/annual subscription ratio, or else.</p><h3>2. Ecosystem metrics</h3><p>Ecosystem metrics are typically top or core KPIs. These are high-level business metrics that often involve calculations, specific date window logic, and filters. This can be MAU, MRR, Churn rate, or the North Star.</p><h3>3. Dark side / Tradeoff metrics</h3><p>Tradeoff, tension, or counter metrics are meant to show a negative effect of a product change that you want to keep your eyes on. For example, it can be % unsubscribes, refunds, reports of spam, or fraud. As I said earlier, if your product change increases usage in one feature, you are likely to see a decline in another. Users shift where they spend time. This metric helps you to measure the balance and fully understand the impact.</p><h3>Success metrics are not Ecosystem metrics.</h3><p>Bellow, I will reference the<a href="https://arxiv.org/abs/2307.01000"> Pareto optimal proxy metrics</a> study conducted by data scientists <a href="https://www.linkedin.com/in/lee-richardson-46180756/">Lee Richardson</a>, <a href="https://www.linkedin.com/in/alessandro-zito-92b250153/">Alessandro Zito</a>, <a href="https://www.linkedin.com/in/jacsor/">Jacopo Soriano</a>, and <a href="https://www.linkedin.com/in/dylan-greaves/">Dylan Greaves</a>.  </p><blockquote><p>&#8220;North star metrics are central to the operations of technology companies like Airbnb, Uber, and Google, amongst many others. Functionally, teams use north star metrics to align priorities, evaluate progress, and determine if features should be launched. Although north star metrics are valuable, there are issues using north star metrics in experimentation. To understand the issues better, it is important to know how experimentation works at large tech companies.&#8221;</p></blockquote><p>Traditional business KPIs are not suitable metrics for assessing the impact of product initiatives. These KPIs primarily serve as ecosystem health indicators meant to describe the state of business and safeguard it against various seasonal, external, and micro effects.</p><p>However, when conducting an A/B test, relying on metrics like retention may not accurately reflect the true impact of an initiative. For example, users in the Variant group might return to your app 1% more than those in the Control group, but 62% of them could also be influenced by other hidden effects you can&#8217;t capture or measure. These could include phone settings, a new viral video on TikTok, or unrelated factors like the plum season or a time change.</p><p>Remember the case study from Strava on <a href="https://dataanalysis.substack.com/p/how-strava-accelerated-user-engagement">how they boosted user engagement through a redesign of the Route Detail page</a>? For the A/B test, while they aimed to improve user retention in the app, they chose to use very sensitive, granular metrics such as average page views per user, saving routes, downloading routes, recording routes, and others, rather than relying on broader measures like week-over-week product usage or retention.</p><p>Retention is an ecosystem KPI. It is not sensitive to isolated fluctuation or noise. It isn&#8217;t designed to be sensitive because it&#8217;s an output metric (read <a href="https://brianbalfour.com/quick-takes/common-mistakes-defining-metrics">Common Mistakes In Defining Metrics</a> by <a href="https://brianbalfour.com/">Brian Balfour</a>). However, clicks, page opens, and transactions are good for capturing immediate responses and variations, thus offering a more nuanced view of specific actions or behaviors.</p><h1>Finding optimal proxy metrics</h1><h3>What is a proxy metric?</h3><blockquote><p>&#8220;The ideal proxy metric is short-term sensitive, and an accurate predictor of the long-term impact of the north star metric.&#8221;&nbsp;</p></blockquote><p>Here are 2 scenarios where the proxy metric helps teams overcome the limitations of the North Star metric:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qLRR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71790b6a-ae07-438d-a499-8bb0cb72be0c_950x468.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qLRR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71790b6a-ae07-438d-a499-8bb0cb72be0c_950x468.png 424w, https://substackcdn.com/image/fetch/$s_!qLRR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71790b6a-ae07-438d-a499-8bb0cb72be0c_950x468.png 848w, https://substackcdn.com/image/fetch/$s_!qLRR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71790b6a-ae07-438d-a499-8bb0cb72be0c_950x468.png 1272w, https://substackcdn.com/image/fetch/$s_!qLRR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71790b6a-ae07-438d-a499-8bb0cb72be0c_950x468.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qLRR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71790b6a-ae07-438d-a499-8bb0cb72be0c_950x468.png" width="716" height="352.7242105263158" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/71790b6a-ae07-438d-a499-8bb0cb72be0c_950x468.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:468,&quot;width&quot;:950,&quot;resizeWidth&quot;:716,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qLRR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71790b6a-ae07-438d-a499-8bb0cb72be0c_950x468.png 424w, https://substackcdn.com/image/fetch/$s_!qLRR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71790b6a-ae07-438d-a499-8bb0cb72be0c_950x468.png 848w, https://substackcdn.com/image/fetch/$s_!qLRR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71790b6a-ae07-438d-a499-8bb0cb72be0c_950x468.png 1272w, https://substackcdn.com/image/fetch/$s_!qLRR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71790b6a-ae07-438d-a499-8bb0cb72be0c_950x468.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Using business KPIs or North Star metrics for A/B testing has 2 main concerns:</p><h4>1. Ecosystem core metrics are not sensitive.&nbsp;</h4><p>Sensitivity refers to the metric&#8217;s ability to detect a statistically significant effect, often associated with statistical power. If the metric is not sensitive, it means that the test results do not clearly indicate whether the hypothesis is improving or degrading the KPI.</p><blockquote><p>&#8220;For example, metrics related to Search quality will be more sensitive in Search experiments, and less sensitive in experiments from other product areas (notifications, home feed recommendations, etc.).&#8221;</p></blockquote><h4>2. Business metrics are not directional to user behavior or experience</h4><blockquote><p>&#8220;Through directionality, we want to capture the alignment between the increase (decrease) in the metric and long-term improvement (deterioration) of the user 4 experience. While this is ideal, getting ground truth data for directionality can be complex.&#8221;</p></blockquote><p>Researchers suggest several methods to measure directionality by comparing the short-term value of a metric against the long-term value of a KPI.&nbsp;</p><blockquote><p>&#8220;The advantage of this approach is that we can compute the measure in every experiment. The disadvantage is that the estimate of the treatment effect of the north star metric is noisy, which makes it harder to separate the correlation in noise from the correlation in the treatment effects. This can be handled, however, by measuring correlation across repeated experiments.&#8221;</p></blockquote><p>Essentially, you need to run correlations between a proxy metric and a KPI, measuring their linear relationship (or it can be a Spearman correlation).</p><p>However, there&#8217;s a catch: there is an inverse relationship between sensitivity and directionality:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vM3m!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda89d547-9950-4c5c-b2dd-af6d15deed77_794x658.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vM3m!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda89d547-9950-4c5c-b2dd-af6d15deed77_794x658.png 424w, https://substackcdn.com/image/fetch/$s_!vM3m!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda89d547-9950-4c5c-b2dd-af6d15deed77_794x658.png 848w, https://substackcdn.com/image/fetch/$s_!vM3m!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda89d547-9950-4c5c-b2dd-af6d15deed77_794x658.png 1272w, https://substackcdn.com/image/fetch/$s_!vM3m!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda89d547-9950-4c5c-b2dd-af6d15deed77_794x658.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vM3m!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda89d547-9950-4c5c-b2dd-af6d15deed77_794x658.png" width="510" height="422.6448362720403" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/da89d547-9950-4c5c-b2dd-af6d15deed77_794x658.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:658,&quot;width&quot;:794,&quot;resizeWidth&quot;:510,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vM3m!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda89d547-9950-4c5c-b2dd-af6d15deed77_794x658.png 424w, https://substackcdn.com/image/fetch/$s_!vM3m!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda89d547-9950-4c5c-b2dd-af6d15deed77_794x658.png 848w, https://substackcdn.com/image/fetch/$s_!vM3m!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda89d547-9950-4c5c-b2dd-af6d15deed77_794x658.png 1272w, https://substackcdn.com/image/fetch/$s_!vM3m!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda89d547-9950-4c5c-b2dd-af6d15deed77_794x658.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Finding the right proxy metric involves balancing sensitivity and directionality: <strong>the more you increase sensitivity, the less likely the proxy metric will relate to business KPIs</strong>. This is a very frustrating aspect of product experimentation. Basically, the faster you can detect a significant lift in a success metric, the further it may be from reflecting true ecosystem KPI improvements, and vice versa.</p><p>Researchers have proposed a new method for identifying proxy metrics that optimizes the trade-off between sensitivity and directionality using a <a href="https://en.wikipedia.org/wiki/Pareto_efficiency">Pareto optimal framework</a> (or Pareto efficiency), a concept from economics and statistics used for problem optimization when dealing with multiple functions or factors. They tested 3 algorithms on over 500 experiments conducted over a period of 6 months. This testing aimed to compare the proxy metric versus the North Star metric, using binary sensitivity and the proxy score to assess the quality of the proxy metric.&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Gk0c!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F274a58c4-61b3-4dec-8b61-32621e6d7958_912x354.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Gk0c!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F274a58c4-61b3-4dec-8b61-32621e6d7958_912x354.png 424w, https://substackcdn.com/image/fetch/$s_!Gk0c!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F274a58c4-61b3-4dec-8b61-32621e6d7958_912x354.png 848w, https://substackcdn.com/image/fetch/$s_!Gk0c!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F274a58c4-61b3-4dec-8b61-32621e6d7958_912x354.png 1272w, https://substackcdn.com/image/fetch/$s_!Gk0c!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F274a58c4-61b3-4dec-8b61-32621e6d7958_912x354.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Gk0c!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F274a58c4-61b3-4dec-8b61-32621e6d7958_912x354.png" width="698" height="270.9342105263158" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/274a58c4-61b3-4dec-8b61-32621e6d7958_912x354.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:354,&quot;width&quot;:912,&quot;resizeWidth&quot;:698,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Gk0c!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F274a58c4-61b3-4dec-8b61-32621e6d7958_912x354.png 424w, https://substackcdn.com/image/fetch/$s_!Gk0c!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F274a58c4-61b3-4dec-8b61-32621e6d7958_912x354.png 848w, https://substackcdn.com/image/fetch/$s_!Gk0c!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F274a58c4-61b3-4dec-8b61-32621e6d7958_912x354.png 1272w, https://substackcdn.com/image/fetch/$s_!Gk0c!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F274a58c4-61b3-4dec-8b61-32621e6d7958_912x354.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><blockquote><p>&#8220;The key idea is that the proxy score rewards both sensitivity, and accurate directionality. More sensitive metrics are more likely to be in the first and third rows, where they can accumulate reward. But metrics in the first and third rows can only accumulate reward if they are in the correct direction. Thus, the proxy score rewards both sensitivity and directionality. Microsoft independently developed a similar score, called Label Agreement&#8221;</p></blockquote><p>Read more about their study - <a href="https://arxiv.org/pdf/2307.01000.pdf">Pareto Optimal Proxy Metrics</a>.</p><p>I'm not aware of any tool allowing product owners to quickly evaluate the relationship between proxy metrics and business KPIs, guiding them toward the appropriate success metric. Ideally, you would have a fellow analyst on your side who can help translate each business KPI into suitable proxy metrics. These metrics can then be used to measure the effectiveness of product initiatives. For example:</p><ul><li><p>Monthly retention &#8594; screen views or particular clicks.&nbsp;</p></li><li><p>MRR/ARR&nbsp; &#8594; successful transactions or completed payments.</p></li><li><p>Churn &#8594; cancellations, requests to cancel, or unsubscribes.&nbsp;</p></li><li><p>Subscription renewals &#8594; successful payments or transactions.</p></li></ul><div><hr></div><p>I am fascinated by this study. Not only is it timely, but it's also easy to understand and interpret. It addressed long-standing pain points we deal with in experimentation - why are improvements in the Variant not reflected in ecosystem core metrics? Why do these improvements only become visible in KPIs 4 - 5 months after the test is completed? And why every method attempting to quantify the relationship between the success proxy metric and the business KPI seems destined to fail.</p><p>Thanks for reading, everyone. Until next Wednesday!</p><h3><strong>Related publications:&nbsp;</strong></h3><ul><li><p><a href="https://dataanalysis.substack.com/p/how-to-measure-new-feature-adoption">How To Measure New Feature Adoption</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/inside-product-analytics-decoding">Inside Product Analytics: Decoding User Behavior</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/an-analysis-of-bias-or-why-ab-testing">An Analysis Of Bias Or Why A/B Testing Fails</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/the-ultimate-guide-to-product-features">The Ultimate Guide To Product Features Analysis</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/my-most-challenging-analysis-issue">My Most Challenging Analysis</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/how-strava-accelerated-user-engagement">How Strava Accelerated User Engagement</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/november-digest-its-okay-not-to-have">It&#8217;s Okay Not To Have All The Answers</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/why-you-shouldnt-stop-ab-tests-early-cbd">Why You Shouldn&#8217;t Stop A/B Tests Early</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/where-to-build-that-metric">Where To Build That Metric</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Why You Shouldn’t Stop A/B Tests Early - Issue 193]]></title><description><![CDATA[Why significance matters, and how long you should run an A/B test]]></description><link>https://dataanalysis.substack.com/p/why-you-shouldnt-stop-ab-tests-early-cbd</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/why-you-shouldnt-stop-ab-tests-early-cbd</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 20 Mar 2024 12:01:50 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F417135f9-7fc9-4a55-a109-078e4aa904d8_1600x455.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to the <a href="https://dataanalysis.substack.com/">Data Analysis Journal</a>, a weekly newsletter about data science and analytics.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://dataanalysis.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://dataanalysis.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>Today, I wanted to cover the most commonly asked questions on experimentation in analytics:&nbsp;</p><ul><li><p>How long should you run an A/B test? It&#8217;s recommended for 2 weeks, but <em><strong>why</strong></em>?&nbsp;&nbsp;</p></li><li><p>Can you (or should you) stop an A/B test early?&nbsp;</p></li><li><p>If you have to, what is the safest approach to handling fast A/B tests?&nbsp;&nbsp;</p></li><li><p>Why are slow rollouts dangerous?&nbsp;</p></li><li><p>What is the recommended procedure for gradually launching A/B tests over time?</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bz-f!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5bc1db42-af3b-4c1d-9edd-95d38036634a_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bz-f!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5bc1db42-af3b-4c1d-9edd-95d38036634a_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!bz-f!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5bc1db42-af3b-4c1d-9edd-95d38036634a_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!bz-f!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5bc1db42-af3b-4c1d-9edd-95d38036634a_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!bz-f!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5bc1db42-af3b-4c1d-9edd-95d38036634a_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bz-f!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5bc1db42-af3b-4c1d-9edd-95d38036634a_200x200.png" width="200" height="200" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5bc1db42-af3b-4c1d-9edd-95d38036634a_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:200,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bz-f!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5bc1db42-af3b-4c1d-9edd-95d38036634a_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!bz-f!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5bc1db42-af3b-4c1d-9edd-95d38036634a_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!bz-f!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5bc1db42-af3b-4c1d-9edd-95d38036634a_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!bz-f!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5bc1db42-af3b-4c1d-9edd-95d38036634a_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>There's a wealth of information on A/B testing available, ranging from academic papers on the frequentist approach (which is often less relevant for marketing or product analytics) to complex probability theories. Today, many product and analytics leaders may not have a background in statistics. Their knowledge of A/B testing often comes from online courses and self-study. As a result, they might not grasp the differences between A/B testing and Hypothesis testing or A/B testing and Split testing. This can lead to applying the same strategies to each or, even worse, having the same expectations for data trust in their results.</p><blockquote><p>&#8220;A Bayesian is one who, vaguely expecting a horse, and catching a glimpse of a donkey, strongly believes he has seen a mule.&#8221; <a href="https://twitter.com/robbalon">Dr. Rob Balon</a>.</p></blockquote><p>While I can't reinvent statistical methods to make Bayesian analysis fit the Frequentist framework, preventing stakeholders from being offended when analysts admit to having low trust in data (breaking news: <a href="https://f.hubspotusercontent00.net/hubfs/215600/qubit-research-ab-test-results-are-illusory-1.pdf">most winning A/B Test results are illusory</a>), I can at least re-iterate and clarify the basics here.</p><h1>Why significance matters</h1><p>Because early experimental data is more likely to be wrong.&nbsp;</p><p>You shouldn&#8217;t stop the test early - even if you think you see a clear winner - because of regression to the mean.</p><p><a href="https://en.wikipedia.org/wiki/Regression_toward_the_mean">Regression to the mean</a> (RTM) theory describes the false-positive result. It&#8217;s an effect when a variable is extreme at first but then moves closer to the average. In real life, the RTM conversion looks approximately like this:&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rVXK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F417135f9-7fc9-4a55-a109-078e4aa904d8_1600x455.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rVXK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F417135f9-7fc9-4a55-a109-078e4aa904d8_1600x455.png 424w, https://substackcdn.com/image/fetch/$s_!rVXK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F417135f9-7fc9-4a55-a109-078e4aa904d8_1600x455.png 848w, https://substackcdn.com/image/fetch/$s_!rVXK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F417135f9-7fc9-4a55-a109-078e4aa904d8_1600x455.png 1272w, https://substackcdn.com/image/fetch/$s_!rVXK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F417135f9-7fc9-4a55-a109-078e4aa904d8_1600x455.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rVXK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F417135f9-7fc9-4a55-a109-078e4aa904d8_1600x455.png" width="1456" height="414" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/417135f9-7fc9-4a55-a109-078e4aa904d8_1600x455.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:414,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rVXK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F417135f9-7fc9-4a55-a109-078e4aa904d8_1600x455.png 424w, https://substackcdn.com/image/fetch/$s_!rVXK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F417135f9-7fc9-4a55-a109-078e4aa904d8_1600x455.png 848w, https://substackcdn.com/image/fetch/$s_!rVXK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F417135f9-7fc9-4a55-a109-078e4aa904d8_1600x455.png 1272w, https://substackcdn.com/image/fetch/$s_!rVXK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F417135f9-7fc9-4a55-a109-078e4aa904d8_1600x455.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>As you can see, the Variant conversion fluctuates significantly at first but then normalizes and begins getting closer to the mean.</p><p>In the examples above, Variant conversion changes a lot during the first 10-12 days and then stabilizes. The chart on the right is a more extreme example to illustrate how impactful and prolonged RTM can be. In my experience, I often see it stabilizing after around a week or so, eventually looking more like this:&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rT3e!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff117869b-588c-48e4-bc76-ff7c4c22103e_1360x762.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rT3e!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff117869b-588c-48e4-bc76-ff7c4c22103e_1360x762.png 424w, https://substackcdn.com/image/fetch/$s_!rT3e!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff117869b-588c-48e4-bc76-ff7c4c22103e_1360x762.png 848w, https://substackcdn.com/image/fetch/$s_!rT3e!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff117869b-588c-48e4-bc76-ff7c4c22103e_1360x762.png 1272w, https://substackcdn.com/image/fetch/$s_!rT3e!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff117869b-588c-48e4-bc76-ff7c4c22103e_1360x762.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rT3e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff117869b-588c-48e4-bc76-ff7c4c22103e_1360x762.png" width="1360" height="762" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f117869b-588c-48e4-bc76-ff7c4c22103e_1360x762.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:762,&quot;width&quot;:1360,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rT3e!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff117869b-588c-48e4-bc76-ff7c4c22103e_1360x762.png 424w, https://substackcdn.com/image/fetch/$s_!rT3e!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff117869b-588c-48e4-bc76-ff7c4c22103e_1360x762.png 848w, https://substackcdn.com/image/fetch/$s_!rT3e!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff117869b-588c-48e4-bc76-ff7c4c22103e_1360x762.png 1272w, https://substackcdn.com/image/fetch/$s_!rT3e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff117869b-588c-48e4-bc76-ff7c4c22103e_1360x762.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Depending on the daily sample volume, your test will show something else. Once you have completed a few tests, it becomes easier to identify the test stages relevant to only your product and tested traffic volume. These stages will guide you in estimating the test timeline to reach significance and decision-making, especially for the &#8220;fast&#8221; tests.</p><p>For the example above, for Bayesian tests, I segment 3 lifecycle stages:&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ve-y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9543c4f-8c06-4cd1-87c8-26fbb56f662d_1290x880.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ve-y!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9543c4f-8c06-4cd1-87c8-26fbb56f662d_1290x880.png 424w, https://substackcdn.com/image/fetch/$s_!Ve-y!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9543c4f-8c06-4cd1-87c8-26fbb56f662d_1290x880.png 848w, https://substackcdn.com/image/fetch/$s_!Ve-y!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9543c4f-8c06-4cd1-87c8-26fbb56f662d_1290x880.png 1272w, https://substackcdn.com/image/fetch/$s_!Ve-y!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9543c4f-8c06-4cd1-87c8-26fbb56f662d_1290x880.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ve-y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9543c4f-8c06-4cd1-87c8-26fbb56f662d_1290x880.png" width="1290" height="880" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c9543c4f-8c06-4cd1-87c8-26fbb56f662d_1290x880.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:880,&quot;width&quot;:1290,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ve-y!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9543c4f-8c06-4cd1-87c8-26fbb56f662d_1290x880.png 424w, https://substackcdn.com/image/fetch/$s_!Ve-y!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9543c4f-8c06-4cd1-87c8-26fbb56f662d_1290x880.png 848w, https://substackcdn.com/image/fetch/$s_!Ve-y!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9543c4f-8c06-4cd1-87c8-26fbb56f662d_1290x880.png 1272w, https://substackcdn.com/image/fetch/$s_!Ve-y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9543c4f-8c06-4cd1-87c8-26fbb56f662d_1290x880.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><ol><li><p>The first stage can be from a few hours to a few days and serves as a validation step. This is the time when you pick random users and check their attributes, distribution, tenure, etc.</p></li><li><p>The second stage is vital and is usually short (often just a day or two). It is a turning point for estimating significance. (It doesn&#8217;t mean the data is significant. But at this stage, for analysts, it&#8217;s usually easy to confirm how long the test will be running).</p></li><li><p>The third stage is the longest and the most stable. The test is getting confidence during these days. Your conversions are not expected to move significantly afterward.</p></li></ol><p>In your analysis, you treat these stages differently. Changing test traffic volume or setup during the first or the second stage is dangerous because you are likely to run into strong RTM. That being said, making test alterations during the third stage can be a compromise, which would be to move faster but at the cost of a lower degree of confidence.</p><p>Other definitions of RTM I run into are &#8220;random measurement error&#8221; or &#8220;non-systemic fluctuations around the true mean&#8221;. RTM is quite tricky and makes the test data you receive look like a meaningful result (when it isn&#8217;t!). It becomes even more concerning when you accept it as a true test outcome and then run other tests against this &#8220;winning&#8221; group. And now you&#8217;ve ended up with multiple flawed tests (which is exactly what split tests do).</p><h1>Why does RTM occur?&nbsp;</h1><p>To simplify, you are likely to run into this effect when:</p><h3><strong>Your samples are not randomly distributed.</strong></h3><p>You are at the mercy of your experimentation instrumentation tooling. If it doesn&#8217;t do a great job of randomly disseminating users (which is often tricky to validate), there isn&#8217;t much you can realistically do except lessen your trust in your A/B tests.</p><h3><strong>Multiple rollout stages, sample size, or variance change.&nbsp;</strong></h3><p>These are the slow rollouts. Let&#8217;s say you launch Control and 2 Variants to 25% traffic with a new 34/33/33 split. After a few days, you remove Variant 2 and launch Control and 1 Variant with a 50/50 split to 50% traffic. You are way more likely to end up with a high probability of RTM.</p><h3><strong>Your target audience has too many properties.&nbsp;</strong></h3><p>The more attributes your target audience has, the higher the opportunity is for RTM. For example, to be qualified for the test, your users have to be in a particular country, use a specific language, be paid subscribers or complete at least 4 transactions, are eligible for a trial or a promo, and have particular settings enabled on their profile, or else. All of these are user attributes that qualify them to receive a test experience. The fewer attributes you apply, the cleaner (and faster) your test will ultimately go.&nbsp;&nbsp;</p><h1>How to prevent RTM?&nbsp;</h1><p>In truth, there is nothing you can do to eliminate the effect, but you can reduce the probability that it will affect your test. Here are some basic common protection practices:&nbsp;</p><ol><li><p>Before any experimentation, know your expected conversion (baseline). Ideally, measure it multiple times, know its range, and compare averages. This will help you identify if your Variant conversion is way off or suspiciously high or low. Don&#8217;t trust Control as your true Baseline.</p></li><li><p>Avoid running experiments on complex user groups that have many properties. The more attributes you introduce, the more complex the experiment becomes.</p></li><li><p>The test groups should be randomized and normally distributed (avoid mixing new and existing users).</p></li><li><p>Avoid slow and disproportional rollouts. For example, once the test has been released to 10% traffic, do not reduce its size.&nbsp;</p></li><li><p>Don&#8217;t stop the test earlier than planned, even if it appears you reached significance or you have a clear winner or loser.&nbsp;</p></li></ol><p>If you have to finish the test ASAP and don&#8217;t trust the results, you can run <a href="https://www.statisticshowto.com/ancova/">the analysis of co-variance ANCOVA</a> (ANOVA+regression). It used to be easy to run in SPSS, SAS, Stata, or other old-world statistical packages, but now you can also run in <a href="https://www.statology.org/ancova-python/">Python Pandas</a>.</p><p>If you don&#8217;t have the resources or time to run ANOVA, here is my &#8220;rule of thumb,&#8221; which you&#8217;re welcome to borrow (with a degree of caution):</p><p><a href="https://emojipedia.org/large-green-circle/">&#128994;</a> If variant conversion doesn&#8217;t fluctuate and stays the same throughout vital stages of the test, it&#8217;s likely trusted. You can end the test early, expand more traffic to it, and move on.</p><p><a href="https://emojipedia.org/large-yellow-circle/">&#128993;</a> If variant conversion fluctuates during the early stage but stabilizes, you will likely be dealing with not normalized or randomly distributed samples. If this is a common effect, it means that your experimental instrumentation is not good. You can&#8217;t go fast with these tests. You have to run it long enough until you get significant data for a prolonged period. Lots of validations and checks are needed for such tests.&nbsp;</p><p>&#128308; If variant conversion fluctuates during all stages of the test, and the significance is reached with a small difference from Control, you are likely dealing with an inconclusive test. Disregard its result and relaunch with a new audience or approach.&nbsp;</p><p>Hopefully, the early pattern you see stays the same or better and becomes more emerging throughout the test. Just don&#8217;t be careless with statistics and follow all the necessary steps and practices to ensure the testing effect you see doesn&#8217;t happen by chance (and remember, there is always a probability it does).</p><h3>Learn more on RTM and test setup:</h3><ul><li><p><a href="https://www.makethebrainhappy.com/2018/12/regression-to-the-mean.html">MakeTheBrainHappy - Regression To The Mean</a></p></li><li><p><a href="https://fs.blog/regression-to-the-mean/">Regression Toward the Mean: An Introduction with Examples</a></p></li><li><p><a href="https://conversion-uplift.co.uk/post/why-are-most-ab-test-results-a-lie/">Why Are Most A/B Test Results A Lie?</a></p></li><li><p><a href="https://f.hubspotusercontent00.net/hubfs/215600/qubit-research-ab-test-results-are-illusory-1.pdf">Most Winning A/B Test Results are Illusory</a></p></li><li><p><a href="https://www.growth-catalyst.in/p/what-goes-wrong-in-ab-tests">What Goes Wrong in A/B Tests</a></p></li><li><p><a href="https://www.evanmiller.org/how-not-to-run-an-ab-test.html">How Not To Run an A/B Test</a></p></li></ul><p>Thanks for reading, everyone. Until next Wednesday!</p>]]></content:encoded></item><item><title><![CDATA[How To Run An A/B Testing On Low Traffic - Issue 181]]></title><description><![CDATA[Strategy and solutions for effective A/B Tests with small samples]]></description><link>https://dataanalysis.substack.com/p/how-to-run-an-ab-testing-on-low-traffic</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/how-to-run-an-ab-testing-on-low-traffic</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 10 Jan 2024 13:02:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!tRG5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d64f663-d16b-4ef8-9125-7891b26fff21_1600x547.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to the <a href="https://dataanalysis.substack.com/">Data Analysis Journal</a>, a weekly newsletter about data science and analytics.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://dataanalysis.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://dataanalysis.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p><em>Before we dive in, I wanted to ask a favor to fill out <a href="https://s.surveyplanet.com/hdui562w">this brief survey</a> (And thank you so much to those who have already done it! You are the best!</em> &#10024;<em>). As my audience grows, I&#8217;d like to better understand how my journal can stay relevant and bring more value.&nbsp;</em></p><div><hr></div><p>The most common question I have received from my readers has been how to do A/B testing with very low traffic.</p><p>There are no specific rules or requirements dictating when to start A/B testing or the minimum number of users necessary to launch a test. There are various opinions on how to approach experimentation with small sample sizes, ranging from more conservative &#8220;<em>You should not run an A/B test until you reach 30,000 users per variant period</em>&#8221; to the more progressive &#8220;<em>It&#8217;s fine, just adjust your thresholds</em>.&#8221;</p><p>Today, I will share my guide on the procedure, specifics, and caveats of experimenting with low traffic. I will cover topics such as the minimum number of users needed to launch an A/B test and offer suggestions on increasing confidence and trust in small sample tests.&nbsp;</p><p>I&#8217;ll present it in a Q&amp;A format to ensure I address all the questions I received from readers regarding low-traffic tests (I apologize for the delay in addressing this):</p><ol><li><p>What is the minimum number of users needed in a sample size for an A/B test?</p></li><li><p>Can we run a test on a very small sample (~50 users in Variant per day) and wait longer to reach significance?</p></li><li><p>What other factors should we consider to ensure we make the right decision?</p></li><li><p>How do you figure out the trade-off between confidence and test timeline? For example, can we test small samples with an 85% confidence instead of 95%? Will the test run faster?</p></li><li><p>What are some rules or testing limitations in low traffic to address to increase statistical rigor?</p></li></ol><h5><em>If you need a refresher on the A/B testing concept or procedure, read the <a href="https://dataanalysis.substack.com/p/ab-test-checklist">A/B Test Checklist</a>.</em></h5><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3K0V!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe865542-8911-4545-969f-4d26a36e0002_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3K0V!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe865542-8911-4545-969f-4d26a36e0002_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!3K0V!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe865542-8911-4545-969f-4d26a36e0002_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!3K0V!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe865542-8911-4545-969f-4d26a36e0002_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!3K0V!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe865542-8911-4545-969f-4d26a36e0002_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3K0V!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe865542-8911-4545-969f-4d26a36e0002_200x200.png" width="200" height="200" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fe865542-8911-4545-969f-4d26a36e0002_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3K0V!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe865542-8911-4545-969f-4d26a36e0002_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!3K0V!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe865542-8911-4545-969f-4d26a36e0002_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!3K0V!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe865542-8911-4545-969f-4d26a36e0002_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!3K0V!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe865542-8911-4545-969f-4d26a36e0002_200x200.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="pullquote"><p>&#8220;When running online experiments, getting numbers is easy; getting numbers you can trust is hard.&#8221; - <a href="https://www.exp-platform.com/Documents/IEEE2010ExP.pdf">Online Experiments: Practical Lessons, Microsoft</a>.</p></div><blockquote><h3><strong>Q1: Is there a set rule for the minimum number of users required to launch an A/B test?</strong></h3></blockquote>
      <p>
          <a href="https://dataanalysis.substack.com/p/how-to-run-an-ab-testing-on-low-traffic">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[An Analysis Of Bias Or Why A/B Testing Fails - Issue 172]]></title><description><![CDATA[A recap of Stanford and Airbnb's collaborative paper on experimentation setup and analysis in two-sided platforms and marketplaces.]]></description><link>https://dataanalysis.substack.com/p/an-analysis-of-bias-or-why-ab-testing</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/an-analysis-of-bias-or-why-ab-testing</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 15 Nov 2023 13:00:57 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!8TcY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6feedf14-c725-4a87-a09e-da74e39e9fbf_1500x1036.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A few years ago, Stanford collaborated with Airbnb on a fascinating research study - <em><strong><a href="https://arxiv.org/pdf/2002.05670.pdf">Experimental Design in Two-Sided Platforms: An Analysis of Bias</a></strong></em><strong>.</strong> Last month, this paper was recognized with the<a href="https://www.informs.org/Recognizing-Excellence/INFORMS-Prizes/George-B.-Dantzig-Dissertation-Award"> Dantzig Dissertation Award</a> and the <a href="https://www.informs.org/Recognizing-Excellence/Community-Prizes/Manufacturing-and-Service-Operations-Management/MSOM-Service-Management-SIG-Best-Paper-Award">MSOM Service Management SIG Best Paper Award</a>.</p><p>It&#8217;s a fascinating study that should be close to any data scientist&#8217;s heart who works with A/B testing and who has, at some point, given up explaining why rapid iteration may not work and why A/B testing often fails.</p><p>The study offers a new way of setting up A/B tests and a new perspective on experimentation. Today, I am sharing a high-level recap of this research and integrating its theory and findings into our daily work. I&#8217;ll review the typical A/B test setups in marketplaces, how they fall short, and what methods the new research offers to address the bias.</p><p>I also want to emphasize how crucial it is for analysts to understand the complexity of interference, how dangerous bias is for the test analysis, and how we can continue to develop ways to navigate communication and uncertainty.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jxVw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9fe3a79-3f36-46c1-ba52-0282aa7270aa_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jxVw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9fe3a79-3f36-46c1-ba52-0282aa7270aa_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!jxVw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9fe3a79-3f36-46c1-ba52-0282aa7270aa_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!jxVw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9fe3a79-3f36-46c1-ba52-0282aa7270aa_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!jxVw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9fe3a79-3f36-46c1-ba52-0282aa7270aa_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jxVw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9fe3a79-3f36-46c1-ba52-0282aa7270aa_200x200.png" width="200" height="200" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a9fe3a79-3f36-46c1-ba52-0282aa7270aa_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jxVw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9fe3a79-3f36-46c1-ba52-0282aa7270aa_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!jxVw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9fe3a79-3f36-46c1-ba52-0282aa7270aa_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!jxVw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9fe3a79-3f36-46c1-ba52-0282aa7270aa_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!jxVw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9fe3a79-3f36-46c1-ba52-0282aa7270aa_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>A/B testing is one of the methods of proving causation. If you want to prove A is causing B, you need to do two things: A/B test it, and then A/B test it <em>right</em>.</p><p>A recap from <a href="https://dataanalysis.substack.com/p/how-to-prove-causation-issue-148">How To Prove Causation</a>:</p><blockquote><p>Causal inference methods include <strong>hypothesis testing</strong> (experimentation) and <strong>observations</strong> (user research). Keep in mind that experimentation (A/B tests, multivariate tests, split tests, etc.) and user research (surveys, field studies, interviews, etc.) don&#8217;t guarantee you clean causality; it&#8217;s subject to randomization, significance, confidence levels, treatment group size, setup, and more.</p></blockquote><p>If you have been reading my newsletter, you should know that the success of your test depends on: </p><ul><li><p>Instrumentation</p></li><li><p>Data maturity</p></li><li><p>The process should be tailored to reflect the nature of your product and business. </p></li></ul><p>(read more: <a href="https://dataanalysis.substack.com/p/embracing-the-new-era-of-accelerated">Embracing the New Era of Accelerated Testing</a>)</p><p>This academic study reiterates my points about the importance of test instrumentation. The test design should appropriately address your unique product offering, customer, and market. <em><strong>If you can&#8217;t validate or prove the test setup, or if you question the ability of your instrumentation to allocate users accurately, your test analysis might be wrong and lead to incorrect assessments.</strong></em></p><p>Another reason why this particular research stands out is that <a href="https://profiles.stanford.edu/ramesh-johari">Ramesh Johari</a> himself led it. Ramesh Johari is a famous scientist, researcher, and professor who spent almost 20 years learning experimentation and has been advising Optimizely, Stitch Fix, Upwork, Airbnb, Uber, Bumble, Stripe, and more. </p><h4>&#128161; Make sure to listen to his recent interview in <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Lenny's Newsletter&quot;,&quot;id&quot;:10845,&quot;type&quot;:&quot;pub&quot;,&quot;url&quot;:&quot;https://open.substack.com/pub/lenny&quot;,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/40b73d36-7447-4f52-83b5-4bbf44b9324a_1200x1200.png&quot;,&quot;uuid&quot;:&quot;eaf87abf-126f-4b2e-b671-53ae9b1864ff&quot;}" data-component-name="MentionToDOM"></span> - <a href="https://www.lennyspodcast.com/marketplace-lessons-from-uber-airbnb-bumble-and-more-ramesh-johari-stanford-professor-startup/">Marketplace lessons from Uber, Airbnb, Bumble, and more</a>:</h4><blockquote><p>&#8220;Many of the changes that are most consequential create winners and losers. And rolling with those changes is about recognizing whether the winners you've created are more important to your business than the losers you've created in the process.&#8220;</p></blockquote><h2>What this research is about:</h2><p>This 56-page academic study offers a framework to better understand how to set up and analyze experiments in <strong>two-sided marketplaces</strong> - products that offer interactions between different user groups (like buyers and sellers, readers and writers, listings and customers). The study is relevant to any platform with a &#8220;dynamic inventory model&#8221; - bookings, services, content, or tutorials, such as Upwork, Udemy, Airbnb, Substack, Uber, eBay, Etsy, Medium, Twitter, Instagram, Meta, and more.</p><h2>The challenge: there is no easy way to eliminate interference.</h2><p>This study focuses on the <strong>interference effect</strong>, where a treatment applied to one group of users unintentionally impacts another group. This phenomenon leads to bias and incorrect estimations, preventing an accurate test read.</p><h3>How tests are run today in marketplaces.</h3>
      <p>
          <a href="https://dataanalysis.substack.com/p/an-analysis-of-bias-or-why-ab-testing">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Embracing the New Era of Accelerated Testing - Issue 150]]></title><description><![CDATA[How to adapt your data science and analytics teams to support the ever-changing framework of A/B tests successfully.]]></description><link>https://dataanalysis.substack.com/p/embracing-the-new-era-of-accelerated</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/embracing-the-new-era-of-accelerated</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 28 Jun 2023 12:01:01 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!SyPw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8dabe88-289d-41e6-8c3e-b99f3a5c2ede_1522x890.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello, and welcome to my<a href="https://dataanalysis.substack.com/"> Data Analytics Journal</a>, where I write about data science and product analytics.</p><p>This month paid subscribers learned about:</p><ul><li><p><a href="https://dataanalysis.substack.com/p/decoding-regression-scores-issue">Decoding Regression Scores</a> - Linear Regression Part 2: How to read regression plots and interpret the equation that drives analysis and predictions.</p></li><li><p><a href="https://dataanalysis.substack.com/p/how-to-prove-causation-issue-148">How To Prove Causation</a> - If correlation doesn&#8217;t imply causation, then what does? Examples of causation analysis, how to use regression for decision-making, and how to recognize when the regression pattern you see is correct and trusted.</p></li><li><p><a href="https://dataanalysis.substack.com/p/a-deep-dive-into-user-engagement">A Deep Dive Into User Engagement Through Tricky Averages</a> - A reminder of how averages can be misleading. Math and steps to report &#8220;average per user per day&#8221; KPI to measure the depth of user engagement.</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://dataanalysis.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://dataanalysis.substack.com/subscribe?"><span>Subscribe now</span></a></p></li></ul><p>A few weeks ago, I went to the first reopened <a href="https://lyftwimldslightningtalks2023.splashthat.com/">DS&amp;ML meetup series organized by WiMLDS and Lyft</a>. One of the delivered talks there about <em>Demonstrating Leadership through AB Testing </em>stuck with me and got me thinking. And thinking.</p><p>The talk was great. It was essentially a textbook-like A/B test methodology nicely laid out in a concise and structured way, illustrating foundational aspects with a solid case study shared by the <a href="https://www.pandora.com/">Pandora</a> product analytics team. And yet, when I tried to map the described way of the A/B tests framework into the current development workflow, it was clear it won&#8217;t set my team up for success.&nbsp;</p><p>The modern boosted &#8220;optimization&#8221; trend makes me question its methods and wonder if the <strong>current school of experimentation is actually set up for it</strong>. It seems to me that the way of doing hypothesis testing, which we learned back in statistical class, doesn&#8217;t serve us well anymore. In order to succeed, we have to make questionable compromises, break some boundaries, and bend backward to make it all work.&nbsp;</p><p>So today I&#8217;ll talk about the dissonance between tempting &#8220;optimize your app in no time and effort using our tool to grow 10x faster&#8221; and being data-driven in its true meaning - making sound decisions based on facts and proven insights that you can trust and validate.&nbsp;</p><p>And the biggest question I have on my plate is how do you set the data science and analytics team up for success to make sure you <strong>empower this new accelerated &#8220;optimization&#8221; culture while keeping the confidence and accuracy of the insights you share</strong>?&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!smvn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff71b52e3-768b-4b7a-a27f-8239f692c04d_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!smvn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff71b52e3-768b-4b7a-a27f-8239f692c04d_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!smvn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff71b52e3-768b-4b7a-a27f-8239f692c04d_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!smvn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff71b52e3-768b-4b7a-a27f-8239f692c04d_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!smvn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff71b52e3-768b-4b7a-a27f-8239f692c04d_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!smvn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff71b52e3-768b-4b7a-a27f-8239f692c04d_200x200.png" width="190" height="190" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f71b52e3-768b-4b7a-a27f-8239f692c04d_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:190,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!smvn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff71b52e3-768b-4b7a-a27f-8239f692c04d_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!smvn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff71b52e3-768b-4b7a-a27f-8239f692c04d_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!smvn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff71b52e3-768b-4b7a-a27f-8239f692c04d_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!smvn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff71b52e3-768b-4b7a-a27f-8239f692c04d_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p><a href="https://dataanalysis.substack.com/p/the-new-age-of-ab-tests-issue-99">As I mentioned before</a>, A/B testing is often the culprit of tension between analysts and product owners. After reading success stories from <a href="https://developers.facebook.com/videos/f8-2017/how-we-shipped-reactions/">Meta (How We Shipped Reactions)</a>, <a href="https://engineering.linkedin.com/blog/2019/06/detecting-interference--an-a-b-test-of-a-b-tests">LinkedIn (Detecting interference: An A/B test of A/B tests)</a>, or <a href="https://medium.com/airbnb-engineering/experiments-at-airbnb-e2db3abf39e7">Airbnb (Experiments at Airbnb)</a>, many product leaders are inspired to follow the trend of constantly iterating with user flows, copies, CTAs positioning, layouts, etc. Then analysts come in and question the test objective, setup, baseline metrics, and experimentation toolkit. Quite often they have to push back on the test launch, delay the test results readout, or even disregard them. They might appear to hinder progress when they&#8217;re really just trying to do their job.</p><p><strong>Your ability to test is linearly correlated with how well your analytics is set up. This comes down to (A) the testing instrumentation, (B) analytics maturity, and (C) test procedure and protocols.</strong></p><p>Now, let&#8217;s imagine you have:&nbsp;</p><ol><li><p>The instrumentation of tomorrow that &#8220;streamlines your feature flagging&#8221; and allows you to &#8220;build and deploy in 10 minutes flat&#8221; and &#8220;easily control multiple flags at once&#8221; (<a href="https://www.kameleoon.com">Kameleoon</a>, <a href="https://superwall.com/">Superwall</a>, <a href="https://launchdarkly.com/">LaunchDarkly</a>, <a href="https://devcycle.com/">DevCycle</a> to name a few).&nbsp;</p></li><li><p>On top of it, let&#8217;s pretend you work at some mystical and magical workplace with a solid analytical foundation set, maintained by dedicated full-time data governance champions. You have your events, definitions, data attributes, and schemas in order (don&#8217;t believe in tooling for this).&nbsp;</p></li></ol><p>Then what? The success of experimentation will come down to the same old standard:&nbsp;</p><ol start="3"><li><p><strong>Test procedure and protocols.</strong></p></li></ol><p>But here&#8217;s the thing. The current &#8220;academic&#8221; workflow we have been taught to follow (and by which I <a href="https://dataanalysis.substack.com/p/ab-test-checklist">stand by</a>) ironically will fail you once you have a solid instrumentation and dream data&amp;analytics governance. And here&#8217;s why.</p><h1><strong>A/B tests become a never-ending lifecycle of quick optimizations</strong></h1><p>Testing today, especially on mobile, evolves into a never-ending series of short A/B tests. There is no beginning, no end of a product feature iteration. It's an eternal evolving spiraling cycle that we may never escape from. And this is a new thing aggressively boosted with modern tooling for app development. </p><p>From <strong><a href="https://superwall.com/">Superwall</a></strong>:</p><blockquote><p>&#8220;We actually allow you to iterate an experiment in flight. Let's say you have a (A: 50) (B: 50) setup and B starts winning, you can maintain the existing users in their groups and either (1) Start assigning more users into B to minimize the opportunity cost of the test, while slowly marching to stat sig or (2) Introduce C, while still assigning users to (A:25) (B:25) (C:50) to quickly iterate. It's not a statistically perfect method but allows for directionally correct iteration w/o having to wait for the experiment to complete. It's something some of our most successful teams do.&#8221;</p></blockquote><p>Our previous academic approach to data handling and evaluation for a test simply doesn't work anymore. <strong>Most analytics teams are not equipped or skilled for this</strong>. We were taught to treat every A/B test as a project - create a ticket for it, assign an id for the experiment, estimate the timeline to reach significance, document a case study for each test, have analysis ready, checks, 3-month follow-ups, <a href="https://dataanalysis.substack.com/p/playbook-for-launching-monitoring">you know the story</a>.&nbsp;</p><p>Now this "old-school" approach doesn't work anymore. It holds you back, it blocks teams, and it's not scalable or efficient.&nbsp;</p><p>I am astonished about how today's modern tooling accelerates optimizations. Today&#8217;s experimentation enabled by <a href="https://superwall.com/">Superwall</a>, <a href="https://qonversion.io/">Qonversion</a>, <a href="https://www.split.io/">Split</a>, and others set the bar very high. Just think about it: a given user is put into 5-10 different tests or "experiences&#8221; simultaneously, which are connected, overlap, or worse of all: <em>dependent on each other</em>. For example, you can test the onboarding flow, paywall conversions (which are connected to the onboarding flow), and accessibility of premium features (which are dependent on paywall conversion) for one user concurrently. It's <em>growth-fascinating</em> and <em>statistically-terrifying</em> at the same time.</p><p>Before, it was "an error" not to exclude users in your test analysis who got in various tests at the same time. Now, quite often you are forced to keep them to make it significant (mind you, most of the tests have slow rollouts and might be running at small % traffic on one mobile platform). You have to have some type of bias measure - how harmful, how polluted users are. It's not Control vs Variant anymore, it's Test_1 Control vs Test_2 Variant_A vs Test_3 Variant_D and possibly a branchy tree downstream of small samples. Teams don't realize they work with complex multi-variable tests, and winner Variant_B from Test_1 might not be "compatible" with a winner Variant_A from Test_2:&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SyPw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8dabe88-289d-41e6-8c3e-b99f3a5c2ede_1522x890.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SyPw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8dabe88-289d-41e6-8c3e-b99f3a5c2ede_1522x890.png 424w, https://substackcdn.com/image/fetch/$s_!SyPw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8dabe88-289d-41e6-8c3e-b99f3a5c2ede_1522x890.png 848w, https://substackcdn.com/image/fetch/$s_!SyPw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8dabe88-289d-41e6-8c3e-b99f3a5c2ede_1522x890.png 1272w, https://substackcdn.com/image/fetch/$s_!SyPw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8dabe88-289d-41e6-8c3e-b99f3a5c2ede_1522x890.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SyPw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8dabe88-289d-41e6-8c3e-b99f3a5c2ede_1522x890.png" width="612" height="357.70054945054943" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e8dabe88-289d-41e6-8c3e-b99f3a5c2ede_1522x890.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:851,&quot;width&quot;:1456,&quot;resizeWidth&quot;:612,&quot;bytes&quot;:301156,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SyPw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8dabe88-289d-41e6-8c3e-b99f3a5c2ede_1522x890.png 424w, https://substackcdn.com/image/fetch/$s_!SyPw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8dabe88-289d-41e6-8c3e-b99f3a5c2ede_1522x890.png 848w, https://substackcdn.com/image/fetch/$s_!SyPw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8dabe88-289d-41e6-8c3e-b99f3a5c2ede_1522x890.png 1272w, https://substackcdn.com/image/fetch/$s_!SyPw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8dabe88-289d-41e6-8c3e-b99f3a5c2ede_1522x890.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em><strong>&#8220;Simpson&#8217;s Paradox:</strong></em></p><p><em>A trend or result that is present when data is put into groups that reverses or disappears when the data is combined.&#8221; read more about <a href="https://towardsdatascience.com/simpsons-paradox-and-interpreting-data-6a0443516765">Simpson&#8217;s Paradox and Interpreting Data</a>.</em></p><p>&#128681;To reiterate, within a one-time frame, you won't be able to confirm for multiple tests if the lift they demonstrate (a) didn&#8217;t happen by chance and (b) is best suitable with all the iterations actively tested.&nbsp;</p><p>I don&#8217;t intend to block or hinder experimentation. It&#8217;s exciting what the latest feature flagging and experimentation tools can do. But by the end of the day, it&#8217;s my team's responsibility to represent trust, confidence, and correctness of the test inference. Here are some of the things I am trying out or that proved to back me up.&nbsp;&nbsp;&nbsp;</p><h2>&#9989; <strong>How to adapt your team to the thriving era of experimentation</strong></h2><ol><li><p><strong>Pre/Post analysis</strong>: always keep a check on high-level KPI. Regardless of what test impact you see, the ecosystem metrics (usually reported monthly) don&#8217;t lie. Set up reminders and follow up in 30 days, or 3 months, and keep it high level. If you can't do it for every test, do it for a series of product iterations - onboarding flow, a new experience for a new feature, the initial activity, etc.</p></li><li><p><strong>Assign an owner for a product feature (with its baseline metrics):</strong> A few years ago, I&#8217;d take the total volume of active A/B tests and equally divide it between my analysts, with more complex tests (like new price, product adoption, notification frequency) assigning to senior analysts and more straight forward (onboarding, upsell, UI change tests) to junior. Today, however, I am leaning towards solely dedicating an analyst to a product feature or experience (e.g. onboarding, activation, conversion, engagement, subscriptions) and letting them own their domain with whatever iteration the product team is testing, regardless of its level of complexity.</p></li><li><p><strong>Automate statistical checks</strong>: test validations (variance, distribution, randomness) have to be automated today. Assign a dedicated DS person to oversee statistics across all tests, and let them own and automate it (via views or dashboarding). For example, back in my time at <a href="https://vidiq.com/">VidIQ</a>, we introduced a consolidated experimentation dashboard that returned us a table with z score, p-value, variance, volume, and user distribution for every active test. It was such a time saver. Don't do manual checks for every test in your impact analysis, it&#8217;s not scalable.&nbsp;</p></li><li><p><strong>Streamline documentation</strong>: your team has to simplify the test evaluation read, and it includes both communication on it and documentation. Readjust your expectation for the readout, for example, not a deck per test, but a slide per test. With such a volume of iterations, detailed notebooks and deep-dive impact analysis per test are not realistically doable anymore. We passed that stage when modern SaaS enabled PMs to make production changes with 1 click without coding or deploying.<strong>&nbsp;</strong></p><p>I live in the past and thus still make attempts to document every product initiative, big or small, to understand how metrics move and user behavior evolves. Moving forward, I am leaning towards documenting only big changes that aim to disrupt user flow or behavior.&nbsp;</p></li><li><p><strong>Eliminate the bias:</strong> At this very moment, I am brainstorming with my fellow analysts to create a way to quantify the bias of users getting into multiple tests. For example, create a new attribute of a cumulative number of currently active tests per user across revenue, onboarding, activation, core usage, secondary features, etc. I&#8217;m envisioning something like this:&nbsp;</p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!alcN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96542cc3-7685-428c-bc5d-d72bfae3886c_1600x246.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!alcN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96542cc3-7685-428c-bc5d-d72bfae3886c_1600x246.png 424w, https://substackcdn.com/image/fetch/$s_!alcN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96542cc3-7685-428c-bc5d-d72bfae3886c_1600x246.png 848w, https://substackcdn.com/image/fetch/$s_!alcN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96542cc3-7685-428c-bc5d-d72bfae3886c_1600x246.png 1272w, https://substackcdn.com/image/fetch/$s_!alcN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96542cc3-7685-428c-bc5d-d72bfae3886c_1600x246.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!alcN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96542cc3-7685-428c-bc5d-d72bfae3886c_1600x246.png" width="1456" height="224" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/96542cc3-7685-428c-bc5d-d72bfae3886c_1600x246.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:224,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!alcN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96542cc3-7685-428c-bc5d-d72bfae3886c_1600x246.png 424w, https://substackcdn.com/image/fetch/$s_!alcN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96542cc3-7685-428c-bc5d-d72bfae3886c_1600x246.png 848w, https://substackcdn.com/image/fetch/$s_!alcN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96542cc3-7685-428c-bc5d-d72bfae3886c_1600x246.png 1272w, https://substackcdn.com/image/fetch/$s_!alcN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96542cc3-7685-428c-bc5d-d72bfae3886c_1600x246.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>If we would have this table, then we would be able to cherry-pick users with the least conflicting test overlaps for the test read. Then, maybe, we could wrap it into a downstream view of something like test_overlap_bias_score being between 1-10. This way, we could use users with 1-2 scores for high-impact test analysis for some important decision-making, and then compare it to users with 5-7 scores. Again, just brainstorming here. This is totally doable for companies with matured analytics.&nbsp;</p><p>Or we could try this solution shared by <strong><a href="https://www.geteppo.com/">Eppo</a></strong>:</p><blockquote><p>&#8220;Another method can be to run a regression every day on a user-level dataset predicting a core metric. In that regression, including a covariance for every active experiment and also every <strong>pair</strong> of experiments. If the pair-wise experiment coefficients show up as significant to some threshold, you'd see that there's an interaction effect issue.&#8221;</p></blockquote><h2><strong>Best practice shared by experimentation experts: </strong>&nbsp;</h2><p><strong><a href="https://superwall.com/">Superwall</a>:&nbsp;</strong></p><blockquote><p>&#8220;When using Superwall we'll automatically report back to you the inclusion of a user in an experiment. People usually then set that as an attribute on their user model to be able to break down by all the permutations of test variants.&#8221;.</p></blockquote><p><strong><a href="https://qonversion.io/">Qonversion</a>:</strong></p><blockquote><p>&#8220;There are a few different ways that help you avoid data contamination and optimize traffic for a large number of simultaneous experiments:</p><p>- Run mutually exclusive experiments. Usually, if experiments are located in the same app area and aimed at affecting the same metric, it&#8217;s highly recommended not to expose them to the same set of users. Such an approach does require having an advanced splitter at your disposal. However, eventually, you get crystal-clear data. For example, do not test onboarding changes and a journey to the Aha! moment on the same users.</p><p>- Expose your user to the experiment only once they reach a specific point in their user flow. For example, if you&#8217;re experimenting with a new paywall design, do not add users to the experiment on the app launch but once the paywall is shown.</p><p>- Keep a global control group (usually no more than 5%). This approach requires extra effort to maintain and correctly compare with your treatments. However, if you find it workable, you can avoid adding control groups to each experiment in favor of manually extracting the same segment for the comparison from the global control group.</p><p>Remember that having relevant users in your groups is crucial. Do not assign to the test those who are unaffected by the new experience at all. For example</p><p>- You&#8217;re validating the paywall-related monetization hypothesis, which means adding to the test users with active subscriptions is irrelevant.</p><p>- You&#8217;re testing new logic for guiding your users to the Aha! Moment, then exposing the test to those, who already have passed it, adds only extra noise to your data.&#8221;</p></blockquote><p><strong><a href="https://www.split.io/">Split</a>:</strong>&nbsp;</p><blockquote><p>&#8220;There is one area to keep in mind when running parallel experiments. That is the case where two changes directly interfere with one another, creating a different impact on behavior when combined than when in isolation. This happens when experiments are being run on the same page, the same user flow, etc. To avoid interaction effects, review concurrent experiments for interactions: use tags and naming conventions to track changed areas, manually test changes as part of the rollout, and look for other feature flags in nearby code. You should also design colliding tests to highlight interactions and compare each variant&#8217;s performance against one another.&#8221;</p></blockquote><div><hr></div><p>Remote configuration and rapid feature flagging open a new age of A/B testing. Nowadays, most experimentation toolkits are flexible and configurable. They empower any product leader to pretend that they have enough design and DS resources and can unleash their power of product iterations in no time. As exciting as that sounds, that means, as a data scientist, you won&#8217;t have the luxury of calculating the timeline for reaching significance, running the post-rollout impact check, working on a communication plan for every test, or creating a case study deck for every product change. Brace for a tsunami of a wild mix of kinda-test and not-a-full-rollout launches. And then: learn to swim in it.</p><p>Thanks for reading, everyone. Until next Wednesday!&nbsp;</p><h3><strong>Related publications:</strong></h3><ul><li><p><a href="https://dataanalysis.substack.com/p/how-to-develop-a-highly-trusted-experiment">How To Develop a Highly Trusted Experiment Analysis Workflow</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/playbook-for-launching-monitoring">Playbook For Launching, Monitoring, and Analyzing A/B Tests</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/ab-test-checklist">A/B Test Checklist</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/build-or-buy-how-we-developed-a-platform">Build or Buy: How We Developed A Platform For A/B Tests</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/why-you-shouldnt-stop-ab-tests-early">Why You Shouldn&#8217;t Stop A/B Tests Early</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/how-to-prove-causation-issue-148">How To Prove Causation</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/5-mistakes-to-avoid-when-running-9c3">5 Mistakes To Avoid When Running A/B Tests</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/how-to-get-randomly-distributed-users">How To Get Randomly Distributed Users in SQL</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/growth-loops-and-some-hard-truths">Growth, Loops, And Some Hard Truths - A Recap Of Amplitude Cohort 2022</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/how-we-optimized-the-onboarding-funnel">How we optimized the onboarding funnel by 220%</a></p></li><li><p><a href="https://dataanalysis.substack.com/p/love-and-hate-between-product-and">Love And Hate Between Product Management And Analytics</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Playbook For Launching, Monitoring, and Analyzing A/B Tests - Issue 134]]></title><description><![CDATA[A framework of product experimentation - the procedure and analysis.]]></description><link>https://dataanalysis.substack.com/p/playbook-for-launching-monitoring</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/playbook-for-launching-monitoring</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 01 Mar 2023 13:01:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!b6Yf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95e4a893-8f16-49de-a81f-5371ff0dbaea_1600x916.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to the <a href="https://dataanalysis.substack.com/">Data Analysis Journal</a>, a weekly newsletter about data science and analytics.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://dataanalysis.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://dataanalysis.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>Today, I will share my step-by-step process for launching, monitoring, analyzing, and reporting A/B tests. I&#8217;ll also cover the roles and expectations in test monitoring and support between product managers and data analysts, as this area of responsibility can overlap and is often <a href="https://dataanalysis.substack.com/p/love-and-hate-between-product-and">a culprit of tensions</a>.&nbsp;</p><p>Through working at different companies, I ended up developing a few different frameworks that were each depending on the data team structure and analyst reporting role. Depending on if analysts are embedded into Product, Business, or Engineering, or if they are part of a team squad or a tiger group, the framework will be different. That said, for today I&#8217;ll keep it at a high level to make it applicable to any organizational and reporting structure.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!g4wx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77aa0bbc-07bc-4793-94a4-e93dae7fe443_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!g4wx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77aa0bbc-07bc-4793-94a4-e93dae7fe443_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!g4wx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77aa0bbc-07bc-4793-94a4-e93dae7fe443_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!g4wx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77aa0bbc-07bc-4793-94a4-e93dae7fe443_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!g4wx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77aa0bbc-07bc-4793-94a4-e93dae7fe443_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!g4wx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77aa0bbc-07bc-4793-94a4-e93dae7fe443_200x200.png" width="166" height="166" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/77aa0bbc-07bc-4793-94a4-e93dae7fe443_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:166,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!g4wx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77aa0bbc-07bc-4793-94a4-e93dae7fe443_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!g4wx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77aa0bbc-07bc-4793-94a4-e93dae7fe443_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!g4wx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77aa0bbc-07bc-4793-94a4-e93dae7fe443_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!g4wx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77aa0bbc-07bc-4793-94a4-e93dae7fe443_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>First thing first. I prefer to differentiate experimentation between introducing a new feature or optimizing an existing feature and classify all product tests into 3 categories:&nbsp;</p><ol><li><p>Optimizing an existing product or feature.</p></li><li><p>Introducing a change to an existing product or feature.</p></li><li><p>Introducing a new product or feature that didn&#8217;t exist before.</p></li></ol><p>While these sound similar (and in many, many companies are treated the same), they actually have different lifecycles and rollouts and should be evaluated differently. </p>
      <p>
          <a href="https://dataanalysis.substack.com/p/playbook-for-launching-monitoring">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[A/B Test Checklist]]></title><description><![CDATA[A short guide to product experimentation steps and process.]]></description><link>https://dataanalysis.substack.com/p/ab-test-checklist</link><guid isPermaLink="false">https://dataanalysis.substack.com/p/ab-test-checklist</guid><dc:creator><![CDATA[Olga Berezovsky]]></dc:creator><pubDate>Wed, 28 Dec 2022 06:35:42 GMT</pubDate><enclosure url="https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/cd7029b3-f274-4215-ac43-d275f496ecf8_200x200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!fcCQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F7336868a-9de5-439b-9905-0fa313943267_200x200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fcCQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F7336868a-9de5-439b-9905-0fa313943267_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!fcCQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F7336868a-9de5-439b-9905-0fa313943267_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!fcCQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F7336868a-9de5-439b-9905-0fa313943267_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!fcCQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F7336868a-9de5-439b-9905-0fa313943267_200x200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fcCQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F7336868a-9de5-439b-9905-0fa313943267_200x200.png" width="146" height="146" data-attrs="{&quot;src&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/7336868a-9de5-439b-9905-0fa313943267_200x200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:200,&quot;width&quot;:200,&quot;resizeWidth&quot;:146,&quot;bytes&quot;:2197,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fcCQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F7336868a-9de5-439b-9905-0fa313943267_200x200.png 424w, https://substackcdn.com/image/fetch/$s_!fcCQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F7336868a-9de5-439b-9905-0fa313943267_200x200.png 848w, https://substackcdn.com/image/fetch/$s_!fcCQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F7336868a-9de5-439b-9905-0fa313943267_200x200.png 1272w, https://substackcdn.com/image/fetch/$s_!fcCQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F7336868a-9de5-439b-9905-0fa313943267_200x200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><h3><strong>&#9997;&#65039; Steps for conducting a product experiment:&nbsp;</strong></h3><ol><li><p><strong>Can you test it?&nbsp;</strong></p></li></ol><p>You can&#8217;t A/B test every little thing. New experiences or new product releases can&#8217;t be run through the A/B test (read - <a href="https://dataanalysis.substack.com/p/how-to-measure-product-adoption-issue-39c">&nbsp;How To Measure Product Adoption</a>). Potential bias - <a href="https://productds.com/wp-content/uploads/Novelty_Effect.html">novelty effect</a> or change aversion.&nbsp;&nbsp;</p><ol start="2"><li><p><strong>Formulate A Hypothesis</strong></p></li></ol><p>Why do you have to run the experiment? What is the ROI? Is it a good time to run the test? Consider seasonality, new version releases, open bugs, etc.&nbsp;</p><p>Set the rate you expect - this is your <a href="https://splitmetrics.com/resources/minimum-detectable-effect-mde/">Minimum Detectable Effect</a> (MDE).  This is the smallest acceptable difference between the Control and Variant. If the Variant is 0.0001% better than the Control, would you still want to run the test? Is it worth the cost and time?</p><ol start="3"><li><p><strong>Finalize your set of metrics</strong></p></li></ol><p>For A/B analysis, I use a set of 3 metrics:</p><ul><li><p>Success metrics</p></li><li><p>Ecosystem metrics (company KPIs)</p></li><li><p>Tradeoff metrics</p></li></ul><p>More described here - <a href="https://dataanalysis.substack.com/p/how-to-pick-the-right-metric-issue">How To Pick The Right Metric</a>.</p><ol start="4"><li><p><strong>Define your audience:</strong></p></li></ol><p>Is the change relevant for new or active users, or all? The region, platform, and language? If you work with matured analytics, do you need to limit the test exposure to some type of persona?&nbsp;</p><p>The more user attributes and filters you add, the longer the test will likely run, as it reduces your sample size. That being said, it reduces variance as well, and thus, it will be more precise.&nbsp;&nbsp;</p><ol start="5"><li><p><strong>Calculate</strong> <strong>sample size:</strong></p></li></ol><ul><li><p>Set your significance, confidence interval, and power.</p></li><li><p>Your group experiment sizes should be the same.</p></li><li><p>Your sample should be randomly distributed. Recognize traffic, devices, returning users, etc. Work with the engineering team on testing and ensuring that the randomization algorithm works as expected (hashing, clustering, sample stratification?).&nbsp;</p></li><li><p>Make sure there is no bias introduced with other tests running.&nbsp;</p></li></ul><ol start="6"><li><p><strong>Run the test</strong> until you reach significance and a little longer. Monitor the test timeline and events.</p></li><li><p><strong>Evaluate</strong> <strong>results</strong>:</p></li></ol><ul><li><p>Run sanity checks. Control metrics and conversions should match the Baseline. If they don&#8217;t, question the test setup.&nbsp;</p></li><li><p>Check sample variance and distribution. High variance often leads to low trust.</p></li><li><p>Run spot checks. Pick a few users from Control and Variant samples and check them to ensure they are random, not overlapping with other tests, and are meeting the test requirements.&nbsp;</p></li><li><p>If the result is not what you expected, think of the potential bias - novelty effect, learning effect, network effect.&nbsp;</p></li></ul><ol start="8"><li><p><strong>Draw conclusions</strong> and provide a recommendation on the next steps to product owners.&nbsp;</p></li></ol><h3><strong>&nbsp;&#128293; Things to remember:&nbsp;&nbsp;</strong></h3><ul><li><p>Run <a href="https://vwo.com/blog/aa-test-before-ab-testing/">the A/A test</a> first. It helps you check the software, outside factors, and natural variance. You would need to know the sample variance to estimate the significance level and statistical power.&nbsp;</p></li><li><p>Don&#8217;t pick metrics that are either too sensitive (views) or too robust (Day 7 or Day 30 retention). They are not helpful and tend to mislead you. The best test metric would show a change in the result and would not fluctuate much when other events are occurring.</p></li><li><p>Don't run the experiment for too long, as you might experience data pollution - the effect when multiple devices, cookies, and other outside factors affect your result.&nbsp;</p></li><li><p>Don&#8217;t run the experiment for too little time either, as you might get a false positive (<a href="https://dataanalysis.substack.com/p/these-tricky-ab-tests-or-why-significance">regression to the mean</a>). In other words, when a variable is extreme at first but then moves closer to the average.</p></li><li><p>When introducing a new change, run the test on a smaller sample for a longer period of time to eliminate the novelty or learning effect bias.</p></li></ul><h3><strong>&#128163; Statistical terminology</strong></h3><p>To approach A/B testing, you can think of <strong><a href="https://www.invespcro.com/blog/calculating-sample-size-for-an-ab-test/">Null-Hypothesis</a></strong><a href="https://www.invespcro.com/blog/calculating-sample-size-for-an-ab-test/"> </a><strong><a href="https://www.invespcro.com/blog/calculating-sample-size-for-an-ab-test/">testing</a></strong> and apply the following terms:</p><ul><li><p><a href="https://www.cs.purdue.edu/homes/ribeirob/courses/Fall2016/lectures/hyp_tests.pdf">P-value</a> - assuming Null-H is true, what is the probability of seeing a specific result? If data is in the "not expected" region, we reject Null-H.&nbsp;</p></li><li><p><a href="https://hbr.org/2016/02/a-refresher-on-statistical-significance">Statistical Significance</a> (or Significance level, alpha) is the probability of seeing the effect when none exists (false positive).</p></li><li><p><a href="https://productcoalition.com/start-here-statistics-for-a-b-testing-5f5c7e02ce1e">Statistical Power</a> (or 1-beta) is the probability of seeing the effect when it does exist.</p></li><li><p><a href="https://cxl.com/blog/confidence-intervals/">Confidence Interval</a> is the number of allowed errors or measurements of estimated reliability: the smaller CI, the more accurate the result.</p></li><li><p><a href="https://evolytics.com/resources/calculators/abtesting-statistical-significance/">z-score</a> is the number of Standard Deviations from the mean.</p></li></ul><h3><strong>&#129300; If you are lost in conversions and numbers, check this guide:&nbsp;</strong></h3><ol><li><p>If your baseline conversion is 20%, you may set the MDE to 10%, and the test may detect 18% - 22% conversion results.</p></li><li><p>The higher your baseline conversion, the smaller the sample size you&#8217;ll need.</p></li><li><p>The smaller the MDE, the larger sample you&#8217;ll need.</p></li><li><p>Low p values are good. They indicate the result didn&#8217;t occur by chance.</p></li><li><p>Your confidence level could be 95% and statistical power - 80%</p></li><li><p>It&#8217;s often recommended to run the experiment for 2 business cycles (2-4 weeks)&nbsp;&nbsp;</p></li></ol><p>&#128226; Use <a href="https://www.optimizely.com/sample-size-calculator/">this calculator</a> or <a href="https://www.evanmiller.org/ab-testing/sample-size.html">this one</a> to determine the needed sample size for your experiment.&nbsp;</p><p>&#128226; Use <a href="http://www.abtestcalculator.com/">this calculator</a> to evaluate your test significance and result.&nbsp;</p><h3><strong>&#128269; Other types of product testing</strong></h3><ol><li><p><a href="https://vwo.com/blog/difference-ab-testing-multivariate-testing/">Multivariate testing</a> (MVT) - multiple variants and their combinations within a single test.</p></li><li><p><a href="https://help.vwo.com/hc/en-us/articles/360020072833-What-is-Split-URL-Testing-">Split URL testing</a> - multiple versions of your webpage posted on different URLs.</p></li><li><p><a href="https://blog.optimizely.com/2013/06/12/4-ways-to-maximize-conversions-with-multi-page-testing/">Multipage testing</a> -&nbsp; testing changes across different pages. There is both funnel Multi-Page testing and Conventional Multi-Page testing. Read more <a href="https://vwo.com/ab-testing/">here</a>.</p></li></ol><p>Check out <a href="https://blog.hubspot.com/marketing/how-to-do-a-b-testing">this guide</a> if you want an A/B experiment checklist.</p>]]></content:encoded></item></channel></rss>