<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Causal Inference on Big Muddy</title><link>https://muddy.jprs.me/tags/causal-inference/</link><description>Recent content in Causal Inference on Big Muddy</description><generator>Hugo</generator><language>en-US</language><lastBuildDate>Mon, 16 Feb 2026 22:49:00 -0500</lastBuildDate><atom:link href="https://muddy.jprs.me/tags/causal-inference/index.xml" rel="self" type="application/rss+xml"/><item><title>In the multiverse of forking paths</title><link>https://muddy.jprs.me/links/2026-02-16-in-the-multiverse-of-forking-paths/</link><pubDate>Mon, 16 Feb 2026 22:49:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-16-in-the-multiverse-of-forking-paths/</guid><description>&lt;p&gt;&lt;img src="https://muddy.jprs.me/media/stark-strange-one.jpg" alt="A scene from Avengers: Infinity War, where Tony Stark asks Dr. Strange, “How many did we win?” Strange replies, “One.”"&gt;&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;STRANGE: I went forward in time to view alternate modelling decisions, to see all the possible outcomes of the coming analysis.&lt;br&gt;
STAR-LORD: How many did you see?&lt;br&gt;
STRANGE: 14,000,605.&lt;br&gt;
STARK: How many did we achieve statistical significance?&lt;br&gt;
STRANGE: One.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Prof. Jessica Hullman recently wrote a piece on Andrew Gelman&amp;rsquo;s blog discussing the use of &amp;lsquo;multiverse analysis&amp;rsquo;, i.e., what if we could see the results of the many slightly different decisions we could have made when constructing a model. This problem is commonly known as the &lt;a href="https://en.wikipedia.org/wiki/Forking_paths_problem"&gt;garden of forking paths&lt;/a&gt;—during an analysis, a researcher is forced to make many small, sometimes arbitrary decisions that can lead to a different result if another researcher tries to independently replicate the analysis. While usually an innocent and inevitable part of the modelling process, these &amp;lsquo;researcher degrees of freedom&amp;rsquo; can also be manipulated to produce a desired result.&lt;/p&gt;
&lt;p&gt;Prof. Hullman points out that multiverse analysis will only become salient as AI coding tools such as Claude Code make it easier than ever to iterate on how we model our research questions.&lt;/p&gt;
&lt;p&gt;Her longer paper with Julia M. Rohrer and Andrew Gelman, &amp;ldquo;What&amp;rsquo;s a multiverse good for anyway?&amp;rdquo; is available &lt;a href="https://osf.io/preprints/psyarxiv/37g29_v1"&gt;here&lt;/a&gt;.&lt;/p&gt;</description></item><item><title>A/B testing for advertising is not randomized</title><link>https://muddy.jprs.me/links/2026-02-01-a-b-testing-for-advertising-is-not-randomized/</link><pubDate>Sun, 01 Feb 2026 23:09:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-01-a-b-testing-for-advertising-is-not-randomized/</guid><description>&lt;p&gt;Florian Teschner writes about a &lt;a href="https://www.sciencedirect.com/science/article/pii/S0167811624001149"&gt;recent paper&lt;/a&gt; from Bögershausen, Oertzen, &amp;amp; Bock arguing that online ad platforms like Facebook and Google misrepresent the meaning of &amp;ldquo;A/B testing&amp;rdquo; for ad campaigns. In A/B testing, we might assume the platform is randomly assigning users to see ad A or ad B, in an attempt to get a clean causal interpretation about which ad is more likely to drive a click (or whatever outcome you’re tracking).&lt;/p&gt;
&lt;p&gt;But according to the paper, this is usually not what is happening. Instead, the platform optimizes delivery for each ad independently, &lt;em&gt;steering each one toward the users most likely to click it&lt;/em&gt;. In other words, the two ads may be shown to different groups of users, and differences in click-through rates may be attributable to who is seeing the ad, as opposed to the overall appeal of the ad. Ad platforms convert A/B tests from simple randomized experiments into murky observational comparisons. For example, an ad may appear to do better because it happened to be shown disproportionately to a group with a high click-through rate, not because it presents a more compelling overall message. Advertisers get the warm glow of &amp;ldquo;experimentally backed&amp;rdquo; marketing without the assurances of randomization.&lt;/p&gt;</description></item></channel></rss>