<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Statistics on Big Muddy</title><link>https://muddy.jprs.me/tags/statistics/</link><description>Recent content in Statistics on Big Muddy</description><generator>Hugo</generator><language>en-US</language><lastBuildDate>Sun, 12 Apr 2026 22:55:00 -0400</lastBuildDate><atom:link href="https://muddy.jprs.me/tags/statistics/index.xml" rel="self" type="application/rss+xml"/><item><title>Adjusting for recalled past vote in political polling</title><link>https://muddy.jprs.me/links/2026-04-12-adjusting-for-recalled-past-vote-in-political-polling/</link><pubDate>Sun, 12 Apr 2026 22:55:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-04-12-adjusting-for-recalled-past-vote-in-political-polling/</guid><description>&lt;p&gt;The founder of Abacus Data, a Canadian polling firm, &lt;a href="https://x.com/DavidColetto/status/2043026334046392624"&gt;dropped&lt;/a&gt; kind of an interesting URL yesterday: &lt;a href="https://abacus-weighting.com/"&gt;abacus-weighting.com&lt;/a&gt;. It&amp;rsquo;s a advertisement in the form of a case study on why Abacus weights their political polls on past vote. It fits perfectly with the theme of &lt;a href="https://muddy.jprs.me/links/2026-04-11-how-do-pollsters-get-different-results-from-the-same-data/"&gt;yesterday&amp;rsquo;s post&lt;/a&gt; on how pollster&amp;rsquo;s get different results from the same data (the answer is they weight the raw data differently).&lt;/p&gt;
&lt;p&gt;If you follow Nate Silver (or American political polling in general), you probably know that pollsters &lt;a href="https://www.nbcnews.com/politics/2024-election/-polls-missed-decisive-slice-trump-voters-2024-rcna182488"&gt;undercounted Trump support&lt;/a&gt; in all three elections where he was on the ballot. What I learned from this post is that support for the Conservative Party of Canada has been underestimated in their firm&amp;rsquo;s polling data in every polling wave for every election since 2011:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;In &lt;strong&gt;every single wave&lt;/strong&gt;, across every single election cycle, Conservative voters are underrepresented in our demographically weighted sample relative to their actual share of the vote. Not in most waves. Not in some elections. In every case we can observe.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Weighting for recalled past vote improves the estimate in every case, sometimes dramatically so:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;In every election, past vote weighting moved our Conservative estimates upward and our Liberal estimates downward — consistently in the direction of the actual result. The 2021 election shows the most dramatic correction: a 7-point improvement in our Conservative estimate.&lt;/p&gt;</description></item><item><title>How do pollsters get different results from the same data?</title><link>https://muddy.jprs.me/links/2026-04-11-how-do-pollsters-get-different-results-from-the-same-data/</link><pubDate>Sat, 11 Apr 2026 22:36:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-04-11-how-do-pollsters-get-different-results-from-the-same-data/</guid><description>&lt;p&gt;Nate Silver linked to this throwback article from 2016 in &lt;em&gt;The New York Times&lt;/em&gt; in his recent article on &lt;a href="https://www.natesilver.net/p/ai-polls-are-fake-polls"&gt;fake AI polls&lt;/a&gt;, which &lt;a href="https://muddy.jprs.me/links/2026-04-07-what-is-a-public-opinion-poll-without-the-public/"&gt;I also wrote about a few days ago&lt;/a&gt;. The article, entitled &amp;ldquo;We Gave Four Good Pollsters the Same Raw Data. They Had Four Different Results.&amp;rdquo; is a good reminder that modern polling diverges very far from the theoretical ideal of a &lt;a href="https://en.wikipedia.org/wiki/Simple_random_sample"&gt;simple random sample&lt;/a&gt;. Even after deciding on a methodology to sample participants and collecting the data, a lot of work goes into interpreting raw poll responses to give us top-line polling numbers. Every pollster needs to figure out how to weight the responses they get, since poll response rates are abysmal and variable across different demographic groups. As in the example given in this piece, these choices can result in large differences in those top-line numbers: from +4 Clinton to +1 Trump, all from the same raw data!&lt;/p&gt;
&lt;p&gt;For an interesting follow-up: &amp;ldquo;&lt;a href="https://www.natesilver.net/p/polling-is-becoming-more-of-an-art"&gt;Polling is becoming more of an art than a science&lt;/a&gt;&amp;rdquo;, also on Nate Silver&amp;rsquo;s Substack.&lt;/p&gt;</description></item><item><title>AI makes it easier to generate fake papers, too</title><link>https://muddy.jprs.me/links/2026-04-08-ai-makes-it-easier-to-generate-fake-papers-too/</link><pubDate>Wed, 08 Apr 2026 20:09:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-04-08-ai-makes-it-easier-to-generate-fake-papers-too/</guid><description>&lt;p&gt;Here&amp;rsquo;s a fun project from Tyler Vigen, creator of the famous &lt;a href="https://tylervigen.com/spurious-correlations"&gt;Spurious Correlations&lt;/a&gt; page (which has been cited as a cautionary tale in many a science class). Using his database of real but spurious correlations (created by calculating the Pearson correlation coefficient &lt;em&gt;r&lt;/em&gt; between a very large number of variables and picking out the hits), he used AI to create amusing fake manuscripts expounding on these statistical flukes as if they were real research questions.&lt;/p&gt;
&lt;p&gt;These papers were generated in January 2024, and as &lt;a href="https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/"&gt;previously discussed&lt;/a&gt; on this blog, the pipeline for end-to-end paper generation has come a long way in two years. I have no doubt Tyler could make these paper&amp;rsquo;s sound much more convincing using today&amp;rsquo;s models, though of course his goal here is to make you laugh (and think), not to trick you. But I have no doubt there will be many scholars adopting this data dredging strategy to generate &amp;ldquo;real&amp;rdquo; papers, contributing to a deluge of papers &lt;a href="https://muddy.jprs.me/links/2026-03-03-the-productivity-shock-coming-to-academic-publishing/"&gt;flooding &lt;/a&gt;&lt;a href="https://muddy.jprs.me/links/2026-03-03-the-productivity-shock-coming-to-academic-publishing/"&gt;the academic publishing system&lt;/a&gt;.&lt;/p&gt;</description></item><item><title>Using Claude Claude for cross-package statistical audits</title><link>https://muddy.jprs.me/links/2026-03-15-using-claude-claude-for-cross-package-statistical-audits/</link><pubDate>Sun, 15 Mar 2026 22:49:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-15-using-claude-claude-for-cross-package-statistical-audits/</guid><description>&lt;p&gt;Economist Scott Cunningham shared an important example of why we should always report the statistical package and version used in our analyses, as he used Claude Code to produce six versions of the exact same analysis using six different packages in R, Python, and Stata. In a &lt;a href="https://en.wikipedia.org/wiki/Difference_in_differences"&gt;difference-in-differences&lt;/a&gt; analysis of the mental health hospital closures on homicide using the standard &lt;a href="https://bcallaway11.github.io/did/articles/multi-period-did.html"&gt;Callaway and Sant’Anna estimator&lt;/a&gt; (for DiD with multiple time periods), he got very different results for some model specifications.&lt;/p&gt;
&lt;p&gt;Since the specifications and the data were identical between packages, he discovered the divergences occurred due to how the packages handled problems with &lt;a href="https://www.tandfonline.com/doi/full/10.1080/00273171.2011.568786#d1e368"&gt;propensity score&lt;/a&gt; weights. Packages were not necessarily transparent about issues with these weights. If you were not running multiple analyses and comparing results across packages, or else carefully checking propensity score diagnostics, you might never have realized how precarious your results were.&lt;/p&gt;
&lt;p&gt;Prof. Cunningham closes with the following advice:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The fifth point, and the broader point, is that this kind of cross-package, cross-language audit is exactly what Claude Code should be used for. Why? Because this is a task that is time-intensive, high-value, and brutally easy to get wrong. But just one mismatched diagnostic across languages invalidates the entire comparison, even something as simple as sample size values differing across specifications, would flag it. This is both easy and not easy — but it is not the work humans should be doing by hand given how easy it would be to even get that much wrong.&lt;/p&gt;</description></item><item><title>In the multiverse of forking paths</title><link>https://muddy.jprs.me/links/2026-02-16-in-the-multiverse-of-forking-paths/</link><pubDate>Mon, 16 Feb 2026 22:49:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-16-in-the-multiverse-of-forking-paths/</guid><description>&lt;p&gt;&lt;img src="https://muddy.jprs.me/media/stark-strange-one.jpg" alt="A scene from Avengers: Infinity War, where Tony Stark asks Dr. Strange, “How many did we win?” Strange replies, “One.”"&gt;&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;STRANGE: I went forward in time to view alternate modelling decisions, to see all the possible outcomes of the coming analysis.&lt;br&gt;
STAR-LORD: How many did you see?&lt;br&gt;
STRANGE: 14,000,605.&lt;br&gt;
STARK: How many did we achieve statistical significance?&lt;br&gt;
STRANGE: One.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Prof. Jessica Hullman recently wrote a piece on Andrew Gelman&amp;rsquo;s blog discussing the use of &amp;lsquo;multiverse analysis&amp;rsquo;, i.e., what if we could see the results of the many slightly different decisions we could have made when constructing a model. This problem is commonly known as the &lt;a href="https://en.wikipedia.org/wiki/Forking_paths_problem"&gt;garden of forking paths&lt;/a&gt;—during an analysis, a researcher is forced to make many small, sometimes arbitrary decisions that can lead to a different result if another researcher tries to independently replicate the analysis. While usually an innocent and inevitable part of the modelling process, these &amp;lsquo;researcher degrees of freedom&amp;rsquo; can also be manipulated to produce a desired result.&lt;/p&gt;
&lt;p&gt;Prof. Hullman points out that multiverse analysis will only become salient as AI coding tools such as Claude Code make it easier than ever to iterate on how we model our research questions.&lt;/p&gt;
&lt;p&gt;Her longer paper with Julia M. Rohrer and Andrew Gelman, &amp;ldquo;What&amp;rsquo;s a multiverse good for anyway?&amp;rdquo; is available &lt;a href="https://osf.io/preprints/psyarxiv/37g29_v1"&gt;here&lt;/a&gt;.&lt;/p&gt;</description></item><item><title>There is only one statistical test</title><link>https://muddy.jprs.me/links/2026-02-11-there-is-only-one-statistical-test/</link><pubDate>Wed, 11 Feb 2026 23:58:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-11-there-is-only-one-statistical-test/</guid><description>&lt;p&gt;A classic article by computer scientist Allen Downey on why there is only one statistical test: compute a test statistic from your observed data, simulate a null hypothesis, and finally compute/approximate a p-value by calculating the fraction of test statistics from the simulated data exceeding the test statistic from your observed data.&lt;/p&gt;
&lt;p&gt;&lt;img src="https://muddy.jprs.me/media/one-statistical-test.png" alt="Diagram illustrating a single hypothesis-testing workflow: observed data are converted into a test statistic (effect δ∗); a null model H0 generates many simulated datasets to form the distribution of δ under H0; the p-value is the tail area of that distribution beyond δ∗."&gt;&lt;/p&gt;
&lt;p&gt;Downey suggests using general simulation methods over the canon of rigid, inflexible tests invented when computation was difficult and expensive.&lt;/p&gt;
&lt;p&gt;&lt;sup&gt;Hat tip to Ryan Briggs on &lt;a href="https://x.com/ryancbriggs/status/2021614352328003642"&gt;Twitter&lt;/a&gt;.&lt;/sup&gt;&lt;/p&gt;</description></item><item><title>Anthropic's statistical analysis skill doesn't get statistical significance quite right</title><link>https://muddy.jprs.me/links/2026-02-06-anthropic-s-statistical-analysis-skill-doesn-t-get-statistical-significance-quite-right/</link><pubDate>Fri, 06 Feb 2026 19:30:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-06-anthropic-s-statistical-analysis-skill-doesn-t-get-statistical-significance-quite-right/</guid><description>&lt;p&gt;Anthropic&amp;rsquo;s new statistical analysis skill demonstrates a common misunderstanding of statistical significance:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Statistical significance means the difference is unlikely due to chance.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;But this phrasing isn&amp;rsquo;t quite right. The p-value in Null Hypothesis Significance Testing is not about the probability the results are &amp;ldquo;due to chance&amp;rdquo;; it is the probability—under the null hypothesis and the model assumptions—of observing results at least as extreme as the ones we obtained. In other words, the p-value summarizes how compatible the data are with the null, given our modelling choices. What it does not tell you is the probability that the null hypothesis is true.&lt;/p&gt;
&lt;p&gt;Statistician Andrew Gelman gave a good definition for statistical significance in a 2015 &lt;a href="https://statmodeling.stat.columbia.edu/2015/07/21/a-bad-definition-of-statistical-significance-from-the-u-s-department-of-health-and-human-services-effective-health-care-program/"&gt;blog post&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;A mathematical technique to measure the strength of evidence from a single study. Statistical significance is conventionally declared when the p-value is less than 0.05. The p-value is the probability of seeing a result as strong as observed or greater, under the &lt;em&gt;null hypothesis&lt;/em&gt; (which is commonly the hypothesis that there is no effect). Thus, the smaller the p-value, the less consistent are the data with the null hypothesis under this measure.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;As some of the commenters in this blog post observe, simply being able to parrot a technically accurate definition of a p-value does not necessarily make us better at applying statistical significance in practice. It is certainly true that statistical significance is widely misused in scientific publishing as a threshold to distinguish signal from noise (or to be fancy, a &amp;ldquo;lexicographic decision rule&amp;rdquo;), which is why &lt;a href="https://sites.stat.columbia.edu/gelman/research/published/abandon.pdf"&gt;some scientists have argued that we should abandon it as the default statistical paradigm for research&lt;/a&gt;.&lt;/p&gt;</description></item><item><title>A/B testing for advertising is not randomized</title><link>https://muddy.jprs.me/links/2026-02-01-a-b-testing-for-advertising-is-not-randomized/</link><pubDate>Sun, 01 Feb 2026 23:09:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-01-a-b-testing-for-advertising-is-not-randomized/</guid><description>&lt;p&gt;Florian Teschner writes about a &lt;a href="https://www.sciencedirect.com/science/article/pii/S0167811624001149"&gt;recent paper&lt;/a&gt; from Bögershausen, Oertzen, &amp;amp; Bock arguing that online ad platforms like Facebook and Google misrepresent the meaning of &amp;ldquo;A/B testing&amp;rdquo; for ad campaigns. In A/B testing, we might assume the platform is randomly assigning users to see ad A or ad B, in an attempt to get a clean causal interpretation about which ad is more likely to drive a click (or whatever outcome you’re tracking).&lt;/p&gt;
&lt;p&gt;But according to the paper, this is usually not what is happening. Instead, the platform optimizes delivery for each ad independently, &lt;em&gt;steering each one toward the users most likely to click it&lt;/em&gt;. In other words, the two ads may be shown to different groups of users, and differences in click-through rates may be attributable to who is seeing the ad, as opposed to the overall appeal of the ad. Ad platforms convert A/B tests from simple randomized experiments into murky observational comparisons. For example, an ad may appear to do better because it happened to be shown disproportionately to a group with a high click-through rate, not because it presents a more compelling overall message. Advertisers get the warm glow of &amp;ldquo;experimentally backed&amp;rdquo; marketing without the assurances of randomization.&lt;/p&gt;</description></item><item><title>Twyman's law</title><link>https://muddy.jprs.me/links/2026-01-30-twyman-s-law/</link><pubDate>Fri, 30 Jan 2026 19:25:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-01-30-twyman-s-law/</guid><description>&lt;p&gt;From Wikipedia:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Twyman&amp;rsquo;s law&lt;/strong&gt; states that &amp;ldquo;Any figure that looks interesting or different is usually wrong&amp;rdquo;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;A bit different from that &lt;a href="https://quoteinvestigator.com/2015/03/02/eureka-funny/"&gt;oft-quoted line attributed to Isaac Asimov&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The most exciting phrase in science is not ‘Eureka!’ but ‘that&amp;rsquo;s funny’&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;But Twyman&amp;rsquo;s law is much truer in my experience. Surprising results are usually a signal that something is screwy with my data, my assumptions, or my pipeline.&lt;/p&gt;
&lt;p&gt;&lt;sub&gt;Hat tip to DJ Rich on &lt;a href="https://x.com/DuaneJRich/status/2016919349135888562"&gt;Twitter&lt;/a&gt;.&lt;/sub&gt;&lt;/p&gt;</description></item></channel></rss>