<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Ai on Big Muddy</title><link>https://muddy.jprs.me/tags/ai/</link><description>Recent content in Ai on Big Muddy</description><generator>Hugo</generator><language>en-US</language><lastBuildDate>Fri, 10 Apr 2026 18:27:00 -0400</lastBuildDate><atom:link href="https://muddy.jprs.me/tags/ai/index.xml" rel="self" type="application/rss+xml"/><item><title>Scientists invent a fake disease, AI picks it up, other scientists cite it</title><link>https://muddy.jprs.me/links/2026-04-10-scientists-invent-a-fake-disease-ai-picks-it-up-other-scientists-cite-it/</link><pubDate>Fri, 10 Apr 2026 18:27:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-04-10-scientists-invent-a-fake-disease-ai-picks-it-up-other-scientists-cite-it/</guid><description>&lt;p&gt;A somewhat disturbing bit of reporting from &lt;em&gt;Nature&lt;/em&gt; tells the story of bixonimania, a fake eye disease invented by Swedish medical researcher Almira Osmanovic Thunström and her team. She seeded the idea for the fake disease in a series of ridiculous, joke-filled blog posts and preprints in mid-2024.&lt;/p&gt;
&lt;p&gt;Because AI can be overly credulous with its sourcing (how often do Google&amp;rsquo;s AI answers confident cite random Reddit posts for the bulk of an answer?), the disease got picked up as an &amp;ldquo;emerging term&amp;rdquo; by the leading chatbots. The preprints even got cited a handful of times in real publications, which is further evidence that scientists don&amp;rsquo;t read the papers they cite (I guess the modern equivalent of copying citations from other papers is having AI dredge the literature for you).&lt;/p&gt;
&lt;p&gt;I can see AI agents being exploited by those pushing dubious medical diagnoses to flood the Internet and preprint servers with articles aimed at convincing LLMs of the validity of their positions. That is if the agents aren&amp;rsquo;t too busy &lt;a href="https://theshamblog.com/an-ai-agent-published-a-hit-piece-on-me/"&gt;spinning of websites to defame&lt;/a&gt; those who incur their wrath.&lt;/p&gt;</description></item><item><title>A data point against the idea that AI will freeze/homogenize culture</title><link>https://muddy.jprs.me/links/2026-04-09-a-data-point-against-the-idea-that-ai-will-freeze-homogenize-culture/</link><pubDate>Thu, 09 Apr 2026 07:00:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-04-09-a-data-point-against-the-idea-that-ai-will-freeze-homogenize-culture/</guid><description>&lt;p&gt;Here&amp;rsquo;s an interesting figure and accompanying passage from this 2023 preprint entitled &amp;ldquo;Machine Culture&amp;rdquo;:&lt;/p&gt;
&lt;p&gt;&lt;img src="https://muddy.jprs.me/media/go-performance.png" alt="Chart of professional Go players’ decision quality over time. Performance is mostly flat from 1950 to 2015, then increases steeply after AlphaGo’s 2016 match with Lee Sedol, highlighted by a shaded region on the right."&gt;&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The innovations generated by AlphaGo and AlphaGo Zero soon entered human culture, as shown by research comparing human gameplay before and after the algorithms’ introduction. The decision quality, as measured by an open-source variant of AlphaGo Zero, showed very little improvement in human gameplay from 1950 to 2016, followed by a sudden improvement after the introduction of AlphaGo in March 2016. However, this improvement wasn’t solely due to humans adopting strategies developed by AlphaGo. It also reflected an unexpected shift, wherein humans started developing moves that were qualitatively distinct both from previous human moves and from the novel moves introduced by AlphaGo. In summary, AlphaGo served as an early, quantifiable exemplar of machine culture, generating novel cultural variations through genuine, nonhuman innovation. This was followed by a major transition into an even broader range of traits as the result of humans building on the previous discoveries made by machines. As the methods underpinning AlphaGo have been generalized to other games and extended to scientific problems, we anticipate a continued infusion of machine-generated discoveries across diverse domains of human culture.&lt;/p&gt;</description></item><item><title>AI makes it easier to generate fake papers, too</title><link>https://muddy.jprs.me/links/2026-04-08-ai-makes-it-easier-to-generate-fake-papers-too/</link><pubDate>Wed, 08 Apr 2026 20:09:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-04-08-ai-makes-it-easier-to-generate-fake-papers-too/</guid><description>&lt;p&gt;Here&amp;rsquo;s a fun project from Tyler Vigen, creator of the famous &lt;a href="https://tylervigen.com/spurious-correlations"&gt;Spurious Correlations&lt;/a&gt; page (which has been cited as a cautionary tale in many a science class). Using his database of real but spurious correlations (created by calculating the Pearson correlation coefficient &lt;em&gt;r&lt;/em&gt; between a very large number of variables and picking out the hits), he used AI to create amusing fake manuscripts expounding on these statistical flukes as if they were real research questions.&lt;/p&gt;
&lt;p&gt;These papers were generated in January 2024, and as &lt;a href="https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/"&gt;previously discussed&lt;/a&gt; on this blog, the pipeline for end-to-end paper generation has come a long way in two years. I have no doubt Tyler could make these paper&amp;rsquo;s sound much more convincing using today&amp;rsquo;s models, though of course his goal here is to make you laugh (and think), not to trick you. But I have no doubt there will be many scholars adopting this data dredging strategy to generate &amp;ldquo;real&amp;rdquo; papers, contributing to a deluge of papers &lt;a href="https://muddy.jprs.me/links/2026-03-03-the-productivity-shock-coming-to-academic-publishing/"&gt;flooding &lt;/a&gt;&lt;a href="https://muddy.jprs.me/links/2026-03-03-the-productivity-shock-coming-to-academic-publishing/"&gt;the academic publishing system&lt;/a&gt;.&lt;/p&gt;</description></item><item><title>What is a public opinion poll without the public?</title><link>https://muddy.jprs.me/links/2026-04-07-what-is-a-public-opinion-poll-without-the-public/</link><pubDate>Tue, 07 Apr 2026 18:27:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-04-07-what-is-a-public-opinion-poll-without-the-public/</guid><description>&lt;p&gt;A few days ago, two professors (Leif Weatherby and Benjamin Recht) published an opinion piece in the &lt;em&gt;New York Times&lt;/em&gt; calling attention to Axios publishing a &lt;a href="https://www.axios.com/2026/03/19/olivia-walton-heartland-forward-maternal-health"&gt;story&lt;/a&gt; on maternal health using invented polling results:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;A recent Axios story on maternal health policy referred to “findings” that a majority of people trusted their doctors and nurses. On the surface, there’s nothing unusual about that. What wasn’t originally mentioned, however, was that these findings were made up.&lt;/p&gt;
&lt;p&gt;Clicking through the links revealed (as did a subsequent editor’s note and clarification by Axios) that the public opinion poll was a computer simulation run by the artificial intelligence start-up Aaru. No people were involved in the creation of these opinions.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The piece goes on to argue that this so-called &amp;ldquo;silicon sampling&amp;rdquo; is seductive because good public opinion polling is expensive, hard to do, and still prone to bias. But this shortcut magnifies the the problem of bias rather than solving it.&lt;/p&gt;
&lt;p&gt;I&amp;rsquo;ve read a little bit about this strategy of using LLM-generated survey participants in the context of social science research in a series of posts (mostly from Prof. Jessica Hullman) over on &lt;a href="https://statmodeling.stat.columbia.edu/"&gt;Andrew Gelman&amp;rsquo;s blog&lt;/a&gt;:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href="https://statmodeling.stat.columbia.edu/2025/12/19/validating-language-models-as-study-participants-how-its-being-done-why-it-fails-and-what-works-instead/"&gt;Validating language models as study participants: How it’s being done, why it fails, and what works instead&lt;/a&gt; (2025-12-19)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://statmodeling.stat.columbia.edu/2025/08/26/survey-statistics-thomas-lumley-writes-about-interviewing-your-laptop/"&gt;Survey Statistics: Thomas Lumley writes about Interviewing your Laptop&lt;/a&gt; (2025-08-26)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://statmodeling.stat.columbia.edu/2025/08/15/when-does-it-make-sense-to-talk-about-llms-having-beliefs/"&gt;When does it make sense to talk about LLMs having beliefs?&lt;/a&gt; (2025-08-15)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://statmodeling.stat.columbia.edu/2025/06/12/better-and-worse-ways-to-mix-human-and-llm-responses-in-behavioral-research-but-you-still-have-to-figure-what-youre-measuring/"&gt;Better and worse ways to mix human and LLM responses in behavioral research (but you still have to figure what you’re measuring)&lt;/a&gt; (2025-06-12)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://statmodeling.stat.columbia.edu/2025/05/29/llms-as-behavioral-study-participants/"&gt;LLMs as behavioral study participants&lt;/a&gt; (2025-05-29)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Silicon sampling seems moderately interesting from a research perspective, but I can&amp;rsquo;t help but agree with the &lt;em&gt;New York Times&lt;/em&gt; opinion piece authors that this will be ruinous for the already waning trust in public opinion polling. If you didn&amp;rsquo;t bother to ask the public, then why should the public care what you &amp;ldquo;find&amp;rdquo;? I think there is probably a lot of utility in using LLM samples to aid in designing and validating surveys, though.&lt;/p&gt;</description></item><item><title>The definition of "agent"</title><link>https://muddy.jprs.me/notes/2026-04-04-the-definition-of-agent/</link><pubDate>Sat, 04 Apr 2026 23:59:00 -0400</pubDate><guid>https://muddy.jprs.me/notes/2026-04-04-the-definition-of-agent/</guid><description>&lt;p&gt;An interesting exchange between &lt;a href="https://x.com/gvanrossum/status/2039045160156426463"&gt;Guido van Rossum&lt;/a&gt; and &lt;a href="https://x.com/karpathy/status/2039054981719089202"&gt;Andrej Karpathy&lt;/a&gt; a few days ago on Twitter:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Guido van Rossum:&lt;/strong&gt;
I think I finally understand what an agent is. It&amp;rsquo;s a prompt (or several), skills, and tools. Did I get this right?&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Andrej Karpathy&lt;/strong&gt;:
LLM = CPU (data: tokens not bytes, dynamics: statistical and vague not deterministic and precise)
Agent = operating system kernel&lt;/p&gt;
&lt;/blockquote&gt;</description></item><item><title>The triumph of the data raccoons</title><link>https://muddy.jprs.me/notes/2026-04-03-the-triumph-of-the-data-raccoons/</link><pubDate>Fri, 03 Apr 2026 23:40:00 -0400</pubDate><guid>https://muddy.jprs.me/notes/2026-04-03-the-triumph-of-the-data-raccoons/</guid><description>&lt;p&gt;My PhD co-supervisor at the University of Toronto, Dr. David Fisman, liked to use the term &amp;ldquo;data raccoon&amp;rdquo; to describe the work of using messy, incomplete, hard-to-work-with data to do serious research. Or, as he described it in &lt;a href="https://www.ourcommons.ca/documentviewer/en/43-1/HESA/meeting-22/evidence#Int-10851517"&gt;testimony&lt;/a&gt; to the Canadian House of Commons in May 2020 (emphasis mine):&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;I&amp;rsquo;ll tell you, my group at University of Toronto call ourselves &amp;ldquo;data raccoons&amp;rdquo;, because we&amp;rsquo;ve sort of managed to thrive for about 15 years on &lt;strong&gt;data that most people regard as garbage&lt;/strong&gt;, so it&amp;rsquo;s sort of a bit of the normal state of affairs for us with public health data analysis.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;It&amp;rsquo;s an unmistakably Toronto metaphor—the city isn&amp;rsquo;t called the &lt;a href="https://www.npr.org/2018/09/16/647599627/theres-no-stopping-toronto-s-uber-raccoon"&gt;raccoon capital of the world&lt;/a&gt; for nothing!&lt;/p&gt;
&lt;p&gt;It occurred to me recently that data raccoons have basically taken over the world. The basis of the AI revolution is vast quantities of text dredged from the Internet, none of which was written for its final purpose of training the &lt;em&gt;deus ex machina&lt;/em&gt;. Arguably the most important dataset for training LLMs has been &lt;a href="https://en.wikipedia.org/wiki/Common_Crawl_Foundation"&gt;Common Crawl&lt;/a&gt;, a mostly uncurated snapshot of the Internet that has been running since 2007. According to a &lt;a href="https://www.mozillafoundation.org/en/research/library/generative-ai-training-data/common-crawl/"&gt;Mozilla report&lt;/a&gt; from 2024, Common Crawl was used in two thirds of LLMs developed in the formative period between 2019 and 2023, and the archive also comprised 80% of tokens in OpenAI&amp;rsquo;s GPT-3. Unsurprisingly, the Common Crawl Foundation has received &lt;a href="https://archive.is/NS9MI"&gt;financial support&lt;/a&gt; from AI companies in recent years, all the while being accused of abetting these same companies to train their models on paywalled articles.&lt;/p&gt;</description></item><item><title>Testing ZeroClaw, Part 2.5: ZeroClaw is alive!</title><link>https://muddy.jprs.me/notes/2026-04-01-testing-zeroclaw-part-2-5-zeroclaw-is-alive/</link><pubDate>Wed, 01 Apr 2026 23:59:00 -0400</pubDate><guid>https://muddy.jprs.me/notes/2026-04-01-testing-zeroclaw-part-2-5-zeroclaw-is-alive/</guid><description>&lt;p&gt;&lt;a href="https://muddy.jprs.me/notes/2026-03-31-testing-zeroclaw-part-2-zeroclaw-is-dead/"&gt;Yesterday&lt;/a&gt;, I wrote about how the &lt;a href="https://github.com/zeroclaw-labs/zeroclaw"&gt;ZeroClaw GitHub repository&lt;/a&gt; had been down for two days with little explanation. Earlier today, the project provided a little more information on &lt;a href="https://x.com/zeroclawlabs/status/2039419021494264055"&gt;Twitter&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;They flagged our org which is why we’re down. Code is safe and we’re still working, just waiting for @github&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Since March 30 (the day after their repo started 404ing), they project has been promising a blog post to explain the situation. As of now, &lt;a href="https://www.zeroclawlabs.ai/blog"&gt;that post is now available&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Over the past few days, a maintainer used aggressive AI automation to review and merge PRs:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Merges went through that shouldn&amp;rsquo;t have.&lt;/li&gt;
&lt;li&gt;In the process of trying to undo the damage, the maintainer&amp;rsquo;s GitHub account was flagged, which triggered enforcement actions on the ZeroClaw org itself.&lt;/li&gt;
&lt;li&gt;That maintainer has been removed from the project.&lt;/li&gt;
&lt;/ul&gt;
&lt;/blockquote&gt;
&lt;p&gt;This sounds strikingly similar to the incident that &lt;a href="https://web.archive.org/web/20260331214457/https://t.me/ZeroClawLabs/11"&gt;occurred about a month ago&lt;/a&gt;, which I also mentioned in yesterday&amp;rsquo;s post:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Earlier today, during routine maintenance, the visibility of the ‎`zeroclaw-labs/zeroclaw` repository was accidentally changed from public to private and was later restored to public.&lt;/p&gt;
&lt;p&gt;After reviewing the GitHub API audit logs and collecting detailed feedback from our engineers, we confirmed that the incident was caused by improper use of an AI agent tool during maintenance.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Obviously, the use agentic workflows in open source projects is an emerging field where best practices have not yet been established. The case of ZeroClaw should be a warning to other projects to keep human review in the loop, or at least to limit the autonomy of agents when a project has numerous contributors. As they say in their blog post:&lt;/p&gt;</description></item><item><title>Testing ZeroClaw, Part 2: ZeroClaw is dead?</title><link>https://muddy.jprs.me/notes/2026-03-31-testing-zeroclaw-part-2-zeroclaw-is-dead/</link><pubDate>Tue, 31 Mar 2026 20:57:00 -0400</pubDate><guid>https://muddy.jprs.me/notes/2026-03-31-testing-zeroclaw-part-2-zeroclaw-is-dead/</guid><description>&lt;p&gt;&lt;a href="https://muddy.jprs.me/notes/2026-03-02-testing-zeroclaw-part-1-setup/"&gt;Earlier this month&lt;/a&gt;, I wrote about setting up one of the many lightweight OpenClaw alternatives, namely ZeroClaw. I had some issues with initial setup, but I got to the point where I could talk with my bot over Telegram.&lt;/p&gt;
&lt;p&gt;Some of my initial enthusiasm for ZeroClaw was dampened by the divergence between the docs and the features available in the release build. The release build was quite out of date due to the breakneck pace of development. In the week or two following my initial setup, the release build pipeline was broken, so even when they released a new tag, there were no new precompiled binaries available. Being forced to compile the Rust binary yourself kind of goes against the project&amp;rsquo;s philosophy of ultra-low resource consumption.&lt;/p&gt;
&lt;p&gt;They eventually fixed the release pipeline and I started casually working on a system where I could send notes and ideas for blog posts to my bot through Telegram and have it turn them into structured Markdown files.&lt;/p&gt;
&lt;p&gt;But two days ago (March 29), I noticed that the ZeroClaw &lt;a href="https://github.com/zeroclaw-labs/zeroclaw"&gt;GitHub repo&lt;/a&gt; was 404ing. On the same day, the project posted the following on &lt;a href="https://x.com/zeroclawlabs/status/2038407120312299524"&gt;Twitter&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Our GitHub repo is currently returning a 404 for some users. We&amp;rsquo;re aware and actively investigating. The repo is public and all code is safe.&lt;/p&gt;</description></item><item><title>How to avoid cognitive surrender to AI</title><link>https://muddy.jprs.me/links/2026-03-29-how-to-avoid-cognitive-surrender-to-ai/</link><pubDate>Sun, 29 Mar 2026 22:59:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-29-how-to-avoid-cognitive-surrender-to-ai/</guid><description>&lt;p&gt;I am sharing a thoughtful article today from Alex Panetta&amp;rsquo;s &lt;em&gt;A.I. For You&lt;/em&gt; on avoiding over-reliance on AI: &amp;ldquo;cognitive debt&amp;rdquo;, &amp;ldquo;epistemic debt&amp;rdquo;, or &amp;ldquo;cognitive surrender&amp;rdquo;.&lt;/p&gt;
&lt;p&gt;A particularly interesting nugget regarding the &amp;ldquo;&lt;a href="https://www.brainonllm.com/"&gt;Your Brain on ChatGPT&lt;/a&gt;&amp;rdquo; article from the MIT Media Lab (yes, &lt;a href="https://en.wikipedia.org/wiki/MIT_Media_Lab#Connections_to_Jeffrey_Epstein"&gt;that&lt;/a&gt; MIT Media Lab):&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The paper is even written to get LLMs to read it carefully. The paper carries instructions telling LLMs which section to read first, which appears to be a clever way to force relevant context atop the context window, as LLMs tend to best remember the beginning and end of conversations — not the middle.&lt;/p&gt;
&lt;/blockquote&gt;</description></item><item><title>Will AI help Canadian police counter a tsunami of fraud?</title><link>https://muddy.jprs.me/links/2026-03-24-will-ai-help-canadian-police-counter-a-tsunami-of-fraud/</link><pubDate>Tue, 24 Mar 2026 21:48:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-24-will-ai-help-canadian-police-counter-a-tsunami-of-fraud/</guid><description>&lt;p&gt;Zak Vescera, writing for the &lt;em&gt;Investigative Journalism Foundation&lt;/em&gt;, observes that fraud cases reported to Canadian police has more than doubled between 2013 and 2024:&lt;/p&gt;
&lt;p&gt;&lt;img src="https://muddy.jprs.me/media/20260324-212544.png" alt="Line chart titled “A tsunami of fraud.” From 2013 to 2024, reported fraud cases in Canada rise sharply from about 80,000 to nearly 180,000, with the steepest increase after 2022. Over the same period, cases cleared by police decline from about 25,000 to under 20,000. The widening gap highlights that fraud reports are surging while police are resolving fewer cases. Source: Statistics Canada."&gt;&lt;/p&gt;
&lt;p&gt;At the same time, the number of cases cleared by Canadian police has fallen. In 2013, the ratio between reported cases and cleared cases was about 3:1; by 2024, this ratio was over 9.5:1.&lt;/p&gt;
&lt;p&gt;The vast majority of fraud cases go unsolved. This is unsurprising given that many are perpetrated over the Internet by individuals overseas and involve methods of sending money that are difficult to recover, such as crypto, gift cards, and physical transfers of cash.&lt;/p&gt;
&lt;p&gt;In response, the National Cybercrime Coordination Centre (NC3) of the RCMP—Canada&amp;rsquo;s national police service—have built a case management system and data portal they hope will eventually be adopted by all Canadian police forces. According to the article, this system is aimed at improving coordination, data sharing, and analysis. The platform will also host a set of AI tools, though the RCMP is vague on details and which are currently implemented. The article gives a few examples: OCR allowing victims to scan gift cards used in fraud rather than typing numbers manually, a tool to classify reports to help police target their investigative resources, and a report generator to simply data sharing when investigations go international.&lt;/p&gt;</description></item><item><title>Using Claude Claude for cross-package statistical audits</title><link>https://muddy.jprs.me/links/2026-03-15-using-claude-claude-for-cross-package-statistical-audits/</link><pubDate>Sun, 15 Mar 2026 22:49:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-15-using-claude-claude-for-cross-package-statistical-audits/</guid><description>&lt;p&gt;Economist Scott Cunningham shared an important example of why we should always report the statistical package and version used in our analyses, as he used Claude Code to produce six versions of the exact same analysis using six different packages in R, Python, and Stata. In a &lt;a href="https://en.wikipedia.org/wiki/Difference_in_differences"&gt;difference-in-differences&lt;/a&gt; analysis of the mental health hospital closures on homicide using the standard &lt;a href="https://bcallaway11.github.io/did/articles/multi-period-did.html"&gt;Callaway and Sant’Anna estimator&lt;/a&gt; (for DiD with multiple time periods), he got very different results for some model specifications.&lt;/p&gt;
&lt;p&gt;Since the specifications and the data were identical between packages, he discovered the divergences occurred due to how the packages handled problems with &lt;a href="https://www.tandfonline.com/doi/full/10.1080/00273171.2011.568786#d1e368"&gt;propensity score&lt;/a&gt; weights. Packages were not necessarily transparent about issues with these weights. If you were not running multiple analyses and comparing results across packages, or else carefully checking propensity score diagnostics, you might never have realized how precarious your results were.&lt;/p&gt;
&lt;p&gt;Prof. Cunningham closes with the following advice:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The fifth point, and the broader point, is that this kind of cross-package, cross-language audit is exactly what Claude Code should be used for. Why? Because this is a task that is time-intensive, high-value, and brutally easy to get wrong. But just one mismatched diagnostic across languages invalidates the entire comparison, even something as simple as sample size values differing across specifications, would flag it. This is both easy and not easy — but it is not the work humans should be doing by hand given how easy it would be to even get that much wrong.&lt;/p&gt;</description></item><item><title>The other half of the ATM–bank teller story</title><link>https://muddy.jprs.me/links/2026-03-11-the-other-half-of-the-atm-bank-teller-story/</link><pubDate>Wed, 11 Mar 2026 23:49:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-11-the-other-half-of-the-atm-bank-teller-story/</guid><description>&lt;p&gt;David Oks had a great post yesterday on the classic parable of how the adoption of ATMs did not lead to the predicted job losses among bank tellers. In fact, the opposite occurred: the number of bank tellers rose. I heard this story recounted several times in early discussions I had about the anticipated effect of AI on labour. I think I first heard it from &lt;a href="https://fee.org/articles/how-ai-and-looser-labor-laws-will-create-jobs/"&gt;Ryan Khurana&lt;/a&gt;. More recently it has been trotted out by US Vice President JD Vance.&lt;/p&gt;
&lt;p&gt;The problem with this story is that the key statistic quoted alongside it, namely that there are more bank tellers than ever before, is no longer true. The famous graph supporting this assertion stops in 2010, and with good reason: the number of bank tellers has sharply fallen since then.&lt;/p&gt;
&lt;p&gt;I think I had come across this fact before, this second half of the famous ATM–bank teller story, but it wasn&amp;rsquo;t until I read David Oks&amp;rsquo;s post that I understood the reason behind it. Quite simply, mobile banking ate physical banks. The ATM didn&amp;rsquo;t reduce the demand for bank tellers because it simply changed the kind of labour they did inside the bank. The iPhone made it so we didn&amp;rsquo;t need to go to the bank at all. It changed the paradigm. Explained this way, it seems obvious. Many new banks (including my own) do not have physical locations and never did.&lt;/p&gt;</description></item><item><title>Editors hate this one weird trick</title><link>https://muddy.jprs.me/notes/2026-03-05-editors-hate-this-one-weird-trick/</link><pubDate>Thu, 05 Mar 2026 20:05:00 -0500</pubDate><guid>https://muddy.jprs.me/notes/2026-03-05-editors-hate-this-one-weird-trick/</guid><description>&lt;p&gt;Given my &lt;a href="https://muddy.jprs.me/links/2026-03-03-the-productivity-shock-coming-to-academic-publishing/"&gt;recent&lt;/a&gt; &lt;a href="https://muddy.jprs.me/notes/2026-02-26-these-academic-journal-ai-policies-aren-t-going-to-last/"&gt;posts&lt;/a&gt; on AI in academic publishing, I just wanted to share this joke from Prof. Arthur Spirling on &lt;a href="https://x.com/arthur_spirling/status/2029006543765520471"&gt;Twitter&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Actually you cant run my paper through Claude to desk reject it because Claude is a regular coauthor of mine. Conflict of interest. Checkmate, editors&lt;/p&gt;
&lt;/blockquote&gt;</description></item><item><title>The productivity shock coming to academic publishing</title><link>https://muddy.jprs.me/links/2026-03-03-the-productivity-shock-coming-to-academic-publishing/</link><pubDate>Tue, 03 Mar 2026 19:33:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-03-03-the-productivity-shock-coming-to-academic-publishing/</guid><description>&lt;p&gt;Today, I wanted to share this piece from economist Scott Cunningham (Baylor University), who wrote about how AI is widening the gap between research and publishing. Or, in economics terms (emphasis mine):&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;But what happens when the same &lt;strong&gt;productivity shock&lt;/strong&gt; hits a system where the bottleneck was never really production in the first place, but rather was a hierarchical journal structure that depended immensely on editor time, skill, discretion and voluntary workers with the same talents called referees for screening quality deemed sufficient for publication?&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The post mentions the Autonomous Policy Evaluation project—the end-to-end AI paper pipeline I &lt;a href="https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/"&gt;wrote about a few weeks ago&lt;/a&gt;—and discusses the likely consequences of this flood of AI-generated papers. Assuming the number of publication slots in reputable journals is relatively fixed, AI-generated papers should add a very large amount of mass to the left side of the paper quality distribution. Acceptance rates will plummet and journals may rely on other signals of quality (name recognition, pedigree, institution) to thin the herd before actually reviewing content. As always, the rich get richer!&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;But this is imperfect, not to mention unfair, and so desk rejection gets noisier: good papers get killed by tired editors and marginally lower quality papers slip through to referees. It’s a cascading failure: volume breaks editors, broken editing wastes referees, wasted referees slow science.&lt;/p&gt;</description></item><item><title>Testing ZeroClaw, Part 1: Setup</title><link>https://muddy.jprs.me/notes/2026-03-02-testing-zeroclaw-part-1-setup/</link><pubDate>Mon, 02 Mar 2026 19:15:00 -0500</pubDate><guid>https://muddy.jprs.me/notes/2026-03-02-testing-zeroclaw-part-1-setup/</guid><description>&lt;p&gt;As mentioned &lt;a href="https://muddy.jprs.me/links/2026-02-24-comparing-the-claw-like-agent-ecosystem/"&gt;last week&lt;/a&gt;, I&amp;rsquo;ve been meaning to test out a personal agent from the Claw-like ecosystem. I settled on testing out &lt;a href="https://github.com/zeroclaw-labs/zeroclaw"&gt;Zeroclaw&lt;/a&gt;, a popular and lightweight OpenClaw alternative that should run well on my Raspberry Pi 4 4GB.&lt;/p&gt;
&lt;p&gt;I wanted to harden my setup as much as possible and opted to running everything in Docker. I started with the &lt;a href="https://github.com/zeroclaw-labs/zeroclaw/blob/main/docker-compose.yml"&gt;official Docker compose file&lt;/a&gt; and added my OpenRouter key. I brought up the pre-built container image and tried sending the basic &amp;ldquo;Hello&amp;rdquo; message to the agent using the CLI. However, I got error because the automatically generated config file defaulted to a version of Claude Sonnet 4 that wasn&amp;rsquo;t available on OpenRouter. I switched to &lt;code&gt;claude-sonnet-4.6&lt;/code&gt; and then &lt;code&gt;gpt-oss-20b&lt;/code&gt; (for &lt;em&gt;much&lt;/em&gt; cheaper testing).&lt;/p&gt;
&lt;p&gt;The Zeroclaw web gateway was a bit of a mess. Of the features I tried, only memory management and the basic status dashboard worked. Trying to talk to the agent through the web interface would give me a black screen (&lt;a href="https://www.reddit.com/r/ZeroClaw/comments/1rc525q/zeroclaw_web_issues/"&gt;here&amp;rsquo;s someone complaining about the same error&lt;/a&gt;). I&amp;rsquo;m still being charged for the tokens, though! The cost tracker always displayed zero, even as I sent CLI and Telegram messages (more on that soon). The configuration editor gave me an error and so did the diagnostics tool.&lt;/p&gt;
&lt;p&gt;The project docs/wiki were helpful for figuring things out, but development is running so far ahead of releases that a bunch of the features referred to aren&amp;rsquo;t available in the current stable version (v0.1.7, from last week). This includes getting and setting specific config options from the CLI and resetting the gateway pairing token. To use these features, you have to compile yourself.&lt;/p&gt;</description></item><item><title>Some examples of just-build-things-ism</title><link>https://muddy.jprs.me/notes/2026-03-01-some-examples-of-just-build-things-ism/</link><pubDate>Sun, 01 Mar 2026 11:58:00 -0500</pubDate><guid>https://muddy.jprs.me/notes/2026-03-01-some-examples-of-just-build-things-ism/</guid><description>&lt;p&gt;The best mantra to come out of the AI era is: &amp;ldquo;You can just build things&amp;rdquo;. (So good OpenAI ripped it off for their &lt;a href="https://www.theverge.com/ai-artificial-intelligence/875550/openais-super-bowl-ad-claims-you-can-just-build-things-with-codex"&gt;Super Bowl ad&lt;/a&gt;.)&lt;/p&gt;
&lt;p&gt;I&amp;rsquo;ve been pretty inspired to see how many people are now building all kinds of incredible tools thanks to advances in AI coding agents, even if they have no previous background in coding (see &lt;a href="https://muddy.jprs.me/links/2026-02-22-oral-texts/"&gt;my post on Havelack.AI from a few days ago&lt;/a&gt;).&lt;/p&gt;
&lt;p&gt;Here are a few more examples I&amp;rsquo;ve been following:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Canadian journalist Alex Panetta writes about his AI-augmented workflow at &lt;a href="https://alexpanetta.substack.com/"&gt;&lt;em&gt;A.I. For You&lt;/em&gt;&lt;/a&gt;. I first came across his work with his debut article &amp;ldquo;&lt;a href="https://alexpanetta.substack.com/p/i-killed-my-doomscrolling-habit-with"&gt;I killed my doomscrolling habit with AI. You can too&lt;/a&gt;&amp;rdquo;. In it, he explains how to vibe code an automated, personalized daily news digest. I&amp;rsquo;ve tried to build something for myself but I haven&amp;rsquo;t gotten it quite right yet. A great follow for big news consumers.&lt;/li&gt;
&lt;li&gt;Economics professor Scott Cunningham, author of the great textbook &lt;a href="https://mixtape.scunning.com/"&gt;&lt;em&gt;Causal Inference: The Mixtape&lt;/em&gt;&lt;/a&gt;, has a &lt;a href="https://x.com/causalinf/status/2026760090947145830"&gt;presentation explaining&lt;/a&gt; how to encourage AI adoption among academic faculty. This starts with faculty &lt;em&gt;experiencing&lt;/em&gt; a killer use case for AI, which he suggests is building slide decks. He shares his tools/agent skills for this use case and more on &lt;a href="https://github.com/scunning1975/MixtapeTools"&gt;GitHub&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;Another economist, Chris Blattman, built a website to share the &lt;a href="https://x.com/cblatts/status/2027018464670491065"&gt;productivity tools&lt;/a&gt; he developed with Claude Code. He provides a tutorial and code on &lt;a href="https://claudeblattman.com/"&gt;Claude Blattman&lt;/a&gt;.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;And of course, &lt;a href="https://tools.simonwillison.net/"&gt;Simon Willison&lt;/a&gt; has been building and sharing tools habitually for years now.&lt;/p&gt;</description></item><item><title>These academic journal AI policies aren't going to last</title><link>https://muddy.jprs.me/notes/2026-02-26-these-academic-journal-ai-policies-aren-t-going-to-last/</link><pubDate>Thu, 26 Feb 2026 16:51:00 -0500</pubDate><guid>https://muddy.jprs.me/notes/2026-02-26-these-academic-journal-ai-policies-aren-t-going-to-last/</guid><description>&lt;p&gt;I recently came across the following policy on the &lt;a href="https://spectrumjournal.ca/index.php/spectrum/about/submissions"&gt;submission page&lt;/a&gt; of an academic journal:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Use of Artificial Intelligence (AI) tools&lt;/strong&gt;: One of the goals of &lt;em&gt;Spectrum&lt;/em&gt; is to stimulate critical thinking and skill development among authors and reviewers alike. &lt;em&gt;Spectrum&lt;/em&gt; discourages the submission of content generated by artificial intelligence (AI)-assisted technologies (such as chatGPT and similar tools). This includes tools that generate text, data, images, figures, or other materials, as well as tools that are used to summarize and synthesize sources. Authors should be aware that such tools are vulnerable to factual inaccuracies, biases, and logical fallacies, and may pose risks to privacy, confidentiality, and copyright.&lt;/p&gt;
&lt;p&gt;If authors choose to submit work created with the assistance of AI tools, such use &lt;strong&gt;must be disclosed&lt;/strong&gt; and described in the submission. The disclosure must include: 1) what system was used, 2) who used it, 3) the time/date of the use, 4) the prompt(s) used to generate the content, and 5) the content in the submission that resulted from use of AI tools. The output from the AI system should also be submitted as supplementary material. Authors must accept full responsibility for the accuracy and integrity of the submission. AI systems do not meet the criteria for authorship, and should not be listed as a co-author.&lt;/p&gt;</description></item><item><title>Agentic engineering patterns</title><link>https://muddy.jprs.me/links/2026-02-25-agentic-engineering-patterns/</link><pubDate>Wed, 25 Feb 2026 16:15:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-25-agentic-engineering-patterns/</guid><description>&lt;p&gt;Simon Willison is building a library of posts covering best practices for using agentic coding tools like Claude Code and OpenAI&amp;rsquo;s Codex. The existing articles cover test-driven development (red/green—ensure tests fail before the change and succeed after it) and AI-assisted code walkthroughs.&lt;/p&gt;</description></item><item><title>Comparing the Claw-like agent ecosystem</title><link>https://muddy.jprs.me/links/2026-02-24-comparing-the-claw-like-agent-ecosystem/</link><pubDate>Tue, 24 Feb 2026 22:44:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-24-comparing-the-claw-like-agent-ecosystem/</guid><description>&lt;p&gt;Chrys Bader has created ClawCharts to track the popularity and growth of OpenClaw and its growing number of competitors.&lt;/p&gt;
&lt;p&gt;I have an unused Raspberry Pi 4 4GB that I&amp;rsquo;ve been meaning to test one of these Claw-like personal agents on (locked down to prevent the security nightmare scenarios we&amp;rsquo;ve seen play out since OpenClaw took off).&lt;/p&gt;
&lt;p&gt;OpenClaw is a bit of a resource hog (which is why so many people are running out to buy Mac Minis), so I&amp;rsquo;ve been looking at the list of lightweight competitors. There is no obvious reason to prefer one over the other, so I&amp;rsquo;ll probably go with the fast-growing &lt;a href="https://github.com/zeroclaw-labs/zeroclaw"&gt;ZeroClaw&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;ZeroClaw offers OAuth connectors for OpenAI and Anthropic subscription plans, but presently neither company is clear on whether this usage is permissible or not. Anthropic recently blew up the OpenClaw community by &lt;a href="https://web.archive.org/web/20260221222303/https://code.claude.com/docs/en/legal-and-compliance#authentication-and-credential-use"&gt;updating their docs&lt;/a&gt; to specifically ban using OAuth outside of Claude Code. An Anthropic employee &lt;a href="https://thenewstack.io/anthropic-agent-sdk-confusion/"&gt;partially walked this back on Twitter&lt;/a&gt;, but there is still no clear statement whether this use case is permitted. Regarding the use of OAuth from OpenAI for OpenClaw (specifically, GPT Codex), Peter Steinberger, creator of OpenClaw, &lt;a href="https://x.com/steipete/status/2024182608746217953"&gt;stated on Twitter&lt;/a&gt;: &amp;ldquo;that already works, OAI publicly said that&amp;rdquo;. No one can seem to find this public statement, but it&amp;rsquo;s worth noting that Steinberger himself is now an OpenAI employee. So, will you get banned for using your ChatGPT Plus/Pro or Claude Pro/Max subscriptions with OpenClaw? Nobody knows.&lt;/p&gt;</description></item><item><title>LLMs automate the erosion of online anonymity</title><link>https://muddy.jprs.me/links/2026-02-23-llms-automate-the-erosion-of-online-anonymity/</link><pubDate>Mon, 23 Feb 2026 22:37:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-23-llms-automate-the-erosion-of-online-anonymity/</guid><description>&lt;p&gt;Economist Florian Ederer &lt;a href="https://x.com/florianederer/status/2025978347809915056"&gt;linked a new preprint&lt;/a&gt; describing the creation of an automated LLM-based pipeline for linking anonymous users across datasets based on unstructured text written by or about them. Prof Ederer is himself famous for &lt;a href="https://florianederer.github.io/ejmr.pdf"&gt;unmasking the IP addresses of users&lt;/a&gt; of the infamous (but influential) Economics Job Market Rumors message board, exploiting a flaw in how usernames were assigned to anonymous posters. For platforms not encoding a user&amp;rsquo;s IP address in their &amp;ldquo;anonymous&amp;rdquo; username, the LLM-based approach involves:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Extracting structured features from free text&lt;/li&gt;
&lt;li&gt;Encoding extracted features to embeddings to compare to candidate profiles&lt;/li&gt;
&lt;li&gt;Reasoning using all available context to identify the most likely match among top candidates&lt;/li&gt;
&lt;li&gt;Calibrate the quality of match by asking the LLM to report confidence&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;I guess it&amp;rsquo;s only a matter of time before someone uses this strategy to unmask Reviewer 2. (Currently this is only possible if Reviewer 2 insists you cite all of the work of the brilliant Dr. X.)&lt;/p&gt;</description></item><item><title>Oral texts</title><link>https://muddy.jprs.me/links/2026-02-22-oral-texts/</link><pubDate>Sun, 22 Feb 2026 13:18:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-22-oral-texts/</guid><description>&lt;p&gt;A major intellectual current in the post-social media age is the rediscovery of media theorists like Marshall McLuhan, Walter Ong, and Neil Postman, whose works seem incredibly prescient in the age of the Internet and the instantaneous and omnipresent mass communication it enables.&lt;/p&gt;
&lt;p&gt;A particular sub-current of this trend is the &lt;a href="https://lindynewsletter.beehiiv.com/p/the-return-of-oral-culture"&gt;return to orality&lt;/a&gt;, a culture rooted in the spoken rather than written word. Indeed, the vast majority of human history is defined by oral culture, and the world&amp;rsquo;s brief sojourn to the written tradition may have finally ended thanks to the Internet.&lt;/p&gt;
&lt;p&gt;One of the most impressive projects to come out of this domain is &lt;a href="https://havelock.ai/"&gt;Havelock.AI&lt;/a&gt;, a tool created by journalist Joe Weisenthal and entirely vibe coded with Claude. The tool analyzes text to give an &amp;ldquo;orality score&amp;rdquo; with supporting analysis. For example, qualified assertions are considered literate, whereas categorical statements are considered oral. The tool defines &lt;a href="https://havelock.ai/methodology"&gt;68 oral/literate markers&lt;/a&gt; based on the framework of Walter Ong. It really is an impressive tool that I recommend checking out.&lt;/p&gt;
&lt;p&gt;I plugged a few of my old articles into the tool and apparently my writing is very much rooted in the written tradition! (This post also scores as strongly literate.)&lt;/p&gt;
&lt;p&gt;&lt;img src="https://muddy.jprs.me/media/havelock-ai-2025-02-22.png" alt="Output from Havelock.AI for this post, referencing the use of a technical term, an epistemic hedge, and an institutional subject as markers of the written tradition"&gt;&lt;/p&gt;</description></item><item><title>Democratizing voice cloning scams</title><link>https://muddy.jprs.me/links/2026-02-18-democratizing-voice-cloning-scams/</link><pubDate>Wed, 18 Feb 2026 22:26:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-18-democratizing-voice-cloning-scams/</guid><description>&lt;p&gt;Jamie Pine has launched Voicebox, a new voice cloning studio built upon the open weight &lt;a href="https://github.com/QwenLM/Qwen3-TTS"&gt;Qwen3-TTS&lt;/a&gt; model. The project is positioned as a free, local alternative to the well-known ElevenLabs voice generator. A &lt;a href="https://voicebox.sh/"&gt;short demo video&lt;/a&gt; is available.&lt;/p&gt;
&lt;p&gt;Obviously, there are legitimate uses for voice cloning technology. But in practice, this will be used to enable AI impersonation scams and spam on a massive scale. The GitHub page for this release isn&amp;rsquo;t exactly encouraging on this front. &lt;a href="https://github.com/jamiepine/voicebox/blob/eb2cd861b19baa16720fd31747071c187c054bc5/README.md"&gt;Demo screenshots&lt;/a&gt; show voice clones of YouTuber Linus Tech Tips, Minecraft creator Markus &amp;ldquo;Notch&amp;rdquo; Persson, and deceased streamer twomad.&lt;/p&gt;
&lt;p&gt;Make sure you have a secret passphrase set up with your family, since your voice is no longer uniquely your own.&lt;/p&gt;</description></item><item><title>Don't let AI do your thinking for you</title><link>https://muddy.jprs.me/links/2026-02-17-don-t-let-ai-do-your-thinking-for-you/</link><pubDate>Tue, 17 Feb 2026 21:11:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-17-don-t-let-ai-do-your-thinking-for-you/</guid><description>&lt;p&gt;Here&amp;rsquo;s a thought-provoking article from Harry Law on &amp;ldquo;The last temptation of Claude&amp;rdquo;—the urge to outsource all of your thinking to AI (and remember, writing is thinking).&lt;/p&gt;
&lt;p&gt;A common theme in the AI commentary I&amp;rsquo;ve been reading lately is the growing importance of taste. AI is sending the cost of creating &amp;ldquo;content&amp;rdquo; (articles, analyses, video, etc.) to zero, even as the attention to consume it all remains fixed. If we want to keep living in a world where AI serves us, we need—more than ever—the discernment to choose the questions worth asking.&lt;/p&gt;
&lt;p&gt;As I put it in my Globe and Mail op-ed on AI and journalism a few years ago:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;AI won’t replace the sort of journalism that holds power accountable, but it could certainly enhance it. After all, you can teach a machine to spot patterns, but you can’t force it to care about your community.&lt;/p&gt;
&lt;/blockquote&gt;</description></item><item><title>In the multiverse of forking paths</title><link>https://muddy.jprs.me/links/2026-02-16-in-the-multiverse-of-forking-paths/</link><pubDate>Mon, 16 Feb 2026 22:49:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-16-in-the-multiverse-of-forking-paths/</guid><description>&lt;p&gt;&lt;img src="https://muddy.jprs.me/media/stark-strange-one.jpg" alt="A scene from Avengers: Infinity War, where Tony Stark asks Dr. Strange, “How many did we win?” Strange replies, “One.”"&gt;&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;STRANGE: I went forward in time to view alternate modelling decisions, to see all the possible outcomes of the coming analysis.&lt;br&gt;
STAR-LORD: How many did you see?&lt;br&gt;
STRANGE: 14,000,605.&lt;br&gt;
STARK: How many did we achieve statistical significance?&lt;br&gt;
STRANGE: One.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Prof. Jessica Hullman recently wrote a piece on Andrew Gelman&amp;rsquo;s blog discussing the use of &amp;lsquo;multiverse analysis&amp;rsquo;, i.e., what if we could see the results of the many slightly different decisions we could have made when constructing a model. This problem is commonly known as the &lt;a href="https://en.wikipedia.org/wiki/Forking_paths_problem"&gt;garden of forking paths&lt;/a&gt;—during an analysis, a researcher is forced to make many small, sometimes arbitrary decisions that can lead to a different result if another researcher tries to independently replicate the analysis. While usually an innocent and inevitable part of the modelling process, these &amp;lsquo;researcher degrees of freedom&amp;rsquo; can also be manipulated to produce a desired result.&lt;/p&gt;
&lt;p&gt;Prof. Hullman points out that multiverse analysis will only become salient as AI coding tools such as Claude Code make it easier than ever to iterate on how we model our research questions.&lt;/p&gt;
&lt;p&gt;Her longer paper with Julia M. Rohrer and Andrew Gelman, &amp;ldquo;What&amp;rsquo;s a multiverse good for anyway?&amp;rdquo; is available &lt;a href="https://osf.io/preprints/psyarxiv/37g29_v1"&gt;here&lt;/a&gt;.&lt;/p&gt;</description></item><item><title>More on vibe researching</title><link>https://muddy.jprs.me/links/2026-02-13-more-on-vibe-researching/</link><pubDate>Fri, 13 Feb 2026 23:49:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-13-more-on-vibe-researching/</guid><description>&lt;p&gt;To follow on &lt;a href="https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/"&gt;yesterday&amp;rsquo;s post&lt;/a&gt; on AI-produced research, here is a reflection on &amp;ldquo;vibe researching&amp;rdquo; from Prof. Joshua Gans of the University of Toronto&amp;rsquo;s Rotman School of Management. Since the release of the first &amp;ldquo;reasoning&amp;rdquo; models in late 2024, he has gone all in on experimenting with AI-first research.&lt;/p&gt;
&lt;p&gt;One of the key takeaways is that he found himself pursuing low quality ideas to completion more often, precisely because the cost of choosing to continue to pursue a questionable idea has been lowered. Sycophancy is a problem, too. With an AI cheerleader, it is easy to convince yourself you have a result when you do not.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Those ideas were all fine but not high quality, and what is worse, I didn’t realise that they weren’t that significant until external referees said so. I didn’t realise it because they were reasonably hard to do, and I was happy to have solved them.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I will note that (human) peer reviewers cannot be the levee that stops the flood of middling AI research: the system of uncompensated labour that undergirds all of academic publishing is already strained to bursting, as every editor desperate to find referees for a paper will tell you.&lt;/p&gt;
&lt;p&gt;Prof. Gans concludes his year-long experiment in &amp;ldquo;vibe researching&amp;rdquo; was a failure, despite publishing many working papers and publishing a handful of them:&lt;/p&gt;</description></item><item><title>An end-to-end AI pipeline for policy evaluation papers</title><link>https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/</link><pubDate>Thu, 12 Feb 2026 19:11:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/</guid><description>&lt;p&gt;Prof. David Yanagizawa-Drott from the Social Catalyst Lab at the University of Zurich has launched Project APE (Autonomous Policy Evaluation), an end-to-end AI pipeline to generate policy evaluation papers. The vast majority of policies around the world are never rigorously evaluated, so it would certainly be useful if we were able to do so in an automated fashion.&lt;/p&gt;
&lt;p&gt;Claude Code is the heart of the project, but other models are used to review the outputs and provide journal-style referee reports. All the coding is done in R (though Python is called in some scripts). Currently, judging is done by Gemini 3 Flash to compare against published research in top economics journals:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Blind comparison: An LLM judge compares two papers without knowing which is AI-generated
Position swapping: Each pair is judged twice with paper order swapped to control for bias
TrueSkill ratings: Papers accumulate skill ratings that update after each match&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The project&amp;rsquo;s home page lists the AI&amp;rsquo;s current &amp;ldquo;win rate&amp;rdquo; at 3.5% in head-to-head matchups against human-written papers.&lt;/p&gt;
&lt;p&gt;Prof. Yanagizawa-Drott says &amp;ldquo;Currently it requires at a minimum some initial human input for each paper,&amp;rdquo; although he does not specify exactly what. If we look at &lt;a href="https://github.com/SocialCatalystLab/ape-papers/blob/main/apep_0264/v1/initialization.md"&gt;&lt;code&gt;initialization.json&lt;/code&gt;&lt;/a&gt; that can be found in each paper&amp;rsquo;s directory, we see the following questions with user-provided inputs:&lt;/p&gt;
&lt;blockquote&gt;
&lt;ol&gt;
&lt;li&gt;Policy domain: What policy area interests you?&lt;/li&gt;
&lt;li&gt;Method: Which identification method?&lt;/li&gt;
&lt;li&gt;Data era: Modern or historical data?&lt;/li&gt;
&lt;li&gt;API keys: Did you configure data API keys?&lt;/li&gt;
&lt;li&gt;External review: Include external model reviews?&lt;/li&gt;
&lt;li&gt;Risk appetite: Exploration vs exploitation?&lt;/li&gt;
&lt;li&gt;Other preferences: Any other preferences or constraints?&lt;/li&gt;
&lt;/ol&gt;
&lt;/blockquote&gt;
&lt;p&gt;The code, reviews, manuscript, and even the results of the initial idea generation process are all available on &lt;a href="https://github.com/SocialCatalystLab/ape-papers"&gt;GitHub&lt;/a&gt;. Their immediate goal is to generate a sample of 1,000 papers and run human evaluations on them (at time of posting, there are 264 papers in the GitHub repository).&lt;/p&gt;</description></item><item><title>Why a Canadian news site just launched an AI publishing tool</title><link>https://muddy.jprs.me/links/2026-02-09-why-a-canadian-news-site-just-launched-an-ai-publishing-tool/</link><pubDate>Mon, 09 Feb 2026 19:49:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-09-why-a-canadian-news-site-just-launched-an-ai-publishing-tool/</guid><description>&lt;p&gt;It&amp;rsquo;s no secret that Canadian journalism (like journalism everywhere) is in trouble. Newsrooms face a steady stream of layoffs despite a couple hundred million Canadian dollars of direct and indirect &lt;a href="https://macdonaldlaurier.ca/government-subsidies-for-canadas-media-were-supposed-to-be-temporary-but-they-keep-on-growing-and-could-be-here-to-stay-dave-snow-in-the-hub/"&gt;government subsidies&lt;/a&gt; every year. The vast majority of outlets eligible for these subsidies take advantage of them, and combined they can &lt;a href="https://macdonaldlaurier.ca/government-subsidies-for-canadas-media-were-supposed-to-be-temporary-but-they-keep-on-growing-and-could-be-here-to-stay-dave-snow-in-the-hub/"&gt;subsidize half of a journalist&amp;rsquo;s salary&lt;/a&gt;. News organizations are desperate to diversify their revenue streams.&lt;/p&gt;
&lt;p&gt;&lt;a href="thehub.ca/2025/03/28/rudyard-griffiths-and-sean-speer-the-hub-is-receiving-over-60000-from-the-government-and-donating-it-all-to-charity-will-the-rest-of-canadas-subsidized-media-disclose-what-theyre-gettin/"&gt;&lt;em&gt;The Hub&lt;/em&gt;&lt;/a&gt; is a right-leaning publication launched in 2021 with a focus on policy and politics. Notably, the outlet &lt;a href="https://macdonaldlaurier.ca/the-ottawa-declaration-on-canadian-journalism/"&gt;declines&lt;/a&gt; or &lt;a href="https://thehub.ca/2025/03/28/rudyard-griffiths-and-sean-speer-the-hub-is-receiving-over-60000-from-the-government-and-donating-it-all-to-charity-will-the-rest-of-canadas-subsidized-media-disclose-what-theyre-gettin/"&gt;donates&lt;/a&gt; their subsidies, citing a valid concern that the scale of such subsidies &lt;a href="https://thehub.ca/2024/07/08/deepdive-government-funding-of-the-news-industry-is-eroding-canadians-trust-in-the-media/"&gt;threaten the perceived trustworthiness and independence&lt;/a&gt; of the media.&lt;/p&gt;
&lt;p&gt;In late January 2026, &lt;em&gt;The Hub&lt;/em&gt; &lt;a href="https://thehub.ca/2026/01/28/why-we-are-launching-newsbox-for-the-hubs-paid-subscribers/"&gt;launched NewsBox&lt;/a&gt;, an AI-powered publishing tool. NewsBox aims to make it easier for creators to transform their content (written, audio, or video) into other formats, such as speeches, essays, or talking points, while maintaining the author&amp;rsquo;s distinct voice. You can see examples of the tool&amp;rsquo;s output on new articles in &lt;em&gt;The Hub&lt;/em&gt;, each of which is accompanied by an AI-generated summary and list of quotes at the top of the page. There is also a &amp;ldquo;Hub AI&amp;rdquo; chatbot in the sidebar of every article.&lt;/p&gt;
&lt;p&gt;The app very much uses &lt;em&gt;The Hub&lt;/em&gt;&amp;rsquo;s branding, prominently featuring the outlet’s co-creators, who also created NewsBox. While their pitch talks about preserving creators&amp;rsquo; voices to avoid the &amp;ldquo;soulless prose&amp;rdquo; and &amp;ldquo;slop&amp;rdquo; outputted by ChatGPT and similar tools, I have to wonder if tighter integration of AI into the news and opinion side of the operation will &lt;a href="https://reutersinstitute.politics.ox.ac.uk/generative-ai-and-news-report-2025-how-people-think-about-ais-role-journalism-and-society"&gt;raise its own issues with trust&lt;/a&gt;. &lt;em&gt;The Hub&lt;/em&gt; has always been fairly tech-friendly, including a &lt;a href="https://thehub.ca/2023/12/20/marc-edge-canadas-news-media-need-a-plan-and-some-help-to-find-a-way-forward/"&gt;longstanding&lt;/a&gt; &lt;a href="https://thehub.ca/category/meta/"&gt;sponsorship&lt;/a&gt; by Meta.&lt;/p&gt;</description></item><item><title>Anthropic's statistical analysis skill doesn't get statistical significance quite right</title><link>https://muddy.jprs.me/links/2026-02-06-anthropic-s-statistical-analysis-skill-doesn-t-get-statistical-significance-quite-right/</link><pubDate>Fri, 06 Feb 2026 19:30:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-06-anthropic-s-statistical-analysis-skill-doesn-t-get-statistical-significance-quite-right/</guid><description>&lt;p&gt;Anthropic&amp;rsquo;s new statistical analysis skill demonstrates a common misunderstanding of statistical significance:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Statistical significance means the difference is unlikely due to chance.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;But this phrasing isn&amp;rsquo;t quite right. The p-value in Null Hypothesis Significance Testing is not about the probability the results are &amp;ldquo;due to chance&amp;rdquo;; it is the probability—under the null hypothesis and the model assumptions—of observing results at least as extreme as the ones we obtained. In other words, the p-value summarizes how compatible the data are with the null, given our modelling choices. What it does not tell you is the probability that the null hypothesis is true.&lt;/p&gt;
&lt;p&gt;Statistician Andrew Gelman gave a good definition for statistical significance in a 2015 &lt;a href="https://statmodeling.stat.columbia.edu/2015/07/21/a-bad-definition-of-statistical-significance-from-the-u-s-department-of-health-and-human-services-effective-health-care-program/"&gt;blog post&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;A mathematical technique to measure the strength of evidence from a single study. Statistical significance is conventionally declared when the p-value is less than 0.05. The p-value is the probability of seeing a result as strong as observed or greater, under the &lt;em&gt;null hypothesis&lt;/em&gt; (which is commonly the hypothesis that there is no effect). Thus, the smaller the p-value, the less consistent are the data with the null hypothesis under this measure.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;As some of the commenters in this blog post observe, simply being able to parrot a technically accurate definition of a p-value does not necessarily make us better at applying statistical significance in practice. It is certainly true that statistical significance is widely misused in scientific publishing as a threshold to distinguish signal from noise (or to be fancy, a &amp;ldquo;lexicographic decision rule&amp;rdquo;), which is why &lt;a href="https://sites.stat.columbia.edu/gelman/research/published/abandon.pdf"&gt;some scientists have argued that we should abandon it as the default statistical paradigm for research&lt;/a&gt;.&lt;/p&gt;</description></item></channel></rss>