<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Academia on Big Muddy</title><link>https://muddy.jprs.me/tags/academia/</link><description>Recent content in Academia on Big Muddy</description><generator>Hugo</generator><language>en-US</language><lastBuildDate>Fri, 10 Apr 2026 18:27:00 -0400</lastBuildDate><atom:link href="https://muddy.jprs.me/tags/academia/index.xml" rel="self" type="application/rss+xml"/><item><title>Scientists invent a fake disease, AI picks it up, other scientists cite it</title><link>https://muddy.jprs.me/links/2026-04-10-scientists-invent-a-fake-disease-ai-picks-it-up-other-scientists-cite-it/</link><pubDate>Fri, 10 Apr 2026 18:27:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-04-10-scientists-invent-a-fake-disease-ai-picks-it-up-other-scientists-cite-it/</guid><description>&lt;p&gt;A somewhat disturbing bit of reporting from &lt;em&gt;Nature&lt;/em&gt; tells the story of bixonimania, a fake eye disease invented by Swedish medical researcher Almira Osmanovic Thunström and her team. She seeded the idea for the fake disease in a series of ridiculous, joke-filled blog posts and preprints in mid-2024.&lt;/p&gt;
&lt;p&gt;Because AI can be overly credulous with its sourcing (how often do Google&amp;rsquo;s AI answers confident cite random Reddit posts for the bulk of an answer?), the disease got picked up as an &amp;ldquo;emerging term&amp;rdquo; by the leading chatbots. The preprints even got cited a handful of times in real publications, which is further evidence that scientists don&amp;rsquo;t read the papers they cite (I guess the modern equivalent of copying citations from other papers is having AI dredge the literature for you).&lt;/p&gt;
&lt;p&gt;I can see AI agents being exploited by those pushing dubious medical diagnoses to flood the Internet and preprint servers with articles aimed at convincing LLMs of the validity of their positions. That is if the agents aren&amp;rsquo;t too busy &lt;a href="https://theshamblog.com/an-ai-agent-published-a-hit-piece-on-me/"&gt;spinning of websites to defame&lt;/a&gt; those who incur their wrath.&lt;/p&gt;</description></item><item><title>AI makes it easier to generate fake papers, too</title><link>https://muddy.jprs.me/links/2026-04-08-ai-makes-it-easier-to-generate-fake-papers-too/</link><pubDate>Wed, 08 Apr 2026 20:09:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-04-08-ai-makes-it-easier-to-generate-fake-papers-too/</guid><description>&lt;p&gt;Here&amp;rsquo;s a fun project from Tyler Vigen, creator of the famous &lt;a href="https://tylervigen.com/spurious-correlations"&gt;Spurious Correlations&lt;/a&gt; page (which has been cited as a cautionary tale in many a science class). Using his database of real but spurious correlations (created by calculating the Pearson correlation coefficient &lt;em&gt;r&lt;/em&gt; between a very large number of variables and picking out the hits), he used AI to create amusing fake manuscripts expounding on these statistical flukes as if they were real research questions.&lt;/p&gt;
&lt;p&gt;These papers were generated in January 2024, and as &lt;a href="https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/"&gt;previously discussed&lt;/a&gt; on this blog, the pipeline for end-to-end paper generation has come a long way in two years. I have no doubt Tyler could make these paper&amp;rsquo;s sound much more convincing using today&amp;rsquo;s models, though of course his goal here is to make you laugh (and think), not to trick you. But I have no doubt there will be many scholars adopting this data dredging strategy to generate &amp;ldquo;real&amp;rdquo; papers, contributing to a deluge of papers &lt;a href="https://muddy.jprs.me/links/2026-03-03-the-productivity-shock-coming-to-academic-publishing/"&gt;flooding &lt;/a&gt;&lt;a href="https://muddy.jprs.me/links/2026-03-03-the-productivity-shock-coming-to-academic-publishing/"&gt;the academic publishing system&lt;/a&gt;.&lt;/p&gt;</description></item><item><title>Some insight into writing a book using Quarto</title><link>https://muddy.jprs.me/links/2026-03-16-some-insight-into-writing-a-book-using-quarto/</link><pubDate>Mon, 16 Mar 2026 20:48:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-16-some-insight-into-writing-a-book-using-quarto/</guid><description>&lt;p&gt;Prof. Kieran Healy (Sociology, Yale University) shares some nice insight into the process of writing a book in Quarto using R in this post. The output screenshots he shares look beautiful, and the idea of deploying the same content as a clean PDF &lt;em&gt;and&lt;/em&gt; a responsive website is awesome. A full draft of the book, &lt;em&gt;Data Visualization: A Practical Introduction (Second Edition)&lt;/em&gt;, is available as a website &lt;a href="https://socviz.co/"&gt;here&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;I have grown increasingly tired of writing in any format other than a plain text file I can easily version control and move around, so the idea of writing a book in Quarto is appealing to me (as long as it has enough technical content to justify the format).&lt;/p&gt;</description></item><item><title>What will the paper of the future look like?</title><link>https://muddy.jprs.me/links/2026-03-10-what-will-the-paper-of-the-future-look-like/</link><pubDate>Tue, 10 Mar 2026 23:48:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-10-what-will-the-paper-of-the-future-look-like/</guid><description>&lt;p&gt;I am sharing today a short blog post by the Institute for Replication: &amp;ldquo;What will the paper of the future look like?&amp;rdquo;&lt;/p&gt;
&lt;p&gt;In short: research looking more like &lt;a href="https://www.youtube.com/watch?v=zwRdO9_GGhY"&gt;software development&lt;/a&gt; (as presaged by Prof. Richard McElreath, author of the excellent &lt;em&gt;Statistical Rethinking&lt;/em&gt;), with the ability to reuse common material, formalize results, and remix analyses built into the pipeline.&lt;/p&gt;</description></item><item><title>Editors hate this one weird trick</title><link>https://muddy.jprs.me/notes/2026-03-05-editors-hate-this-one-weird-trick/</link><pubDate>Thu, 05 Mar 2026 20:05:00 -0500</pubDate><guid>https://muddy.jprs.me/notes/2026-03-05-editors-hate-this-one-weird-trick/</guid><description>&lt;p&gt;Given my &lt;a href="https://muddy.jprs.me/links/2026-03-03-the-productivity-shock-coming-to-academic-publishing/"&gt;recent&lt;/a&gt; &lt;a href="https://muddy.jprs.me/notes/2026-02-26-these-academic-journal-ai-policies-aren-t-going-to-last/"&gt;posts&lt;/a&gt; on AI in academic publishing, I just wanted to share this joke from Prof. Arthur Spirling on &lt;a href="https://x.com/arthur_spirling/status/2029006543765520471"&gt;Twitter&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Actually you cant run my paper through Claude to desk reject it because Claude is a regular coauthor of mine. Conflict of interest. Checkmate, editors&lt;/p&gt;
&lt;/blockquote&gt;</description></item><item><title>The productivity shock coming to academic publishing</title><link>https://muddy.jprs.me/links/2026-03-03-the-productivity-shock-coming-to-academic-publishing/</link><pubDate>Tue, 03 Mar 2026 19:33:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-03-03-the-productivity-shock-coming-to-academic-publishing/</guid><description>&lt;p&gt;Today, I wanted to share this piece from economist Scott Cunningham (Baylor University), who wrote about how AI is widening the gap between research and publishing. Or, in economics terms (emphasis mine):&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;But what happens when the same &lt;strong&gt;productivity shock&lt;/strong&gt; hits a system where the bottleneck was never really production in the first place, but rather was a hierarchical journal structure that depended immensely on editor time, skill, discretion and voluntary workers with the same talents called referees for screening quality deemed sufficient for publication?&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The post mentions the Autonomous Policy Evaluation project—the end-to-end AI paper pipeline I &lt;a href="https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/"&gt;wrote about a few weeks ago&lt;/a&gt;—and discusses the likely consequences of this flood of AI-generated papers. Assuming the number of publication slots in reputable journals is relatively fixed, AI-generated papers should add a very large amount of mass to the left side of the paper quality distribution. Acceptance rates will plummet and journals may rely on other signals of quality (name recognition, pedigree, institution) to thin the herd before actually reviewing content. As always, the rich get richer!&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;But this is imperfect, not to mention unfair, and so desk rejection gets noisier: good papers get killed by tired editors and marginally lower quality papers slip through to referees. It’s a cascading failure: volume breaks editors, broken editing wastes referees, wasted referees slow science.&lt;/p&gt;</description></item><item><title>Some examples of just-build-things-ism</title><link>https://muddy.jprs.me/notes/2026-03-01-some-examples-of-just-build-things-ism/</link><pubDate>Sun, 01 Mar 2026 11:58:00 -0500</pubDate><guid>https://muddy.jprs.me/notes/2026-03-01-some-examples-of-just-build-things-ism/</guid><description>&lt;p&gt;The best mantra to come out of the AI era is: &amp;ldquo;You can just build things&amp;rdquo;. (So good OpenAI ripped it off for their &lt;a href="https://www.theverge.com/ai-artificial-intelligence/875550/openais-super-bowl-ad-claims-you-can-just-build-things-with-codex"&gt;Super Bowl ad&lt;/a&gt;.)&lt;/p&gt;
&lt;p&gt;I&amp;rsquo;ve been pretty inspired to see how many people are now building all kinds of incredible tools thanks to advances in AI coding agents, even if they have no previous background in coding (see &lt;a href="https://muddy.jprs.me/links/2026-02-22-oral-texts/"&gt;my post on Havelack.AI from a few days ago&lt;/a&gt;).&lt;/p&gt;
&lt;p&gt;Here are a few more examples I&amp;rsquo;ve been following:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Canadian journalist Alex Panetta writes about his AI-augmented workflow at &lt;a href="https://alexpanetta.substack.com/"&gt;&lt;em&gt;A.I. For You&lt;/em&gt;&lt;/a&gt;. I first came across his work with his debut article &amp;ldquo;&lt;a href="https://alexpanetta.substack.com/p/i-killed-my-doomscrolling-habit-with"&gt;I killed my doomscrolling habit with AI. You can too&lt;/a&gt;&amp;rdquo;. In it, he explains how to vibe code an automated, personalized daily news digest. I&amp;rsquo;ve tried to build something for myself but I haven&amp;rsquo;t gotten it quite right yet. A great follow for big news consumers.&lt;/li&gt;
&lt;li&gt;Economics professor Scott Cunningham, author of the great textbook &lt;a href="https://mixtape.scunning.com/"&gt;&lt;em&gt;Causal Inference: The Mixtape&lt;/em&gt;&lt;/a&gt;, has a &lt;a href="https://x.com/causalinf/status/2026760090947145830"&gt;presentation explaining&lt;/a&gt; how to encourage AI adoption among academic faculty. This starts with faculty &lt;em&gt;experiencing&lt;/em&gt; a killer use case for AI, which he suggests is building slide decks. He shares his tools/agent skills for this use case and more on &lt;a href="https://github.com/scunning1975/MixtapeTools"&gt;GitHub&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;Another economist, Chris Blattman, built a website to share the &lt;a href="https://x.com/cblatts/status/2027018464670491065"&gt;productivity tools&lt;/a&gt; he developed with Claude Code. He provides a tutorial and code on &lt;a href="https://claudeblattman.com/"&gt;Claude Blattman&lt;/a&gt;.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;And of course, &lt;a href="https://tools.simonwillison.net/"&gt;Simon Willison&lt;/a&gt; has been building and sharing tools habitually for years now.&lt;/p&gt;</description></item><item><title>These academic journal AI policies aren't going to last</title><link>https://muddy.jprs.me/notes/2026-02-26-these-academic-journal-ai-policies-aren-t-going-to-last/</link><pubDate>Thu, 26 Feb 2026 16:51:00 -0500</pubDate><guid>https://muddy.jprs.me/notes/2026-02-26-these-academic-journal-ai-policies-aren-t-going-to-last/</guid><description>&lt;p&gt;I recently came across the following policy on the &lt;a href="https://spectrumjournal.ca/index.php/spectrum/about/submissions"&gt;submission page&lt;/a&gt; of an academic journal:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Use of Artificial Intelligence (AI) tools&lt;/strong&gt;: One of the goals of &lt;em&gt;Spectrum&lt;/em&gt; is to stimulate critical thinking and skill development among authors and reviewers alike. &lt;em&gt;Spectrum&lt;/em&gt; discourages the submission of content generated by artificial intelligence (AI)-assisted technologies (such as chatGPT and similar tools). This includes tools that generate text, data, images, figures, or other materials, as well as tools that are used to summarize and synthesize sources. Authors should be aware that such tools are vulnerable to factual inaccuracies, biases, and logical fallacies, and may pose risks to privacy, confidentiality, and copyright.&lt;/p&gt;
&lt;p&gt;If authors choose to submit work created with the assistance of AI tools, such use &lt;strong&gt;must be disclosed&lt;/strong&gt; and described in the submission. The disclosure must include: 1) what system was used, 2) who used it, 3) the time/date of the use, 4) the prompt(s) used to generate the content, and 5) the content in the submission that resulted from use of AI tools. The output from the AI system should also be submitted as supplementary material. Authors must accept full responsibility for the accuracy and integrity of the submission. AI systems do not meet the criteria for authorship, and should not be listed as a co-author.&lt;/p&gt;</description></item><item><title>LLMs automate the erosion of online anonymity</title><link>https://muddy.jprs.me/links/2026-02-23-llms-automate-the-erosion-of-online-anonymity/</link><pubDate>Mon, 23 Feb 2026 22:37:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-23-llms-automate-the-erosion-of-online-anonymity/</guid><description>&lt;p&gt;Economist Florian Ederer &lt;a href="https://x.com/florianederer/status/2025978347809915056"&gt;linked a new preprint&lt;/a&gt; describing the creation of an automated LLM-based pipeline for linking anonymous users across datasets based on unstructured text written by or about them. Prof Ederer is himself famous for &lt;a href="https://florianederer.github.io/ejmr.pdf"&gt;unmasking the IP addresses of users&lt;/a&gt; of the infamous (but influential) Economics Job Market Rumors message board, exploiting a flaw in how usernames were assigned to anonymous posters. For platforms not encoding a user&amp;rsquo;s IP address in their &amp;ldquo;anonymous&amp;rdquo; username, the LLM-based approach involves:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Extracting structured features from free text&lt;/li&gt;
&lt;li&gt;Encoding extracted features to embeddings to compare to candidate profiles&lt;/li&gt;
&lt;li&gt;Reasoning using all available context to identify the most likely match among top candidates&lt;/li&gt;
&lt;li&gt;Calibrate the quality of match by asking the LLM to report confidence&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;I guess it&amp;rsquo;s only a matter of time before someone uses this strategy to unmask Reviewer 2. (Currently this is only possible if Reviewer 2 insists you cite all of the work of the brilliant Dr. X.)&lt;/p&gt;</description></item><item><title>More on vibe researching</title><link>https://muddy.jprs.me/links/2026-02-13-more-on-vibe-researching/</link><pubDate>Fri, 13 Feb 2026 23:49:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-13-more-on-vibe-researching/</guid><description>&lt;p&gt;To follow on &lt;a href="https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/"&gt;yesterday&amp;rsquo;s post&lt;/a&gt; on AI-produced research, here is a reflection on &amp;ldquo;vibe researching&amp;rdquo; from Prof. Joshua Gans of the University of Toronto&amp;rsquo;s Rotman School of Management. Since the release of the first &amp;ldquo;reasoning&amp;rdquo; models in late 2024, he has gone all in on experimenting with AI-first research.&lt;/p&gt;
&lt;p&gt;One of the key takeaways is that he found himself pursuing low quality ideas to completion more often, precisely because the cost of choosing to continue to pursue a questionable idea has been lowered. Sycophancy is a problem, too. With an AI cheerleader, it is easy to convince yourself you have a result when you do not.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Those ideas were all fine but not high quality, and what is worse, I didn’t realise that they weren’t that significant until external referees said so. I didn’t realise it because they were reasonably hard to do, and I was happy to have solved them.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I will note that (human) peer reviewers cannot be the levee that stops the flood of middling AI research: the system of uncompensated labour that undergirds all of academic publishing is already strained to bursting, as every editor desperate to find referees for a paper will tell you.&lt;/p&gt;
&lt;p&gt;Prof. Gans concludes his year-long experiment in &amp;ldquo;vibe researching&amp;rdquo; was a failure, despite publishing many working papers and publishing a handful of them:&lt;/p&gt;</description></item><item><title>An end-to-end AI pipeline for policy evaluation papers</title><link>https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/</link><pubDate>Thu, 12 Feb 2026 19:11:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/</guid><description>&lt;p&gt;Prof. David Yanagizawa-Drott from the Social Catalyst Lab at the University of Zurich has launched Project APE (Autonomous Policy Evaluation), an end-to-end AI pipeline to generate policy evaluation papers. The vast majority of policies around the world are never rigorously evaluated, so it would certainly be useful if we were able to do so in an automated fashion.&lt;/p&gt;
&lt;p&gt;Claude Code is the heart of the project, but other models are used to review the outputs and provide journal-style referee reports. All the coding is done in R (though Python is called in some scripts). Currently, judging is done by Gemini 3 Flash to compare against published research in top economics journals:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Blind comparison: An LLM judge compares two papers without knowing which is AI-generated
Position swapping: Each pair is judged twice with paper order swapped to control for bias
TrueSkill ratings: Papers accumulate skill ratings that update after each match&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The project&amp;rsquo;s home page lists the AI&amp;rsquo;s current &amp;ldquo;win rate&amp;rdquo; at 3.5% in head-to-head matchups against human-written papers.&lt;/p&gt;
&lt;p&gt;Prof. Yanagizawa-Drott says &amp;ldquo;Currently it requires at a minimum some initial human input for each paper,&amp;rdquo; although he does not specify exactly what. If we look at &lt;a href="https://github.com/SocialCatalystLab/ape-papers/blob/main/apep_0264/v1/initialization.md"&gt;&lt;code&gt;initialization.json&lt;/code&gt;&lt;/a&gt; that can be found in each paper&amp;rsquo;s directory, we see the following questions with user-provided inputs:&lt;/p&gt;
&lt;blockquote&gt;
&lt;ol&gt;
&lt;li&gt;Policy domain: What policy area interests you?&lt;/li&gt;
&lt;li&gt;Method: Which identification method?&lt;/li&gt;
&lt;li&gt;Data era: Modern or historical data?&lt;/li&gt;
&lt;li&gt;API keys: Did you configure data API keys?&lt;/li&gt;
&lt;li&gt;External review: Include external model reviews?&lt;/li&gt;
&lt;li&gt;Risk appetite: Exploration vs exploitation?&lt;/li&gt;
&lt;li&gt;Other preferences: Any other preferences or constraints?&lt;/li&gt;
&lt;/ol&gt;
&lt;/blockquote&gt;
&lt;p&gt;The code, reviews, manuscript, and even the results of the initial idea generation process are all available on &lt;a href="https://github.com/SocialCatalystLab/ape-papers"&gt;GitHub&lt;/a&gt;. Their immediate goal is to generate a sample of 1,000 papers and run human evaluations on them (at time of posting, there are 264 papers in the GitHub repository).&lt;/p&gt;</description></item></channel></rss>