<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Science on Big Muddy</title><link>https://muddy.jprs.me/tags/science/</link><description>Recent content in Science on Big Muddy</description><generator>Hugo</generator><language>en-US</language><lastBuildDate>Tue, 07 Apr 2026 18:27:00 -0400</lastBuildDate><atom:link href="https://muddy.jprs.me/tags/science/index.xml" rel="self" type="application/rss+xml"/><item><title>What is a public opinion poll without the public?</title><link>https://muddy.jprs.me/links/2026-04-07-what-is-a-public-opinion-poll-without-the-public/</link><pubDate>Tue, 07 Apr 2026 18:27:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-04-07-what-is-a-public-opinion-poll-without-the-public/</guid><description>&lt;p&gt;A few days ago, two professors (Leif Weatherby and Benjamin Recht) published an opinion piece in the &lt;em&gt;New York Times&lt;/em&gt; calling attention to Axios publishing a &lt;a href="https://www.axios.com/2026/03/19/olivia-walton-heartland-forward-maternal-health"&gt;story&lt;/a&gt; on maternal health using invented polling results:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;A recent Axios story on maternal health policy referred to “findings” that a majority of people trusted their doctors and nurses. On the surface, there’s nothing unusual about that. What wasn’t originally mentioned, however, was that these findings were made up.&lt;/p&gt;
&lt;p&gt;Clicking through the links revealed (as did a subsequent editor’s note and clarification by Axios) that the public opinion poll was a computer simulation run by the artificial intelligence start-up Aaru. No people were involved in the creation of these opinions.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The piece goes on to argue that this so-called &amp;ldquo;silicon sampling&amp;rdquo; is seductive because good public opinion polling is expensive, hard to do, and still prone to bias. But this shortcut magnifies the the problem of bias rather than solving it.&lt;/p&gt;
&lt;p&gt;I&amp;rsquo;ve read a little bit about this strategy of using LLM-generated survey participants in the context of social science research in a series of posts (mostly from Prof. Jessica Hullman) over on &lt;a href="https://statmodeling.stat.columbia.edu/"&gt;Andrew Gelman&amp;rsquo;s blog&lt;/a&gt;:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href="https://statmodeling.stat.columbia.edu/2025/12/19/validating-language-models-as-study-participants-how-its-being-done-why-it-fails-and-what-works-instead/"&gt;Validating language models as study participants: How it’s being done, why it fails, and what works instead&lt;/a&gt; (2025-12-19)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://statmodeling.stat.columbia.edu/2025/08/26/survey-statistics-thomas-lumley-writes-about-interviewing-your-laptop/"&gt;Survey Statistics: Thomas Lumley writes about Interviewing your Laptop&lt;/a&gt; (2025-08-26)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://statmodeling.stat.columbia.edu/2025/08/15/when-does-it-make-sense-to-talk-about-llms-having-beliefs/"&gt;When does it make sense to talk about LLMs having beliefs?&lt;/a&gt; (2025-08-15)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://statmodeling.stat.columbia.edu/2025/06/12/better-and-worse-ways-to-mix-human-and-llm-responses-in-behavioral-research-but-you-still-have-to-figure-what-youre-measuring/"&gt;Better and worse ways to mix human and LLM responses in behavioral research (but you still have to figure what you’re measuring)&lt;/a&gt; (2025-06-12)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://statmodeling.stat.columbia.edu/2025/05/29/llms-as-behavioral-study-participants/"&gt;LLMs as behavioral study participants&lt;/a&gt; (2025-05-29)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Silicon sampling seems moderately interesting from a research perspective, but I can&amp;rsquo;t help but agree with the &lt;em&gt;New York Times&lt;/em&gt; opinion piece authors that this will be ruinous for the already waning trust in public opinion polling. If you didn&amp;rsquo;t bother to ask the public, then why should the public care what you &amp;ldquo;find&amp;rdquo;? I think there is probably a lot of utility in using LLM samples to aid in designing and validating surveys, though.&lt;/p&gt;</description></item><item><title>How SARS-CoV-2 variants get named on GitHub</title><link>https://muddy.jprs.me/links/2026-03-26-how-sars-cov-2-variants-get-named-on-github/</link><pubDate>Thu, 26 Mar 2026 07:00:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-26-how-sars-cov-2-variants-get-named-on-github/</guid><description>&lt;p&gt;Bioinformatics has long been an unusually collaborative and transparent field, with genomes, protein structures, and other complex biological data habitually deposited into open databases during the course of research. The situation was no different at the outset of the COVID-19 pandemic, when a small group of scientists developed the &lt;a href="https://en.wikipedia.org/wiki/Phylogenetic_Assignment_of_Named_Global_Outbreak_Lineages"&gt;Pango nomenclature&lt;/a&gt; for classifying variants of the SARS-CoV-2 virus. Outside of a handful of Greek-letter &amp;ldquo;variants of concern&amp;rdquo; names assigned by the World Health Organization, the Pango nomenclature is the standard for tracking the evolution of the SARS-CoV-2 virus. You may recall names such as B.1.1.7 (Alpha or the UK variant), B.1.351 (Beta or the South African variant), and P.1 (Gamma or the Brazilian variant). You can see a complete list of active SARS-CoV-2 lineages using the Pango nomenclature &lt;a href="https://cov-lineages.org/lineage_list.html"&gt;here&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;By &lt;a href="https://github.com/cov-lineages/pango-designation/issues/1"&gt;August 2020&lt;/a&gt;, the work of defining new lineages of SARS-CoV-2 had moved to &lt;a href="https://github.com/cov-lineages/pango-designation"&gt;GitHub&lt;/a&gt;, where the scientific process could happen in transparent and collaborative way. The definition of new lineages happens on proposals submitted as GitHub issues. In &lt;a href="https://github.com/cov-lineages/pango-designation/issues/1988"&gt;May 2023&lt;/a&gt;, a second &lt;a href="https://github.com/sars-cov-2-variants/lineage-proposals"&gt;GitHub repository&lt;/a&gt; was opened to move discussions of smaller or less clear lineages out of the main repository. These discussions can be promoted to the main repository, as this &lt;a href="https://github.com/sars-cov-2-variants/lineage-proposals/issues/2199"&gt;issue tracking LP.8.1 sub-lineages&lt;/a&gt; was in &lt;a href="https://github.com/cov-lineages/pango-designation/issues/2978"&gt;May 2025&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;The work of defining new lineages of SARS-CoV-2 continues to this day on the GitHub repository, as the virus continues to mutate and evolve. And bioinformatics continues to be a shining beacon for open science for the rest of us to learn from.&lt;/p&gt;</description></item><item><title>Fight club at the bird feeder</title><link>https://muddy.jprs.me/links/2026-03-20-fight-club-at-the-bird-feeder/</link><pubDate>Fri, 20 Mar 2026 07:00:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-20-fight-club-at-the-bird-feeder/</guid><description>&lt;p&gt;&lt;em&gt;Alternate title: Blue Jay brutally feeder mogs Tufted Titmouse&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src="https://muddy.jprs.me/media/20260319-214037.png" alt="Network showing dominance hierarchy among 13 common feeder birds; the Blue Jay wins against 10 species and loses to 3"&gt;&lt;/p&gt;
&lt;p&gt;From the Cornell Lab of Ornithology, a pretty neat article about dominance hierarchies at the bird feeder using over 7,600 observations collected by citizen scientists contributing to &lt;a href="https://feederwatch.org/"&gt;Project Feeder Watch&lt;/a&gt;. Essentially, bird watchers reported instances when one bird species successfully displaced another at the bird feeder, and used this network of comparisons to build a dominance hierarchy. By using information contained within the network, you can even compare birds that are rarely observed together. Not all dominance patterns are linear, however, as the article reports:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;A separate analysis uncovered some dominance triangles in which three birds had one-to-one relationships independent of each other, like a game of birdy rock-paper-scissors. For example, the House Finch dominates the Purple Finch, and the Purple Finch dominates the Dark-eyed Junco, but the junco dominates House Finch.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The full paper is here: &lt;a href="https://doi.org/10.1093/beheco/arx108"&gt;Fighting over food unites the birds of North America in a continental dominance hierarchy&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;This work is reminiscent of &lt;a href="https://www.cochrane.org/authors/handbooks-and-manuals/handbook/current/chapter-11"&gt;network meta-analysis&lt;/a&gt;, in which three or more interventions (e.g., drugs) are compared using both direct and indirect evidence. For example, if there are studies comparing drug A versus drug B and drug B versus drug C, we can infer the comparison between drug A and drug C, even if no study has ever directly compared them.&lt;/p&gt;</description></item><item><title>More on vibe researching</title><link>https://muddy.jprs.me/links/2026-02-13-more-on-vibe-researching/</link><pubDate>Fri, 13 Feb 2026 23:49:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-13-more-on-vibe-researching/</guid><description>&lt;p&gt;To follow on &lt;a href="https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/"&gt;yesterday&amp;rsquo;s post&lt;/a&gt; on AI-produced research, here is a reflection on &amp;ldquo;vibe researching&amp;rdquo; from Prof. Joshua Gans of the University of Toronto&amp;rsquo;s Rotman School of Management. Since the release of the first &amp;ldquo;reasoning&amp;rdquo; models in late 2024, he has gone all in on experimenting with AI-first research.&lt;/p&gt;
&lt;p&gt;One of the key takeaways is that he found himself pursuing low quality ideas to completion more often, precisely because the cost of choosing to continue to pursue a questionable idea has been lowered. Sycophancy is a problem, too. With an AI cheerleader, it is easy to convince yourself you have a result when you do not.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Those ideas were all fine but not high quality, and what is worse, I didn’t realise that they weren’t that significant until external referees said so. I didn’t realise it because they were reasonably hard to do, and I was happy to have solved them.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I will note that (human) peer reviewers cannot be the levee that stops the flood of middling AI research: the system of uncompensated labour that undergirds all of academic publishing is already strained to bursting, as every editor desperate to find referees for a paper will tell you.&lt;/p&gt;
&lt;p&gt;Prof. Gans concludes his year-long experiment in &amp;ldquo;vibe researching&amp;rdquo; was a failure, despite publishing many working papers and publishing a handful of them:&lt;/p&gt;</description></item><item><title>An end-to-end AI pipeline for policy evaluation papers</title><link>https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/</link><pubDate>Thu, 12 Feb 2026 19:11:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/</guid><description>&lt;p&gt;Prof. David Yanagizawa-Drott from the Social Catalyst Lab at the University of Zurich has launched Project APE (Autonomous Policy Evaluation), an end-to-end AI pipeline to generate policy evaluation papers. The vast majority of policies around the world are never rigorously evaluated, so it would certainly be useful if we were able to do so in an automated fashion.&lt;/p&gt;
&lt;p&gt;Claude Code is the heart of the project, but other models are used to review the outputs and provide journal-style referee reports. All the coding is done in R (though Python is called in some scripts). Currently, judging is done by Gemini 3 Flash to compare against published research in top economics journals:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Blind comparison: An LLM judge compares two papers without knowing which is AI-generated
Position swapping: Each pair is judged twice with paper order swapped to control for bias
TrueSkill ratings: Papers accumulate skill ratings that update after each match&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The project&amp;rsquo;s home page lists the AI&amp;rsquo;s current &amp;ldquo;win rate&amp;rdquo; at 3.5% in head-to-head matchups against human-written papers.&lt;/p&gt;
&lt;p&gt;Prof. Yanagizawa-Drott says &amp;ldquo;Currently it requires at a minimum some initial human input for each paper,&amp;rdquo; although he does not specify exactly what. If we look at &lt;a href="https://github.com/SocialCatalystLab/ape-papers/blob/main/apep_0264/v1/initialization.md"&gt;&lt;code&gt;initialization.json&lt;/code&gt;&lt;/a&gt; that can be found in each paper&amp;rsquo;s directory, we see the following questions with user-provided inputs:&lt;/p&gt;
&lt;blockquote&gt;
&lt;ol&gt;
&lt;li&gt;Policy domain: What policy area interests you?&lt;/li&gt;
&lt;li&gt;Method: Which identification method?&lt;/li&gt;
&lt;li&gt;Data era: Modern or historical data?&lt;/li&gt;
&lt;li&gt;API keys: Did you configure data API keys?&lt;/li&gt;
&lt;li&gt;External review: Include external model reviews?&lt;/li&gt;
&lt;li&gt;Risk appetite: Exploration vs exploitation?&lt;/li&gt;
&lt;li&gt;Other preferences: Any other preferences or constraints?&lt;/li&gt;
&lt;/ol&gt;
&lt;/blockquote&gt;
&lt;p&gt;The code, reviews, manuscript, and even the results of the initial idea generation process are all available on &lt;a href="https://github.com/SocialCatalystLab/ape-papers"&gt;GitHub&lt;/a&gt;. Their immediate goal is to generate a sample of 1,000 papers and run human evaluations on them (at time of posting, there are 264 papers in the GitHub repository).&lt;/p&gt;</description></item></channel></rss>