<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Open Data on Big Muddy</title><link>https://muddy.jprs.me/tags/open-data/</link><description>Recent content in Open Data on Big Muddy</description><generator>Hugo</generator><language>en-US</language><lastBuildDate>Fri, 03 Apr 2026 23:40:00 -0400</lastBuildDate><atom:link href="https://muddy.jprs.me/tags/open-data/index.xml" rel="self" type="application/rss+xml"/><item><title>The triumph of the data raccoons</title><link>https://muddy.jprs.me/notes/2026-04-03-the-triumph-of-the-data-raccoons/</link><pubDate>Fri, 03 Apr 2026 23:40:00 -0400</pubDate><guid>https://muddy.jprs.me/notes/2026-04-03-the-triumph-of-the-data-raccoons/</guid><description>&lt;p&gt;My PhD co-supervisor at the University of Toronto, Dr. David Fisman, liked to use the term &amp;ldquo;data raccoon&amp;rdquo; to describe the work of using messy, incomplete, hard-to-work-with data to do serious research. Or, as he described it in &lt;a href="https://www.ourcommons.ca/documentviewer/en/43-1/HESA/meeting-22/evidence#Int-10851517"&gt;testimony&lt;/a&gt; to the Canadian House of Commons in May 2020 (emphasis mine):&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;I&amp;rsquo;ll tell you, my group at University of Toronto call ourselves &amp;ldquo;data raccoons&amp;rdquo;, because we&amp;rsquo;ve sort of managed to thrive for about 15 years on &lt;strong&gt;data that most people regard as garbage&lt;/strong&gt;, so it&amp;rsquo;s sort of a bit of the normal state of affairs for us with public health data analysis.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;It&amp;rsquo;s an unmistakably Toronto metaphor—the city isn&amp;rsquo;t called the &lt;a href="https://www.npr.org/2018/09/16/647599627/theres-no-stopping-toronto-s-uber-raccoon"&gt;raccoon capital of the world&lt;/a&gt; for nothing!&lt;/p&gt;
&lt;p&gt;It occurred to me recently that data raccoons have basically taken over the world. The basis of the AI revolution is vast quantities of text dredged from the Internet, none of which was written for its final purpose of training the &lt;em&gt;deus ex machina&lt;/em&gt;. Arguably the most important dataset for training LLMs has been &lt;a href="https://en.wikipedia.org/wiki/Common_Crawl_Foundation"&gt;Common Crawl&lt;/a&gt;, a mostly uncurated snapshot of the Internet that has been running since 2007. According to a &lt;a href="https://www.mozillafoundation.org/en/research/library/generative-ai-training-data/common-crawl/"&gt;Mozilla report&lt;/a&gt; from 2024, Common Crawl was used in two thirds of LLMs developed in the formative period between 2019 and 2023, and the archive also comprised 80% of tokens in OpenAI&amp;rsquo;s GPT-3. Unsurprisingly, the Common Crawl Foundation has received &lt;a href="https://archive.is/NS9MI"&gt;financial support&lt;/a&gt; from AI companies in recent years, all the while being accused of abetting these same companies to train their models on paywalled articles.&lt;/p&gt;</description></item><item><title>How SARS-CoV-2 variants get named on GitHub</title><link>https://muddy.jprs.me/links/2026-03-26-how-sars-cov-2-variants-get-named-on-github/</link><pubDate>Thu, 26 Mar 2026 07:00:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-26-how-sars-cov-2-variants-get-named-on-github/</guid><description>&lt;p&gt;Bioinformatics has long been an unusually collaborative and transparent field, with genomes, protein structures, and other complex biological data habitually deposited into open databases during the course of research. The situation was no different at the outset of the COVID-19 pandemic, when a small group of scientists developed the &lt;a href="https://en.wikipedia.org/wiki/Phylogenetic_Assignment_of_Named_Global_Outbreak_Lineages"&gt;Pango nomenclature&lt;/a&gt; for classifying variants of the SARS-CoV-2 virus. Outside of a handful of Greek-letter &amp;ldquo;variants of concern&amp;rdquo; names assigned by the World Health Organization, the Pango nomenclature is the standard for tracking the evolution of the SARS-CoV-2 virus. You may recall names such as B.1.1.7 (Alpha or the UK variant), B.1.351 (Beta or the South African variant), and P.1 (Gamma or the Brazilian variant). You can see a complete list of active SARS-CoV-2 lineages using the Pango nomenclature &lt;a href="https://cov-lineages.org/lineage_list.html"&gt;here&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;By &lt;a href="https://github.com/cov-lineages/pango-designation/issues/1"&gt;August 2020&lt;/a&gt;, the work of defining new lineages of SARS-CoV-2 had moved to &lt;a href="https://github.com/cov-lineages/pango-designation"&gt;GitHub&lt;/a&gt;, where the scientific process could happen in transparent and collaborative way. The definition of new lineages happens on proposals submitted as GitHub issues. In &lt;a href="https://github.com/cov-lineages/pango-designation/issues/1988"&gt;May 2023&lt;/a&gt;, a second &lt;a href="https://github.com/sars-cov-2-variants/lineage-proposals"&gt;GitHub repository&lt;/a&gt; was opened to move discussions of smaller or less clear lineages out of the main repository. These discussions can be promoted to the main repository, as this &lt;a href="https://github.com/sars-cov-2-variants/lineage-proposals/issues/2199"&gt;issue tracking LP.8.1 sub-lineages&lt;/a&gt; was in &lt;a href="https://github.com/cov-lineages/pango-designation/issues/2978"&gt;May 2025&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;The work of defining new lineages of SARS-CoV-2 continues to this day on the GitHub repository, as the virus continues to mutate and evolve. And bioinformatics continues to be a shining beacon for open science for the rest of us to learn from.&lt;/p&gt;</description></item><item><title>Vandalism of OpenStreetMap</title><link>https://muddy.jprs.me/links/2026-03-23-vandalism-of-openstreetmap/</link><pubDate>Mon, 23 Mar 2026 17:26:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-23-vandalism-of-openstreetmap/</guid><description>&lt;p&gt;&lt;a href="https://www.openstreetmap.org/"&gt;OpenStreetMap&lt;/a&gt; (OSM) is an open, community-driven map database powering countless apps and services and used by organizations including Amazon, Apple, Microsoft, Uber, Mapbox, and Wikimedia. In short, it is foundational infrastructure for the web. For regions with active communities (particularly in Europe), OSM is often noted for the superiority of its data on features such as cycling routes, hiking trails, and footpaths.&lt;/p&gt;
&lt;p&gt;The &lt;a href="https://en.wikipedia.org/w/index.php?title=OpenStreetMap&amp;amp;oldid=1340732815#quality_assurance"&gt;Wikipedia article for OpenStreetMap&lt;/a&gt; documents several instances of data vandalism, which OSM is vulnerable to as a crowdsourced project. Three incidents stood out:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;&lt;a href="https://www.wired.com/2012/01/osm-google-accusation/"&gt;In 2012&lt;/a&gt;, Google fired two &amp;ldquo;rogue contractors&amp;rdquo; for vandalizing the OSM database, intentionally adding false data such as reversing the direction of one-way streets.&lt;/li&gt;
&lt;li&gt;&lt;a href="https://arstechnica.com/information-technology/2018/08/data-vandal-changes-name-of-new-york-city-to-jewtropolis-across-multiple-apps/"&gt;In 2018&lt;/a&gt;, a vandal made several viciously antisemitic edits to place names around New York City. While quickly reverted at the source, these changes nonetheless propagated into downstream applications pulling data from MapBox, such as Zillow, Snapchat, Citibike, and Wikipedia.&lt;/li&gt;
&lt;li&gt;Users of the mobile game &lt;a href="https://www.mdpi.com/2220-9964/9/4/197"&gt;Pokémon GO&lt;/a&gt; regularly vandalize the OSM database underlying the game to gain a gameplay advantage, although the authors of the research article on this subject note this vandalism tends to be transitory rather than sustained.&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;Side note: I was amused to note how strong Google&amp;rsquo;s regional results bias is for &amp;ldquo;OSM&amp;rdquo;—the entire first page is taken up by results related to the Orchestre symphonique de Montréal.&lt;/p&gt;</description></item><item><title>Properly the work of federal public health agencies</title><link>https://muddy.jprs.me/links/2026-03-22-properly-the-work-of-federal-public-health-agencies/</link><pubDate>Sun, 22 Mar 2026 23:38:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-22-properly-the-work-of-federal-public-health-agencies/</guid><description>&lt;p&gt;One of the reasons I started this blog was to have a place to put down posts and articles that have lodged themselves in my brain. The wind-down announcement of the &lt;a href="https://en.wikipedia.org/wiki/COVID_Tracking_Project"&gt;COVID Tracking Project&lt;/a&gt;, a volunteer-led COVID-19 data tracking collaboration, is one such article.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;But the work itself—compiling, cleaning, standardizing, and making sense of COVID-19 data from 56 individual states and territories—&lt;em&gt;is properly the work of federal public health agencies&lt;/em&gt;.&lt;/strong&gt; Not only because these efforts are a governmental responsibility—which they are—but because federal teams have access to far more comprehensive data than we do, and can mandate compliance with at least some standards and requirements.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;After one year of work, the COVID Tracking Project decided to quite collecting data on COVID-19 in the United States, because they recognized that the work of collecting a comparable, national-level dataset was the responsibility of federal government agencies.&lt;/p&gt;
&lt;p&gt;As someone who co-led the &lt;a href="https://opencovid.ca/"&gt;COVID-19 Canada Open Data Working Group&lt;/a&gt;, which curated &lt;a href="https://github.com/ccodwg/Covid19Canada"&gt;COVID-19&lt;/a&gt; &lt;a href="https://github.com/ccodwg/CovidTimelineCanada"&gt;data&lt;/a&gt; for Canada until the end of 2023, I think about this article a lot. It&amp;rsquo;s a good read, and it speaks to how essential open data was to filling in the gaps in the national and international understanding of the COVID-19 pandemic.&lt;/p&gt;</description></item><item><title>Fight club at the bird feeder</title><link>https://muddy.jprs.me/links/2026-03-20-fight-club-at-the-bird-feeder/</link><pubDate>Fri, 20 Mar 2026 07:00:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-20-fight-club-at-the-bird-feeder/</guid><description>&lt;p&gt;&lt;em&gt;Alternate title: Blue Jay brutally feeder mogs Tufted Titmouse&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src="https://muddy.jprs.me/media/20260319-214037.png" alt="Network showing dominance hierarchy among 13 common feeder birds; the Blue Jay wins against 10 species and loses to 3"&gt;&lt;/p&gt;
&lt;p&gt;From the Cornell Lab of Ornithology, a pretty neat article about dominance hierarchies at the bird feeder using over 7,600 observations collected by citizen scientists contributing to &lt;a href="https://feederwatch.org/"&gt;Project Feeder Watch&lt;/a&gt;. Essentially, bird watchers reported instances when one bird species successfully displaced another at the bird feeder, and used this network of comparisons to build a dominance hierarchy. By using information contained within the network, you can even compare birds that are rarely observed together. Not all dominance patterns are linear, however, as the article reports:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;A separate analysis uncovered some dominance triangles in which three birds had one-to-one relationships independent of each other, like a game of birdy rock-paper-scissors. For example, the House Finch dominates the Purple Finch, and the Purple Finch dominates the Dark-eyed Junco, but the junco dominates House Finch.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The full paper is here: &lt;a href="https://doi.org/10.1093/beheco/arx108"&gt;Fighting over food unites the birds of North America in a continental dominance hierarchy&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;This work is reminiscent of &lt;a href="https://www.cochrane.org/authors/handbooks-and-manuals/handbook/current/chapter-11"&gt;network meta-analysis&lt;/a&gt;, in which three or more interventions (e.g., drugs) are compared using both direct and indirect evidence. For example, if there are studies comparing drug A versus drug B and drug B versus drug C, we can infer the comparison between drug A and drug C, even if no study has ever directly compared them.&lt;/p&gt;</description></item><item><title>geoBoundaries: An open database of political administrative boundaries</title><link>https://muddy.jprs.me/links/2026-03-13-geoboundaries-an-open-database-of-political-administrative-boundaries/</link><pubDate>Fri, 13 Mar 2026 17:05:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-13-geoboundaries-an-open-database-of-political-administrative-boundaries/</guid><description>&lt;p&gt;Today I discovered geoBoundaries, a &lt;a href="https://creativecommons.org/licenses/by/4.0/"&gt;CC BY 4.0&lt;/a&gt;-licensed database of political administrative boundaries covering the entire world. It is notable for its high level of detail, going from ADM0 (country), ADM1 (states/provinces), ADM2 (counties/departments or municipalities), to ADM3 (municipalities or sub-municipalities) for many countries. My go-to source for world map files is &lt;a href="https://www.naturalearthdata.com/"&gt;Natural Earth&lt;/a&gt;, which is limited to ADM0 and ADM1 but is in the public domain. Natural Earth also includes some physical geography like water and bathymetry, while geoBoundaries is focused solely on political administrative boundaries. Both datasets deal with disputed boundaries, which is an endless source of tension in the &lt;a href="https://github.com/nvkelso/natural-earth-vector/issues"&gt;Natural Earth GitHub&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;An R package for retrieving data from geoBoundaries, &lt;a href="https://dieghernan.github.io/202602_geobounds/"&gt;&lt;code&gt;geobounds&lt;/code&gt;&lt;/a&gt;, was released in February. A similar package for Natural Earth, &lt;a href="https://github.com/ropensci/rnaturalearth"&gt;&lt;code&gt;rnaturalearth&lt;/code&gt;&lt;/a&gt;, has long been maintained by rOpenSci.&lt;/p&gt;</description></item><item><title>What will the paper of the future look like?</title><link>https://muddy.jprs.me/links/2026-03-10-what-will-the-paper-of-the-future-look-like/</link><pubDate>Tue, 10 Mar 2026 23:48:00 -0400</pubDate><guid>https://muddy.jprs.me/links/2026-03-10-what-will-the-paper-of-the-future-look-like/</guid><description>&lt;p&gt;I am sharing today a short blog post by the Institute for Replication: &amp;ldquo;What will the paper of the future look like?&amp;rdquo;&lt;/p&gt;
&lt;p&gt;In short: research looking more like &lt;a href="https://www.youtube.com/watch?v=zwRdO9_GGhY"&gt;software development&lt;/a&gt; (as presaged by Prof. Richard McElreath, author of the excellent &lt;em&gt;Statistical Rethinking&lt;/em&gt;), with the ability to reuse common material, formalize results, and remix analyses built into the pipeline.&lt;/p&gt;</description></item><item><title>Open By Default: A database of access to information requests to the Canadian government</title><link>https://muddy.jprs.me/links/2026-03-07-open-by-default-a-database-of-access-to-information-requests-to-the-canadian-government/</link><pubDate>Sat, 07 Mar 2026 14:32:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-03-07-open-by-default-a-database-of-access-to-information-requests-to-the-canadian-government/</guid><description>&lt;p&gt;In Canada, any person or corporation in the country can make a request for general records to any agency of the federal government through the &lt;em&gt;Access to Information Act&lt;/em&gt; (the equivalent in the United States is the &lt;em&gt;Freedom of Information Act&lt;/em&gt;). The government provides a &lt;a href="https://open.canada.ca/en/search/ati"&gt;searchable database&lt;/a&gt; of completed requests, but includes only a summary of the request and the number of pages of responsive material. The actual documents turned over are not included. However, completed request packages may be informally re-requested, and should you do so, someone from the relevant agency will (usually) send them to you eventually.&lt;/p&gt;
&lt;p&gt;This re-request process has its limits. It can takes weeks or months for the documents to be sent, and the database itself only goes back to January 2020 (they used to delete records older than two years, but stopped doing this some time after 2020). Occasionally, they will never send the documents at all, and all you can do is either re-request them again or open a formal access to information request (which will cost you $5).&lt;/p&gt;
&lt;p&gt;Making it easier to access completed access to information requests is why the Investigative Journalism Foundation built &lt;a href="https://theijf.org/open-by-default"&gt;Open By Default&lt;/a&gt;, &amp;ldquo;the biggest database of internal government documents never before made publicly accessible&amp;rdquo;. It includes documents from completed access to information requests &lt;a href="https://theijf.org/open-by-default-methodology"&gt;obtained&lt;/a&gt; using both automated (presumably the re-request form) and manual processes (donations from trusted partners, particularly of documents from before the online re-request form was available). The files are cleaned and OCRed into one beautiful, searchable database.&lt;/p&gt;</description></item><item><title>US Medicaid data gets DOGE'd</title><link>https://muddy.jprs.me/links/2026-02-14-us-medicaid-data-gets-doge-d/</link><pubDate>Sat, 14 Feb 2026 10:29:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-14-us-medicaid-data-gets-doge-d/</guid><description>&lt;p&gt;The US Health and Human Services DOGE team (I guess DOGE still exists in some form) just released a new aggregated, provider-level Medicaid claims database covering January 2018 through December 2024. With this dataset, you can track the monthly claims for each procedure (by HCPCS Code) and provider over time.&lt;/p&gt;
&lt;p&gt;Even if the &lt;a href="https://www.axios.com/2026/02/14/elon-musk-doge-medicaid-fraud-hhs-database"&gt;framing around this dataset&amp;rsquo;s release is partisan&lt;/a&gt;—tied to allegations of Medicaid fraud in Minnesota—it is a genuine advance in transparency for the US&amp;rsquo;s third largest spending program. No doubt this accomplishment required a lot of work on the backend to harmonize countless fragmented datasets into one tidy schema. These data were difficult to access before, and now they are free for anyone to use. Journalists, policy researchers, and companies working in the US healthcare sector will benefit the most, but every taxpayer benefits from added transparency about where their tax dollars go.&lt;/p&gt;
&lt;p&gt;I would say there is the potential for these data to be misused to spark witch hunts, but this is more or less the stated purpose for this data release. Per Elon Musk: &amp;ldquo;Medicaid data has been open sourced, so the level of fraud is easy to identify.&amp;rdquo; If you go on &lt;a href="https://x.com/DOGE_HHS/status/2022370909211021376"&gt;Twitter&lt;/a&gt;, you will find several people have already plugged in the dataset to Claude Code and trumpeted their ASCII tables of providers flagged for potential fraud. Inevitably, some of these providers targeted by public scrutiny for their unusual billing patterns will have perfectly innocent explanations. But if &lt;a href="https://x.com/charlesornstein/status/2022395807484514409"&gt;ProPublica is excited&lt;/a&gt; about the release of this new dataset, then so am I.&lt;/p&gt;</description></item><item><title>An end-to-end AI pipeline for policy evaluation papers</title><link>https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/</link><pubDate>Thu, 12 Feb 2026 19:11:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-12-an-end-to-end-ai-pipeline-for-policy-evaluation-papers/</guid><description>&lt;p&gt;Prof. David Yanagizawa-Drott from the Social Catalyst Lab at the University of Zurich has launched Project APE (Autonomous Policy Evaluation), an end-to-end AI pipeline to generate policy evaluation papers. The vast majority of policies around the world are never rigorously evaluated, so it would certainly be useful if we were able to do so in an automated fashion.&lt;/p&gt;
&lt;p&gt;Claude Code is the heart of the project, but other models are used to review the outputs and provide journal-style referee reports. All the coding is done in R (though Python is called in some scripts). Currently, judging is done by Gemini 3 Flash to compare against published research in top economics journals:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Blind comparison: An LLM judge compares two papers without knowing which is AI-generated
Position swapping: Each pair is judged twice with paper order swapped to control for bias
TrueSkill ratings: Papers accumulate skill ratings that update after each match&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The project&amp;rsquo;s home page lists the AI&amp;rsquo;s current &amp;ldquo;win rate&amp;rdquo; at 3.5% in head-to-head matchups against human-written papers.&lt;/p&gt;
&lt;p&gt;Prof. Yanagizawa-Drott says &amp;ldquo;Currently it requires at a minimum some initial human input for each paper,&amp;rdquo; although he does not specify exactly what. If we look at &lt;a href="https://github.com/SocialCatalystLab/ape-papers/blob/main/apep_0264/v1/initialization.md"&gt;&lt;code&gt;initialization.json&lt;/code&gt;&lt;/a&gt; that can be found in each paper&amp;rsquo;s directory, we see the following questions with user-provided inputs:&lt;/p&gt;
&lt;blockquote&gt;
&lt;ol&gt;
&lt;li&gt;Policy domain: What policy area interests you?&lt;/li&gt;
&lt;li&gt;Method: Which identification method?&lt;/li&gt;
&lt;li&gt;Data era: Modern or historical data?&lt;/li&gt;
&lt;li&gt;API keys: Did you configure data API keys?&lt;/li&gt;
&lt;li&gt;External review: Include external model reviews?&lt;/li&gt;
&lt;li&gt;Risk appetite: Exploration vs exploitation?&lt;/li&gt;
&lt;li&gt;Other preferences: Any other preferences or constraints?&lt;/li&gt;
&lt;/ol&gt;
&lt;/blockquote&gt;
&lt;p&gt;The code, reviews, manuscript, and even the results of the initial idea generation process are all available on &lt;a href="https://github.com/SocialCatalystLab/ape-papers"&gt;GitHub&lt;/a&gt;. Their immediate goal is to generate a sample of 1,000 papers and run human evaluations on them (at time of posting, there are 264 papers in the GitHub repository).&lt;/p&gt;</description></item><item><title>The case for sharing clinical trial data</title><link>https://muddy.jprs.me/links/2026-02-10-the-case-for-sharing-clinical-trial-data/</link><pubDate>Tue, 10 Feb 2026 19:39:00 -0500</pubDate><guid>https://muddy.jprs.me/links/2026-02-10-the-case-for-sharing-clinical-trial-data/</guid><description>&lt;p&gt;Saloni Dattani of the excellent &lt;a href="https://worksinprogress.co/"&gt;&lt;em&gt;Works in Progress&lt;/em&gt;&lt;/a&gt; magazine (and formerly of &lt;a href="https://ourworldindata.org/"&gt;Our World in Data&lt;/a&gt;) launched a new Substack today called &lt;em&gt;The Clinical Trials Abundance blog&lt;/em&gt;. The first post is on the case for sharing clinical trial data. We have been gradually moving toward mandatory reporting of &lt;a href="https://www.npr.org/sections/health-shots/2019/10/11/769348119/canadas-decision-to-make-public-more-clinical-trial-data-puts-pressure-on-fda"&gt;clinical trial &lt;em&gt;results&lt;/em&gt;&lt;/a&gt; (though enforcement is &lt;a href="https://www.trialstracker.net/"&gt;another&lt;/a&gt; &lt;a href="https://fdaaa.trialstracker.net/"&gt;question&lt;/a&gt;), but sharing data would be one step further. Even though clinical trials rely on the trust (and often money) of the public, it can be very difficult to gain access to the raw results, even if journal article authors claim they are &amp;ldquo;available upon request&amp;rdquo;. A norm of clinical trial data sharing would not only increase the confidence in published results but also aid future drug development, reduce expensive redundancy, and improve meta-analyses (which are often forced to rely on heterogeneous summary measures).&lt;/p&gt;</description></item></channel></rss>