A UN dataset too big for any single journalist to read in full. An AI that read it, found the surprises, rendered them in brand format, and wrote the verification guides — while the human wrote four short sentences.
The UN's SDG dataflow is one of the most comprehensive datasets ever assembled. It tracks 195 countries across indicators spanning remittances, refugee flows, biodiversity, parliamentary representation, water law, and a hundred things in between. UN SDG Global Dataflow: 15 streams, 195 countries, millions of data points per year.
It is also, in its raw form, almost completely unreadable by a human on a deadline. The data is encoded in SDMX: Statistical Data and Metadata eXchange. Dense XML format used by statistical agencies worldwide.SDMX format, spread across dozens of indicator codes, and arrives as compressed CSV slices you have to know how to ask for.
Someone at a news organisation would normally need a data journalist, an API expert, two weeks, and a research budget. Instead, here is what actually happened: one editor typed four short prompts into an AI coding assistant. What came back changed what a newsroom data workflow can look like.
This is not a demo. These are real cards, with real data, ready to print. What follows is the story of how they were made — and what it means for the newsrooms that haven't tried this yet.
Imagine you hand a talented new analyst a single sentence and walk away. You come back an hour later to find a fully documented API client sitting on your desk, handling rate limits, resume-from-failure, parallel downloads, and compressed storage. That is, essentially, what happened.
The AI didn't just execute the instruction — it made engineering choices. It chose
WAL mode
for the SQLite database, knowing that parallel download jobs might write simultaneously.
It chose gzip compression for the CSV outputs. It built a proper --force flag.
These are decisions an experienced data engineer would make — not things the human specified.
The AI's downloader fetched the UN's full catalog, probed each dataflow for recent years, and produced a structured inventory. The SDG Harmonized Global Dataflow (version 1.24) turned out to be the richest source — 2025 and 2026 data available, covering everything the analysis needed.
Raw data isn't news. A country's remittance cost as a percentage of $200 isn't a headline. The fact that sending money from Johannesburg to Mumbai costs 8.5 times more than sending it from London — that's a headline.
The instruction to the AI was deceptively simple. What it had to do was anything but.
The AI ran the analysis three times, each round deeper than the last. Between rounds, the human added two lines of direction — nudges, not rewrites. The AI went online between rounds to research what other journalists had already done with similar datasets, so it could find the angles that hadn't been covered yet.
Remittance cost gap · Biodiversity Target 14 · Gender-balanced but old parliaments · Ukraine vs Afghanistan refugee rate · UN's 3% remittance benchmark nearly unmet
~15 words. That's the entire editorial direction for round 2.
India remittance corridor spread (£3.82 vs $32.78) · India productivity rank (11th of 181) · Migration deaths concentration · Youth employment strategy gap · more
~12 words. The AI had already been documenting its own research ideas in a notes file.
Added: Seed SMTA crop diversity (India #2 globally) · Livestock backup gap · Water law vs actual participation · Biodiversity plan vs reality. Each lead supported by a reproducible CSV.
Take the Ukraine finding. The AI didn't just report that Ukraine has 11,970 refugees per 100,000 people. It framed it against Afghanistan — a country synonymous with refugee crisis in public perception — and showed that Ukraine now exceeds that rate. That reframing is an editorial choice, not a calculation. The AI made it.
Or take the seed story. Among 15 dataflows and hundreds of indicators, the AI surfaced a little-known indicator — Standard Material Transfer Agreements — and noticed that India receives more crop genetic material than any country except Germany. A human analyst would need to know to look for it. The AI found it by exploration.
A data insight is not a published card. Someone has to turn "India ranks 11th in worker productivity growth" into something a reader will stop and absorb. That means choosing the right chart type, placing text so it doesn't overflow, matching the masthead font exactly, and hitting a sub-second print deadline.
The human gave feedback once — on the very first card. The tagline was too small. The colour palette needed to be more print-friendly. Three sentences of feedback, and the AI updated the brand spec file and regenerated everything correctly. Cards 2 through 10 went out without another round of corrections.
Notice what wasn't specified: the chart types. The human never said "use a lollipop chart." The AI chose it for card 06 because a lollipop shows ranked data with large gaps between items more clearly than a bar chart — fewer ink-to-insight wasted pixels. It chose a waffle chart for card 10 because 195 small squares make the minority of 50 "achieved" countries feel like a minority in a way a percentage never can.
Card 06 · Lollipop — ranked seed-sharing data
Card 10 · Waffle — 195 unit squares by status
Card 07 · Proportional strip — part-of-whole
Card 09 · Funnel — rules vs. practice filter
There is an obvious problem with AI-generated data journalism: how does a human editor verify it? The same AI that wrote the insight also wrote the SOP — the Standard Operating Procedure that tells a journalist exactly how to check every number before it goes to print.
Each of the 10 cards has a matching verification guide. They follow the same seven-section structure: how to download the source data, what the card is claiming, the fastest way to verify it, which exact CSV rows contain the numbers, a card-text-to-source-value table, and common verification mistakes to avoid.
This is the key trust mechanism. AI-generated journalism that cannot be verified by a human is a liability. AI-generated journalism that comes with a complete verification trail is an asset. The SOPs mean that a journalist who has never seen this dataset can reproduce every number in under ten minutes.
Try it: open any card in the gallery, click "Verify," and follow the checklist. The numbers hold.
Plus: be resumable, log progress, use sub-directories. That's it.
Discovered SDMX API · 15 dataflows cataloged · Resume-safe · Parallel · Compressed output · Full README
Plus two short follow-up nudges over three rounds. Total human writing: ~120 words.
20+ output CSVs · Web research for unexplored angles · Editorial framing for each insight · notes.md of caveats and follow-ups
Gave feedback on card #1: tagline too small, use print colours. ~3 sentences total.
Chose chart types independently · Lollipop, waffle, funnel, proportional strip — each matched to the story's data shape · Rendered PNG for visual QA · Fixed own errors
Specified the checklist format. That's all.
Exact CSV row filters · Copy-paste search strings · Pre-publication checklists · Common pitfall warnings · Reproducible data pipeline instructions
There is a difference between a tool that helps a journalist work faster and a system that can run an entire pipeline from raw data to published card with minimal human input. This is the second thing.
The human in this workflow didn't know SDMX format. Didn't write SQL. Didn't design a chart. Didn't calculate a percentage. Didn't pixel-check a font size. What the human did was something harder to automate: decided what mattered. "Find the surprising angles. Write for lay readers. Make it verifiable." Those are editorial judgments, and they guided everything.
We tested exactly this. The same process produced 7 cards from India's PLFS survey and 5 cards from PLFS v4 before the UN work began. The prompts were nearly identical. The outputs were equally publication-ready.
What changes when this becomes routine? Data stories that currently take a week of journalist-plus-designer time now take an afternoon — and the human spend that afternoon on the editorial judgment, not the plumbing. The machine handles the plumbing.
It is not autonomous journalism. Every card in this gallery was published because a human editor decided it was ready. The verification SOPs exist precisely because the AI's output needs a human check before it goes to print. The editorial voice — the decision to compare Ukraine to Afghanistan, to focus on India's seed-diversity rank — came from a person with journalistic judgment.
What AI automates is the distance between "I have a dataset" and "I have something a journalist can read and verify." That distance used to take weeks. Now it takes hours.