Dataset

The State of Humanitarian Data

2026-03-20 82 views 39 min read
68%
Crisis Data Up-to-Date (2025)
11/20
Sub-categories Declined YoY
~40%
Data Staff Cut at UNHCR & IOM
3.4M
HDX Dataset Downloads in 2025
Sector Alert Two simultaneous famines were declared in 2025 — in Sudan and the Gaza Strip — the first time in the IPC's 20-year history this has occurred. The data infrastructure that tracks such crises is simultaneously being dismantled by unprecedented funding cuts, with UNHCR and IOM each cutting data-related staff by roughly 40%.
Sources & Methodology: This analysis draws on three primary sources: the OCHA State of Open Humanitarian Data 2026 report (OCHA Centre for Humanitarian Data, March 2026); the video transcript of Safeguarding Humanitarian Memory: Challenges and a Practical Toolkit for Digital Archives (Flora Chatt, University of Manchester Humanitarian Archive, 2025); and the Digital Humanitarian Archives Toolkit (beta), developed by the Humanitarian and Conflict Response Institute, University of Manchester, under the Humanitarian Archive Emergency (HAE) project (Elrha/HIF, Nov 2025–Oct 2026). Quantitative figures are drawn directly from HDX Data Grids covering 22 HNRP locations as of 31 December 2025.

A System Under Stress: Data, Archives, and the Cost of Cutting Memory

Humanitarian data and institutional memory face a compounding crisis in 2025–2026. On one front, the OCHA Centre for Humanitarian Data reports that the share of crisis data that is available and up-to-date has fallen from 74% to 68% across 22 active humanitarian operations — a decline driven overwhelmingly by the largest wave of international funding cuts in recent memory. On another, an entire ecosystem of organisational archives — the documents, field records, WhatsApp threads, and emails that constitute the sector's institutional memory — is deteriorating silently, with no adequate preservation infrastructure for the vast majority of humanitarian organisations.

These two crises are more connected than they appear. The funding cuts that are shrinking real-time data coverage are the same cuts disbanding the enumerator networks, forcing office closures, and eliminating the staff whose knowledge and records will never be formally archived. What is lost is not merely a data point but a body of institutional learning accumulated over years of emergency response. The sector's ability to understand the present is weakening at the same moment its ability to learn from the past is disappearing.

Against this backdrop, two parallel initiatives offer targeted responses. The OCHA Centre for Humanitarian Data continues to manage HDX — a platform used in 237 countries and territories whose nearly 20,000 datasets were downloaded 3.4 million times in 2025. At the University of Manchester, the Humanitarian Archive Emergency (HAE) project and its accompanying Digital Humanitarian Archives Toolkit represent a practical attempt to democratise archiving for resource-poor organisations. Together, they sketch the outlines of what a more resilient humanitarian knowledge infrastructure could look like — if the sector finds the will and resources to build it.


The State of Open Crisis Data: A Fragile 68%

The HDX Data Grids — OCHA's standardised framework for assessing data across 22 active humanitarian operations — provide the clearest snapshot of the sector's data health. As of 31 December 2025, 68% of crisis data is available and up-to-date, down from 74% the previous year. That six-point decline, achieved despite extraordinary mitigation efforts, signals a structural weakening rather than a temporary fluctuation.

The geography of data quality is deeply uneven. Mozambique leads at 90% of sub-categories available and current, followed by Afghanistan, Cameroon, and South Sudan at 85%. At the bottom sit Venezuela (47%), the Syrian Arab Republic (50%), and Burkina Faso (50%) — precisely the contexts where decision-makers most need reliable information. The State of Palestine, where the IPC declared famine conditions in 2025, has only 53% of data available and current.

Key Finding Eleven of the twenty HDX sub-categories saw year-on-year declines in up-to-date data availability. The most dramatic: baseline population data fell 45% — a direct consequence of UNFPA's rapid capacity reduction — and administrative boundary data fell 28% after third-party support was abruptly withdrawn in early 2025.
Dashboard 1
HDX Data Grid: Availability by Location & Category
Data availability as of 31 December 2025 across 22 HNRP humanitarian operations
Status: Available & Up-to-Date Available, Not Current Unavailable
90%
Highest — Mozambique
47%
Lowest — Venezuela
68%
Global Average
11%
Completely Unavailable

Data Availability by Location — Available & Up-to-Date (%)

Two sub-categories registered genuine improvements. Humanitarian needs data rose 30%, driven by the Humanitarian Reset's focus on streamlining planning processes and bringing data online earlier. Food security improved 14% after the Cadre Harmonisé — delayed in 2024 — was published on time in 2025. These positive cases demonstrate that process reforms, not just funding, can shift data quality. They are also important proof of concept as the sector prepares for deeper cuts ahead.


The Human Cost of the Data Drought

Behind every declining percentage point in the HDX Data Grids is a concrete operational decision: a cancelled contract, a closed office, a disbanding enumerator network. The scale of these reductions in 2025–2026 has no recent precedent. UNHCR and IOM have each cut data-related staff by approximately 40%. OCHA has reduced information management capacity in field offices by more than 15%. WFP enumerator visits — the in-person verification that underpins food price and market monitoring — dropped by roughly one-third.

👥
Data Staff Reductions by Organisation
Percentage reduction in data/IM-related staff, 2025–2026
Sources: OCHA Centre for Humanitarian Data 2026 Report; org-level reporting
📉
Operational Data Collection Reductions
Reduction in data collection activities, 2024 → 2025
Sources: WFP, REACH, IOM operational reporting; OCHA 2026

The cascading effects extend well beyond the humanitarian sector. UNHCR's statistical work — which provides socioeconomic data on refugees to governments and development actors — is under pressure. Without adequate resources, government-led solutions that depend on this data may be delayed or misdirected. In Sudan, IOM was forced to cancel contracts for more than 100 enumerators; displacement data collection was disrupted entirely until emergency funding was secured. In Somalia, IOM projects data collection will shrink from nationwide coverage to just 4–5 locations in 2026.

Forward Risk Networks of enumerators and key informants built over many years have been disbanded. Unlike funding, which can be restored, these human networks — built on trust, local knowledge, and institutional relationships — cannot be rapidly reconstructed once dismantled. The sector faces asymmetric risk: cuts are fast, recovery is slow.
OrganisationImpact TypeScaleTimingDownstream Effect
UNHCRData staff reduction~40%Early 2026Refugee registration slowed; statistical work at risk
IOMData staff reduction~40%Early 2026Somalia: nationwide → 4–5 sites; DRC, Ethiopia, Mozambique similar
OCHAIM field capacity>15%Early 2026Reduced field presence; partially offset by standby partners
WFPEnumerator visits~33%2025Food price monitoring maintained in volume but in-person validation cut
UNFPAPopulation data processingMajor2025Baseline population data availability fell 45%
REACHMSNAs completed16 → 142025~50% of remaining MSNAs covered only partial country
WFPRegional office closureFull closure2025Southern Africa consolidated; reduced anticipatory capacity for El Niño shocks
IOM SudanEnumerator contracts100+ contracts2025Displacement data disrupted; MSNA undermined until emergency funds secured

The Bot War: Open Data Platforms Under Siege

While funding cuts constrain the supply of humanitarian data, a new threat has emerged on the demand side: AI bot crawlers deployed by major technology companies to train large language models. In 2025, AI-bot and crawler traffic on the global internet almost quadrupled, climbing from 2.6% of verified bot requests in January to more than 10.1% by the end of Q3. HDX, as a curated open-data platform with structured, high-quality humanitarian datasets, became a prime target.

The platform absorbed periodic server demand surges reaching 20–30 times baseline levels. For the first time, civil society and nonprofits surpassed financial institutions as the most targeted sector for cyberattacks. The very features that make HDX valuable to humanitarian users — open access, no login requirements, rapid download options — also make it maximally attractive to automated scrapers. This creates a structural paradox: openness enables impact but also enables exploitation.

🤖
Top Referral Traffic to HDX (2025)
Referring domains driving unique pageviews, January–December 2025
Source: OCHA Centre for Humanitarian Data 2026 Report
📊
HDX Platform Scale (2025)
Key platform metrics for the year
Source: OCHA Centre for Humanitarian Data 2026 Report
Signal & Opportunity ChatGPT.com was the second largest referral source driving traffic to HDX in 2025, with 47,000 unique pageviews — behind only google.com (669,000). AI platforms are already becoming primary gateways to humanitarian data. Ensuring that AI systems relay this data accurately, and cite HDX as a trusted source, is now a core strategic priority for the sector.

HDX's response — web application firewalls, IP blocking, bot-control rules, geolocation filtering — consumed significant resources. Managing bot identification has become, in the words of the OCHA report, "more of an art than a science." The platform now categorises traffic into three tiers: wanted bots (search crawlers, AI indexers), unwanted bots (scrapers, security probes), and a "grey zone" of unidentifiable traffic. Navigating this complexity will require a strategic reallocation of resources toward infrastructure and specialised skillsets.


The Archive Emergency: What Isn't Preserved Is Lost

Beneath the headline data figures lies a deeper and less visible problem: the systematic failure to preserve the records, field notes, email chains, and operational documents that constitute humanitarian institutional memory. While organisations like UNHCR, ICRC, and IFRC have dedicated archive teams, and others like Oxfam and Save the Children have deposited archives at major universities, the vast majority of humanitarian organisations — particularly smaller national and local NGOs — have no functional archiving practice whatsoever.

Flora Chatt, archivist at the University of Manchester's Humanitarian Archive (founded 2021, currently holding 15 donated collections), frames this as a structural problem compounded by the nature of humanitarian work itself. Fast-paced emergency settings, frequent staff turnover, poor infrastructure for recordkeeping, and the move to digital communications — which require active, ongoing maintenance unlike paper — have created a perfect storm for institutional memory loss.

🏛️ Tier 1 Large orgs with dedicated archive teams (UNHCR, ICRC, IFRC)
🏫 Tier 2 Large orgs deposited at universities (Oxfam, Save the Children)
⚠️ Tiers 3–4 Mid/small orgs — unmanaged archives; serious or critical risk

"Archives need long-term resource and funding. They need a secure place to be kept. They need someone to manage them — to make sure that the sensitive information within them is kept safe. And this work needs to be continued for as long as archives are meant to last, which is permanently."

— Flora Chatt, Archivist, Humanitarian Archive, University of Manchester (2025)

The shift to digital record-keeping has paradoxically made the problem harder. A box of paper documents, left in a room for decades, remains largely readable. Digital files left for years face a cascade of risks: outdated hardware, obsolete file formats, silent corruption through "bit rot." The technical requirements for digital preservation — format migration, checksum verification, regular monitoring — demand more sustained resource and expertise than physical archiving, at precisely the moment when organisational budgets are being cut.

📁
Risk Factors Threatening Humanitarian Archives — by Theme Cluster
Count of identified risk factors per category (qualitative coding from Chatt 2025 & HAE project)
Source: Transcript — Safeguarding Humanitarian Memory (Chatt, 2025); HAE Project documentation
Risk FactorClusterSeverityKey Evidence
Unprecedented international funding cutsFinancialCriticalHAE project launched specifically in response to funding crisis
Staff reductions disbanding data networksOperationalCriticalEnumerator networks built over years; cannot be quickly rebuilt
No organisational will to manage archivesGovernanceHighMany orgs have archives but lack awareness or motivation to manage them
Digital file format obsolescenceTechnicalHighFiles become unreadable without legacy software within years
WhatsApp/email as primary recordGovernanceHighDecisions made and stored in ephemeral, unpreserved channels
Equity gap: small vs large organisationsEquityHighLarge orgs can advocate for archives; small orgs cannot
Geographic coverage shrinkingGeographicHighNon-HNRP countries at risk of going unmonitored in 2026
Bit rot / silent file corruptionTechnicalMediumBinary data flips silently; files corrupt without visible warning
Selective preservation biasGovernanceMediumLower-level operational records rarely preserved; skews the historical record

The Humanitarian Archive Emergency Project and its Toolkit

The Humanitarian Archive Emergency (HAE) project, funded by Elrha's Humanitarian Innovation Fund and running from November 2025 to October 2026, represents the sector's most structured response to the archive crisis. Based at the University of Manchester's Humanitarian and Conflict Response Institute, the project aims to establish a global census of archives at risk, develop reporting mechanisms, create ethical triage frameworks, and shape a medium-term strategy for sustainable rescue. It operates on a one-year timeline — modest for the scale of the problem, but designed to catalyse longer-term action.

The project's most immediate practical output is the Digital Humanitarian Archives Toolkit (beta), an e-learning resource designed to be accessible to any staff member at any humanitarian organisation — regardless of prior archiving knowledge. Its architecture reflects a pragmatic philosophy: the goal is not best practice but good enough practice, scaled to organisations that have almost no dedicated capacity.

Dashboard 2
Digital Humanitarian Archives Toolkit — Structure & Process
Six-module e-learning framework for building sustainable humanitarian archives
6
Total Modules
5
Core (Essential) Steps
1
Optional Step
0
Prior Knowledge Required

The 6-Step Archiving Process

STEP 1 — ADVOCACY
Get organisational buy-in. Build the case for archives internally, identify allies, secure time and resource. Without this, all other steps fail. Includes an Outcomes Form to track progress.
STEP 2 — AUDIT
Map what you have. Create a Digital Asset Register — a structured inventory of files, storage locations (including email inboxes, WhatsApp, portable drives), risks, sensitive data, and value assessment. Template provided.
STEP 3 — ORGANISATION
Impose order. Implement consistent file naming conventions, logical folder structures, remove duplicates. Foundational step before any preservation work begins.
STEP 4 — STORAGE
House files safely. Choose appropriate long-term storage solutions — cloud, local, or hybrid. Consider redundancy, security, accessibility, and cost. Guidance on options provided.
STEP 5 — PRESERVATION
Active ongoing maintenance. Monitor file integrity, migrate formats before they become obsolete, verify checksums, maintain metadata. The most technically demanding step; a checklist tool is provided.
STEP 6 — ACCESS (OPTIONAL)
Make archives usable. Sensitivity review, opening to researchers or internal users, discoverability. Optional but transforms preserved records from a storage exercise into an institutional asset.

Where the Crises Converge: Five Structural Themes

The data availability crisis documented by OCHA and the archive emergency described by the HAE project are not separate phenomena. They are two manifestations of the same underlying structural failure: the humanitarian sector has never treated knowledge infrastructure — data systems, preservation practices, institutional memory — as a core operational asset deserving sustained investment. The following themes recur across all three source documents.

🔗
Cross-Source Theme Convergence
Strength of convergence across OCHA Report, Transcript (Chatt 2025), and DHA Toolkit for 10 key themes
Sources: All three primary documents; convergence scored qualitatively as Strong (3), Medium (2), Weak (1)

1. Funding Cuts as Existential Threat

All three sources converge on this theme with maximum strength. The OCHA report quantifies it: 68% data availability, 40% staff cuts. The Chatt transcript contextualises it: the DHA Toolkit was "particularly kind of accelerated as well over the past few years" by budget cuts, and "this has become particularly kind of acute very recently in times of severe budget cuts as we are encountering now." The toolkit's design philosophy directly addresses it: built for organisations with "very few resources." The implication is clear — the sector cannot technical-solution its way out of a funding problem.

2. The Asymmetric Recovery Problem

Funding can, in principle, be restored. But much of what is lost in these cuts cannot be quickly rebuilt. Enumerator networks built over years of local relationship-building, once disbanded, require years to reconstruct. Archives deleted or left to decay cannot be recovered. Institutional knowledge held by departing staff dissipates. The OCHA report notes this explicitly regarding staffing cuts; the Chatt transcript makes the same point about archives. The humanitarian sector faces a structurally asymmetric risk profile: cuts are fast and cheap, recovery is slow and expensive.

3. The Equity Gap in Knowledge Infrastructure

Both the data availability data and the archive landscape reveal a sharp divide between large, well-funded organisations and the majority of smaller actors. Large organisations can absorb data staff reductions while maintaining some coverage; small organisations cannot. Large organisations have dedicated archivists and can advocate for their archives' importance; small and local NGOs have neither. The Chatt transcript puts this starkly: "the larger organisations can not only afford to maintain their archives but they can afford to kind of advocate for it." This structural inequality means that the historical record of humanitarian work is disproportionately shaped by large, international actors — a distortion with real implications for accountability and learning.

4. Digital Technology as Both Solution and Problem

The shift to digital work is simultaneously what makes data sharing at scale possible (HDX, Data Grids, the toolkit itself) and what makes preservation dramatically harder. Digital files require active, resource-intensive maintenance. AI bots are both a threat to platform stability and a potentially powerful channel for disseminating humanitarian data. The sector's digital infrastructure strategy has consistently underinvested in the maintenance and preservation dimension — optimising for creation and sharing while neglecting the longer-term archival challenge.

5. Data as a Public Good Requiring Collective Governance

Perhaps the most consequential theme across all three sources is the call to treat humanitarian data and memory as a public good — collectively financed, governed by shared standards, and maintained for the benefit of the entire sector and the communities it serves. The OCHA report makes this explicit in its calls for collective financing and the UN80 Initiative. The HAE project embeds it in its use of the Internet Archive as a public repository. The toolkit positions itself as a freely available common resource. The challenge is translating this principle into durable institutional architecture.


Climate Data: A Model for Cross-Sector Resilience

Against the prevailing narrative of decline, climate hazard data availability remained stable year-on-year — one of only a handful of sub-categories to hold its ground in 2025. The reason is instructive: climate data is provided not primarily by humanitarian organisations, but by scientific and research institutions — the EU Joint Research Centre, JANUS Atmospheric and Environmental Research, and the UC Santa Barbara Climate Hazards Center — whose funding and mandate lie outside the humanitarian system's current crisis.

🌍
Climate Data Providers on HDX — Key Dataset Characteristics
Update frequency and historical coverage depth for three core climate datasets
Source: OCHA State of Open Humanitarian Data 2026, Section 9

The three anchor datasets — ASAP (agricultural anomaly hotspots), FloodScan (near-real-time flood mapping), and CHIRPS (rainfall anomalies) — collectively cover the earth's most food-insecure regions with update frequencies ranging from daily to every five days, and historical baselines stretching back to the 1980s and 1990s. WFP processes CHIRPS data for 129 countries. An NGO in East Africa, Sauti, turns this data into simple SMS messages that help farmers and traders plan in real time.

Structural Lesson The stability of climate data availability in 2025 is a direct consequence of building partnerships beyond the humanitarian sector. Research institutions and private-sector data providers operate on different funding cycles and governance structures. Replicating this model — anchoring critical humanitarian data in cross-sector partnerships — offers one of the clearest paths to resilience against future funding shocks.

Key Events in the Humanitarian Knowledge Crisis (2022–2026)

2021
University of Manchester Humanitarian Archive founded. Begins collecting personal and small organisational archives related to humanitarian work; grows to 15 collections by 2025.
2022
HDX introduces climate data sub-category in Data Grids, beginning cross-sector data partnerships with scientific institutions.
2024
Crisis funding cuts announced. OCHA's prior-year report flags risks to data availability for 2025 based on funding reductions already underway. Cadre Harmonisé for West Africa delayed, impacting food security data.
EARLY 2025
Third-party support for administrative divisions data withdrawn. 28% year-on-year decline results; replacement processes built by already-stretched teams. IOM Sudan forced to cancel 100+ enumerator contracts; displacement data collection disrupted.
2025
IPC declares simultaneous famines in Sudan and Gaza Strip — first time in the IPC's 20+ year history. AI-bot traffic to HDX almost quadruples; server demand surges 20–30× baseline.
MID-2025
WFP announces closure of Southern Africa regional office. Multiple organisations freeze vacancies and postpone lay-offs; project-based activities cut immediately.
NOV 2025
HAE project launches at University of Manchester (Elrha/HIF funding). Digital Humanitarian Archives Toolkit released in beta. HAE Steering Committee holds first in-person meeting.
EARLY 2026
Staff reductions actioned at UNHCR and IOM: ~40% cuts in data-related staff at both organisations. OCHA field IM capacity reduced by >15%. HDX Data Grid availability confirmed at 68% — down from 74%.
MAR 2026
OCHA publishes State of Open Humanitarian Data 2026. Partners widely anticipate further deterioration in data coverage through 2026 as organisations adapt to reduced scale. UN80 Initiative and Humanitarian Reset identified as reform opportunities.

What a Resilient Humanitarian Knowledge System Requires

The convergent analysis across all three sources points to eight actionable priorities. These range from immediate emergency interventions to longer-term structural reforms. Absent political will and collective financing, even the most technically sound toolkit or data platform cannot substitute for the institutional commitment that has been lacking.

01
Collective Financing for Core Data Services
Establish pooled, multi-donor financing mechanisms for the common data services the sector relies on — HDX infrastructure, Data Grid curation, baseline population data, administrative divisions — so that no single funder's withdrawal can cascade into system-wide failure.
02
Integrate Archives into the Humanitarian Reset
The UN80 Initiative and Humanitarian Reset must explicitly include humanitarian archives and records management. The DHA Toolkit and HAE project should be positioned as scalable infrastructure for sector-wide memory preservation, not marginal library projects.
03
Protect Enumerator Networks Before Cutting Them
Establish international norms that classify long-standing enumerator and key informant networks as critical data infrastructure. Before any network is disbanded, a transition plan for continuity or handover to local actors should be required.
04
Replicate the Cross-Sector Climate Data Model
The resilience of climate data availability in 2025 — anchored in research-institution partnerships — demonstrates the value of diversifying data provision beyond the humanitarian system. Apply this model to other critical sub-categories: displacement, population, food prices.
05
Mandate Digital Preservation Standards for Humanitarian Data
Donors and coordinating bodies should require, as a condition of funding, that organisations implement basic digital preservation practices for the records they generate. The DHA Toolkit provides a ready-made, low-resource framework for compliance.
06
Develop a Shared Archiving Infrastructure for Smaller Organisations
No shared humanitarian archive exists for the sector as a whole. The Internet Archive Humanitarian Collection (HAE project) is a beginning. Build a co-hosted, ethically governed repository where smaller NGOs can deposit materials they cannot preserve independently.
07
Establish an AI Data Governance Framework for HDX
As AI platforms become primary gateways to humanitarian data — ChatGPT already the second largest referrer to HDX — the sector needs proactive governance: licensing frameworks, accuracy verification, and engagement with AI companies to ensure data is cited and relayed correctly.
08
Invest in Non-HNRP Geographic Coverage Before Crisis Strikes
The Humanitarian Reset's focus on active crises risks creating blind spots in regions currently below emergency threshold. Preemptive investment in data infrastructure for non-HNRP countries — modelled on the early-warning function of climate data — is essential for anticipatory action.

Memory as Infrastructure: The Stakes of Getting This Right

Humanitarian operations have always faced the tension between the urgency of the present and the requirements of the future. The current crisis — in which funding cuts are simultaneously degrading real-time data coverage and destroying the institutional memory that would allow the sector to learn from what it has done — makes that tension acute. A sector that cannot track what is happening now and cannot remember what it did before is flying blind in both directions.

The responses documented here — OCHA's sustained management of HDX, the Humanitarian Archive Emergency project's practical toolkit, the cross-sector climate data partnerships — are serious and well-designed. But they are operating at the margins of a structural problem that requires a structural response. The Humanitarian Reset and the UN80 Initiative offer a rare moment of system-wide reflection. The question is whether knowledge infrastructure — data systems, archive preservation, institutional memory — will be treated as the core operational asset it is, or whether it will again be the first line item cut when budgets tighten.

"Data has a central role to play in a revitalized humanitarian system — one that delivers more effectively and builds trust in multilateral action. It is a core asset that should be managed, with appropriate safeguards, as a public good in service of saving lives."

— OCHA Centre for Humanitarian Data, State of Open Humanitarian Data 2026
humanitarian data archieve OCHA UNHCR IOM crisis knowledge
Downloads & Resources
humanitarian_data_archives_dataset.xlsx
29 KB
Download