Skip to main content
Choose your location
Select the location and the language that you prefer
main content, press tab to continue
Article | Willis Research Network Newsletter

20 years after Hurricane Katrina, are insurers ready for a completely different kind of $100+ billion disaster?

By Jessica Boyd and Daniel Bannister | August 21, 2025

Hurricane Katrina reshaped disaster risk management, but 20 years later, we still plan for the last crisis, not the next one. Here are 5 reasons future $100+ billion catastrophes could be far more complex.
  1. On August 29, 2005, Hurricane Katrina made landfall near Buras-Triumph, Louisiana, as a Category 3 hurricane with sustained winds of 125 mph (200 km/h). It had previously reached Category 5 intensity over the Gulf of Mexico, peaking at 175 mph (280 km/h), before weakening prior to landfall. Katrina generated a storm surge over 20 feet (6 metres) high along parts of the Mississippi coast, overtopping and breaching levee systems in New Orleans.


  1. In total, there were over 1,800 fatalities, more than 200,000 homes destroyed, and around 1.2 million people displaced. Economic losses exceeded USD $250 billion (inflation adjusted to 2025 values), while insured losses reached around USD $100 billion (in 2025 terms), making Katrina the most expensive natural catastrophe on record.

    Twenty years on, Katrina still shapes how we think about catastrophe risk. But the next $100+ billion catastrophe won’t follow Katrina’s script. Risks have changed. So have our cities. Even the hazards themselves are changing. Rather than ask “are we ready for another Katrina?”, the question should be: “are we ready for a completely different kind of disaster?”

    Here are 5 key reasons why the next major catastrophe won’t look like Katrina, and how ready we really are, scored out of five.


  1. 01

    We missed clear warnings before Katrina and future ones may be harder to see

    The devastating flooding from Hurricane Katrina is often described as a surprise, but in hindsight it was alarmingly foreseeable. Just a year earlier, in 2004, Hurricane Ivan came close to a direct hit on New Orleans, prompting warnings of widespread flooding. But when Ivan veered north and made landfall in a less densely populated area, its potential catastrophic impacts were quickly forgotten. And just months earlier, Hurricane Pam — a FEMA simulation exercise — had already modelled widespread flooding from levee overtopping in New Orleans.[1] The warnings were there, in both models and nature. This oversight highlights a critical weakness in risk thinking at the time: failure to learn from counterfactuals. We tend to prepare for what has happened, not for what nearly did.


  1. So how can we avoid being blindsided by the next “predictable surprise”? We can use downward counterfactual analysis, the practice of examining past near-miss events and asking “what if this had been worse?”.[2] Unlike some traditional disaster scenarios, which can feel abstract and unlikely, these are rooted in real past events so are tangible and relatable. Though the concept is gaining traction, downward counterfactuals are not systematically used in exposure management. As a result, we continue to miss chances to prepare for disasters history has already warned us about. Yet even the past has limits. With a changing climate, the same catastrophe hitting the same place today could unfold in entirely different, and potentially more devastating, ways.

    Score: 2/5, inadequately prepared

    True proactive planning for “predictable surprises” is still more the exception than the rule.


  2. 02

    We manage exposure hotspots better now, but new ones are emerging

    Katrina revealed how insurers had concentrated exposure in high-risk areas, leaving them vulnerable to flooding from a single event. In response, (re)insurers began monitoring and capping their exposure in vulnerable zones. Portfolios were diversified geographically to reduce the risk of excessive loss in any one area. In essence, the practice of exposure management was born.


  1. 03

    Levee failures were the known unknown, what’s the next infrastructure domino?

    One of Katrina’s harshest lessons was that infrastructure we take for granted can catastrophically fail. The New Orleans levee breaches — unmodelled by insurers at the time — exposed this blind spot. In the typical (re)insurance reactive style, many models now include the failure probabilities for New Orleans’ levees, but not for other, less extensive yet still vulnerable defences.


  1. The challenge is anticipating critical infrastructure failures that haven’t yet occurred. One of the biggest barriers is the failure of imagination, our tendency to overlook risks we haven’t experienced before. Human cognitive biases, such as availability bias, cause us to overfocus on memorable past events and ignore less obvious possibilities, narrowing our view of what’s plausible.

    To address this, we can use tools that stretch our imagination. Large language models (LLMs), a branch of generative AI, have the potential to rapidly generate diverse, novel catastrophe scenarios that a human (or group of humans) may miss. While many insurers already include scenario analysis to help plan for unforeseen circumstances, current exposure management processes don’t yet systematically include these highly imaginative what-ifs. This must be done with care — LLMs can misinterpret context or generate inaccurate details — but with expert review, they can expand, not replace, our ability to imagine what could go wrong.

    Score: 1/5, severely underprepared

    Where Katrina exposed a physical weakness in a known location, the next failure could be digital, decentralised, and invisible until it happens. We don’t yet use the tools available to expand our imagination to help us plan for unfamiliar future events.


  2. 04

    Post-event loss amplification could be bigger and messier

    When infrastructure fails, the consequences go far beyond physical damage. Katrina showed how cascading effects — costly repairs, delayed claims settlement, and a paralysed local economy — can drive insured losses far beyond expectations. At the time, catastrophe models couldn’t account for such post-disaster complexities. Today’s models apply adjustments for things like demand surge and claims inflation, but those adjustments remain relatively unsophisticated. Also post-event loss amplification (PLA) factors are still largely based on historical averages and expert judgement.

    What’s different now is the scale and interconnectedness of exposure. Major cities are more densely built, more reliant on just-in-time supply chains, and more digitally dependent. A future disaster could disrupt essential services — power, water, internet — across multiple cities or regions simultaneously, triggering ripple effects in housing, healthcare, finance, and beyond. Add new vulnerabilities like cloud data storage or gig-economy labour, and the system-wide impacts become even harder to model or contain.


  1. 05

    Coverage disputes could escalate faster and cost more

    Hurricane Katrina sparked public outrage over insurance coverage gaps, especially the distinction between wind or flood (storm surge) damage. These conflicts triggered protracted court cases that proved costly and reputationally damaging for insurers, who subsequently tightened up policy wordings relating to peril type coverage.


Authors


Jessica Boyd, Head of Model Research
Head of Model Research, Willis Research Network
Email

Daniel Bannister - Weather & Climate Risks Research Lead Willis Research Network
Weather & Climate Risks Research Lead
Email

Contact us