Jannah Theme License is not validated, Go to the theme options page to validate the license, You need a single license for each domain name.
Tactical

Did Neil Ferguson’s Modelling of the COVID Pandemic Prove Once & For All That Epidemiology Is Useless?

This article was originally published by Rhoda Wilson at The Daily Exposé.

The UK and Imperial College London were the intellectual nerve center of the global covid pandemic response.  As you will recall, it was Neil Ferguson, a physicist at Imperial College, who developed the main epidemiology model behind the justification to shut down economies. 

However, Ferguson’s model was undocumented 13-year-old code that needed to be “cleaned up” and his assumptions were garbage.  This is been known since at least March 2020, around the same time that the UK government implemented its first shutdown of the economy spuriously called “lockdowns.”

The American Institute for Economic Research (“AEIR”) noted in April 2021, “Ferguson predicted catastrophic death tolls back on 16 March 2020 unless governments around the world adopted his preferred suite of non-pharmaceutical interventions (“NPIs”) to ward off the pandemic. Most countries followed his advice, particularly after the United Kingdom and United States governments explicitly invoked his report as a justification for lockdowns.”

Yet Ferguson’s model was written using some 13-year-old undocumented code which then needed to be “cleaned up” with the help of Microsoft to make it reusable by anyone else. That isn’t even the worst of it – key assumptions forming the inputs to the model were garbage, Jonathan Engler writes introducing an article written by Mike Hearn in March 2020.

By Jonathan Engler, 22 June 2024

As the years march on, it is easy to forget how egregiously wrong every single assumption around “the pandemic” was.

This piece (see below for link) from 31 Mar 2020 – which I highly recommend – lays bare the shortcomings of using GIGO1 modeling as a driver of public policy, something which we see across many domains, most notably the “climate” agenda but also many other fields.

In fact, I would characterize the age in which we live as one which is witnessing pseudoscientific modeling replacing empiricism.

Anyway, do enjoy. The commentary on Ferguson’s / Imperial’s Foot and Mouth disease modeling is particularly interesting. How wrong do you have to be before someone hesitates to trust you to determine policy (because that is what he did) which will cause death and misery for millions of people2?

It needs to be emphasized that the UK / Imperial was the intellectual nerve center of the global “pandemic response.”

Yet Ferguson’s model was written using some 13-year-old undocumented code which then needed to be “cleaned up” with the help of Microsoft to make it reusable by anyone else.

That isn’t even the worst of it – which is that the key assumptions forming the inputs to the model were garbage.

The summary of the piece is reproduced below. Click on it or HERE to go to the full article. [Note: We have reproduced the article below.]

Notes:

  • 1 Garbage in, garbage out.
  • 2 Because whilst many things weren’t known in spring 2020, what WAS predictable was that you cannot shut down the global economy without – over the medium and longer term – causing millions of deaths from economic deprivation. The immediate harms to health from the healthcare shutdown were also predicted by many.

By Mike Hearn, 31 March 2020

There’s a famous saying about simulations of the real world: “All models are wrong, but some are useful.” A critical question to ask right now is whether epidemiological models are wrong but useful, or just wrong.

Update 2nd April 2020

Here are links to other analyses that I found after the publication of this article

Are Epidemiological Models Useful?

It’s up to governments to decide what advice to follow. Nothing in this analysis should be taken as a suggestion to ignore their recommendations or laws. Don’t start ignoring the local rules because you read this blog post.

ICL’s modeling of SARS-CoV-2 spread is what triggered the UK’s switch to full lockdown mode and reinforced or also triggered it in many other countries. Given the dire consequences of shutting down the planet, it deserves a level of scrutiny no scientific work has ever been subjected to before. That’s now slowly starting to happen.

In this article, I’ll look primarily at the history of epidemiology and methodological problems found within. If you’re interested in problems that are really covid-19 specific, I recommend ‘Coronavirus Disease 2019: The Harms of exaggerated information and non-evidence-based Measures’ by Dr. John Ioannidis, accepted by the European Journal of Clinical Investigation.

Citations will be provided for all claims.

Summary

  • Imperial College London (“ICL”) is described as being the best in the world at epidemiological modeling.
  • Despite this, they have a history of major misses that don’t seem to be admitted. Their recommendations have led to disastrous overreaction by governments in the past.
  • Although presented as scientists they engage in unscientific practices e.g. making unfalsifiable statements, lack of peer review, refusing to show their code, and using vague terms instead of statistical confidence bounds.
  • They have predicted disease outbreaks to be orders of magnitude more serious than they really were.
  • It’s unclear what the field of epidemiology learned from these mistakes, if anything.
  • They are using data for their models that are known to not be statistically meaningful.
  • Sniping has broken out among epidemiologists at different universities, many of whom are contradicting each other in public, and in very major ways.

Whilst most of this article will focus on the track record of ICL the problem is really a wider one than that – ICL’s report is being listened to because of the “brand” of academia and academic science more generally. Oxford has produced its own paper which is no better and if anything, even worse. “We’re guided by the science” is the motto of governments everywhere because they assume the science is correct, or at least, better than nothing. If that assumption is false, it’s a huge problem.

Who Are the ICL Modelling Team?

So, they’re a pretty big deal. If they’re the best epidemiologists in the world it seems we might judge the field of epidemiology by their performance.

It’s worth noting at this point that epidemiology isn’t the same thing as medicine. Prof. Ferguson did his PhD in theoretical physics. Modelers can be computer programmers who specialize in applied mathematics, not doctors in the hospital sense. This isn’t meant as a knock: applied mathematics is a highly valuable field of course but as we’ll see later, a key criticism of epidemiology is the way in which it elevates abstract mathematical calculations above the experience of people with on-the-ground medical experience.

Unscientific Practices

The scientific method needs several things to work right.

Peer review. ’Impact of non-pharmaceutical interventions (NPIs) to reduce COVID-19 mortality and healthcare demand’ is dated 16 March 2020 and was released to the press immediately, simultaneous with the sudden change of government strategy it caused. Although many papers on COVID-19 have been put through peer review, in this case, it seems it was skipped.

It’s possible peer review would have rejected the paper; certainly, a lot of their peers have problems with it.

Reproducibility. The analysis can’t be replicated for multiple reasons. Amongst other problems the actual model itself isn’t available anywhere because the code is of such low quality only Prof. Ferguson’s team understands it:

He doesn’t plan to ever release the original code, only a version rewritten by Microsoft:

In maths class children who don’t show their workings get a “fail” grade, doubly so if they submit work done by others. In epidemiology, it’s no big deal.

This is an exceptionally critical problem. I cannot emphasize this enough. Academia is in the middle of the replication crisis (that’s the real name). Entire fields are having their credibility shredded because nobody can replicate ‘discoveries’ that were widely accepted for decades.

In the time since ICL released their study, other non-epidemiologists managed to build models that are not only publicly documented and with available source, but which are fully interactive and can be run by anyone in a browser. “My code is too complex to understand without personal training” is not an acceptable thing for publicly funded researchers to say, especially not when that code is now over a decade old. Imperial had all the time in the world to make their results reproducible and of acceptable quality, yet never did.

The analysis can’t be replicated for other reasons: it relies on private correspondence for key data (“personal communication” is listed as a source twice), makes vague reference to “the NHS providing increasing certainty around the limits of hospital surge capacity” but doesn’t say where this certainty was published – especially curious given that the analysis shows capacity as constant for a year but the NHS is building three new emergency hospitals, the first of which is one of the largest hospitals in the world. Where did this completely flat line at 8 ICU beds per 100,000 come from?

Non-reproducible work is being slowly stamped out in other fields like psychology: it needs to be unacceptable here too. Especially now!

Falsifiable predictions. Scientists make testable predictions.

Prof Ferguson has a habit of making “heads I win, tails I win” type predictions. To be fair, in his paper, various estimates were provided for death tolls given various combinations of reproduction values (R0), levels of lockdown, and trigger values. Those are fairly precise and in hindsight, we could measure how far away reality is from them. For instance, with an R0 of 2.2 and with a lockdown triggered at a rate of 300 ICU cases per week, they predict 26,000 deaths. Some people have claimed he later revised his prediction downward from over 500,000 deaths – but that claim isn’t true.

But there is a slightly more subtle problem. Later he did change what he was predicting to be “20,000 deaths and could be much lower.” If deaths are much higher, he can argue his recommendations weren’t followed closely enough – and as those recommendations are virtually impossible to implement fully, who can argue? If deaths are around 20,000, he can say “Our analysis correctly predicted the outcome.” If deaths are much lower, he can say “Deaths were within our predictions.”

This sort of problem has cropped up before. When asked to model the outbreak of bovine spongiform encephalopathy (a.k.a. mad cow disease) he predicted a human death toll of between 50 and 150,000.

When queried about the vagueness of this prediction the response was “Yes, the range is wide, but it didn’t actually lead to any change in government policy” (see Daily Telegraph). That answer is a non-sequitur but shows a deep concern with whether epidemiological advice steers outcomes.

A second critical change was the claim that “probably 2/3rds of those people would have died anyway.” The notion of excess mortality appears nowhere in the original report; most likely ICL found out the Italian data they used reported deaths with infection and not deaths because of infection at the same time as everyone else. It would presumably change the conclusions fundamentally. Indeed, it muddies the very concept of “number of deaths.”

In case you think I’m picking on ICL, the recent Oxford epidemiology paper on COVID-19 says “… the proportion of the UK population that has already been infected could be anywhere between 0.71% and 56% (95% credible intervals…).”

These sorts of ranges in predictions really mean epidemiology has nothing useful to contribute. However, they aren’t saying that in plain English.

Appearance of neutrality. Trust in science falls when people believe scientists are pushing political agendas. This largely explains why, as the Financial Times reports, ‘Economists among least trusted professionals in the UK’. It’s also a common concern floated by climatology skeptics.

An easy way to avoid this problem is for scientists to simply publish what they’ve found and leave discussion of policy changes to politicians who are – unlike academics – directly accountable to those who policy affects.

Epidemiologists don’t seem to do this. So far in every case I’ve examined epidemiologists recommend extremely specific social and agricultural policies, and some papers spend about half their word count directly addressing policymakers.

Foot and Mouth Disease

Let’s take a look at how some of these problems can lead to disaster.

Epidemiological modeling is a relatively young field. Its first test drive in the UK came with an epidemic in 2001 of foot-and-mouth disease (“FMD”) amongst pigs and sheep. The result was so catastrophic it has been the subject of many papers. Whilst this event is now 20 years old, I’ll also look at a second prediction from about 5 years ago to demonstrate that not much has changed.

Here are a few of the papers I read on the topic this weekend but there are many more. It’s obvious the events were highly traumatic and highly studied as a result. As you can guess from the titles, the authors were very critical of what happened:

  1. Destructive tension: mathematics versus experience ‘, referred to here as Mansley et al (authors all have veterinary backgrounds)
  2. Wrong, but Useful: Negotiating Uncertainty in Infectious Disease Modelling’, Christley et al. (varied backgrounds)
  3. Carnage by computer: The blackboard economics of the 2001 foot and mouth epidemic’, Campell & Lee (Cardiff Law School)

The government’s “FMD Science Group” consisted of a range of disciplines but epidemiologists were by all accounts dominant. Modellers came from four different universities, but fairly quickly the ICL model became the primary model used (from the paper ‘Media, metaphors, modelling’).

Their programs predicted a severe epidemic. They argued it could only be averted by an instant and extreme policy: the so-called “contiguous cull.” Any susceptible animal living within 3km of any farm that had an infected animal was to be killed immediately, even if it was healthy.

Virtually all the killed animals were uninfected:

Was it justified? Mansley et al say:

There was an accidental control group. Cumbrian authorities didn’t have the resources to implement the contiguous cull everywhere:

They failed the acid test of usefulness.

Why did they go wrong?

Along with bad assumptions, the model was fundamentally concerned with the geographical spread between farms but the data used about farms was of very low quality, having been originally collected for the management of CAP subsidies:

As Kitching et al. put it: “The UK experience provides a salutary warning of how models can be abused in the interest of scientific opportunism.”

Although Mansley et al is the most comprehensive, there are lots of papers making similar points.

Lessons Learned, or Not

In the wake of this event, ICL produced the following chart:

The chief scientific advisor at the time gave this testimony:

As far as the epidemiologists were concerned, it was a great success. What conclusion did the government ultimately reach? We know because some years later another outbreak occurred:

The lesson learned was to ignore epidemiology:

What happened?

The failure of epidemiology in this event was absolute: the only consumers of their product were politicians and civil servants. Those people chose not to use models in the next outbreak.

A key question is what the field learned from this. And that’s where it gets very troubling indeed. From The Telegraph: “Professor Ferguson said of his modeling for FMD: A number of factors going into deciding policy, of which science – particularly modeling – is only one. It is ludicrous to say now that our model changed government policy. A number of factors did’.”

The highlighted statement is problematic because it’s both very strong and just doesn’t match anything else written about the outbreak. It’s clear from every document on the FMD epidemic that epidemiological modeling was the primary driver of government policy. How can the professor claim it’s ludicrous to believe ICL’s work determined government policy when so many authors writing papers about that time believed otherwise?

From ‘Media, metaphors and modeling’: “As the modeling undertaken at Imperial College became the prime source for political decision-making, the press focused mainly on the models produced there, not the models produced by the Edinburgh and Cambridge teams.”

From ‘Destructive tension: mathematics vs experience’: “The models that supported the contiguous culling policy were severely flawed.”

As of today, the ICL team still believes they were basically correct about the FMD epidemic:

Zika: Another Miss

You might remember the 2015 Zika outbreak in Latin America, a terrifying virus that caused babies of infected women to be born with shrunken heads and serious brain damage. ICL modeled the epidemic and said in a July 2016 paper:

And here’s what happened. It didn’t take 3 years with seasonal returns. It was gone within one.

Zika is eradicated from the USA. Globally there have been so few cases since 2017 that Wikipedia’s page about the virus doesn’t even bother with any news about it after that (the last update is about two cases in Angola).

Whilst the disease still circulates in Latin America, the worst affected region making up 60% of all reports (Brazil) sees an average of about 365 cases per week, of which only about 30 are laboratory-confirmed – too low to see on the graph above. If the predicted seasonal oscillations exist at all they are lost in the noise.

The analysis was also unable to explain the behavior up to that point:

About half the paper was devoted to the section, ‘What should policy-makers do’ but no concrete recommendations were made beyond recommending women avoid getting pregnant. The paper does admit:  “Advising against pregnancy has been criticized for being infeasible for many women – especially long term”

No kidding.

Conclusions

Similar problems seem to crop up repeatedly in epidemiological analysis:

  • Models are based on input data of extremely low quality. This is acknowledged briefly but doesn’t stop anyone from making predictions even though it should.
  • Statistical uncertainty often isn’t formally analyzed. Vague adjectives like “largely,” “broadly,” “probably,” “substantially” and “typically” are used instead.
  • Modelers are highly involved in policy-making and clearly perceive that as one of their primary purposes. Advice to political leaders can make up half or more of supposedly scientific papers.
  • Predictions routinely have bounds so enormous they are rendered useless.
  • Modelers don’t seem to have made any obvious methodological changes in response to prior prediction failures

Should epidemiologists be treated with the nearly god-like respect they currently command?

I’ve got nothing against any of the people or institutions discussed in this article, and in principle, I don’t see why epidemics shouldn’t be simulatable. But it could be prudent to hold off talking to journalists and politicians until the field has had a string of undeniable successes and the outcomes have become seen as routine. Clearly, we’re not there yet.

About the Author

Mike Hearn is a former Google engineer, the original author of Bitcoinj and a former contributor to Bitcoin Core. He quit Bitcoin in January 2016; one reason given was rising fees (see The resolution of the Bitcoin experiment).  He publishes essays on the online publishing platform Medium under the title ‘Mike’s Blog’.

Featured image: Neil Ferguson taken from UK coronavirus adviser resigns after reports his lover visited during lockdown, CNN, 6 May 2020 (right).

Read the full article here

Back to top button