Do we need to run a trial for everything?
This is an excerpt from my healthcare + comedy newsletter Out of Pocket. You can sign up for it here to get weekly analysis. You can read the full version of this post here.
Spurred by a spat between Keith Rabois and Zach Weinberg, last week I talked about randomized controlled trials and their shortcomings. Congrats if you read the whole thing, I didn’t even read the whole thing.
It was meant to give a basic understanding of why randomized controlled trials are important. Now I can get to the core of their actual debate: when is observational data (aka. real-world data) useful?
*As a quick terminology point, real-world data is the data itself while real-world evidence is the usage of that data to actually prove something. I’m going to use real-world evidence throughout the entire post to make it easier.
Observe
Should we be running a randomized controlled trial on every single thing we do?
Nah. We know for a fact that some things have direct causation, we don’t need a trial to prove it. We know if it’s good or not based on observation. Actually this is how medicine first started - people would notice things that worked, try them out, and then those recommendations would spread.
But at what point do you KNOW something is good or bad? When two things seem related, we all know to appropriately and smugly say “correlation does not imply causation” while basking in our clear intellectual superiority. But there needs to be a point at which we DO infer causality, right?
Keith is right in that there was no randomized controlled trial done to link smoking and lung cancer. After a few papers showing the harmful effects of tobacco in animals, a 1954 paper by Sir Richard Doll and Sir Bradford Hill showed such a strong correlation between smoking and lung cancer in humans via observational data that the scientific community agreed that causality could be inferred. Several other epidemiology studies came out around the same time with similar results.
The author of the paper, Sir Bradford Hill, established 9 principles to figure out when correlation should imply causation if you’re looking at the relationship between some intervention/environmental factor and a disease:
Strength Of Association - Pretty self-explanatory. Is the strength of association between the two strong or weak?
Consistency - Is the association between the two factors found consistently?
Specificity - Is there a 1:1 relationship between exposure to X that causes Y disease. Today we know that exposure to different factors can cause a multitude of different diseases, but at the time it was believed the relationship was more specific.
Temporality - Did the exposure happen before the disease manifested? (bonus points if the disease manifests in a consistent amount of time post-exposure)
Biological Gradient - Does more exposure yield more severe versions of the disease?
Plausibility - Can the relationship between exposure and disease be explained? Is the mechanism actually understood? Even Sir Bradford Hill said this probably wouldn’t be possible in most cases.
Coherence - Does that mechanism/reason for association defy what we know previously about how this disease usually manifests?
Experiment - Do we have any other experiments we can point to (e.g. animal models).
Analogy - Are there other exposure/disease relationships we can point to that are similar?
These have been debated over time, but a useful framework. If we do not have enough evidence via these criteria, a randomized controlled trial would help establish a more clear relationship.
Also the kicker? Sir Bradford Hill ALSO pioneered the randomized clinical trial! Find someone that looks at you the way he looked at experiment design.
Establishing the link between smoking and cancer used observational data instead of a randomized controlled trial to make this conclusion. So while Keith is correct that a randomized controlled trial was never performed, that exact situation is actually why we had to set up frameworks to understand when observational data is actually causal. And it turns out the bar to make that claim is quite high.
You can see the benefits and obstacles of real-world evidence in the full post, including who won the debate. For weekly healthcare analysis, sign up for Out Of Pocket.
Value-based Care Operator
4 年Love the humor! If only statistics were normally taught with rap and pop culture references, more people might like it, let alone understand it.
Data and Technology Leader | Advisor | Investor
4 年Check out Judea Pearl’s work on casual inference. Mostly academic but also some (reasonably) readable popular science books too.