@davidgerard , not sure if you saw this great post about ML models in radiology?

The boosters love to say ‘look how good AI is in medical diagnosis, therefore LLMs are good’. Only, it turns out (from the article):

while models beat humans on benchmarks, the standardized tests designed to measure AI performance, they struggle to replicate this performance in hospital conditions. Most tools can only diagnose abnormalities that are common in training data, and models often don’t work as well outside of their test conditions

It also highlights a problem that’s actually quite general in medicine: we have far more data about unhealthy people than healthy ones. I was talking to a cardiologist almost ten years ago who was very excited about the data things like the Apple Watch could collect. Apparently they know that a lot of people who have heart attack have arrhythmia, but they have no idea if this is a meaningful correlation. Healthy people tend to have their heart monitored for a minute or less on a visit to a doctor every few years. People with known heart problems wear heart monitors that can record a load of things, so you have very good data on their heart rhythms but no baseline to compare it against.

This is also true for radiology. You really want to do anomaly detection: take a few million scans of healthy people, wait a few years to see if any of them have undiagnosed conditions, and then use that dataset to train a model of what a healthy lung (or whatever) looks like. Then feed new scans to the model, have it flag anomalies, and loop in an expert to figure out what kind of anomaly it is and whether it’s important.

But what you have is a load of very examples of things that are wrong, in very specific ways. And these also have artefacts that are specific to individual devices, so it’s easy for a model to learn that people who are scanned with this class of machine have this condition.

And that’s just the start of the issues they discuss.

AI isn't replacing radiologists

Radiology combines digital images, clear benchmarks, and repeatable tasks. But demand for human radiologists is ay an all-time high.

The Works in Progress Newsletter

@david_chisnall @davidgerard
One of the frustrations with the ML boom is that 'Data Scientists' seemingly have little understanding of basic statistical concepts.

Machine Learning is just computational statistics. The same laws about sampling, bias, reliability, etc apply. But very few data scientists seem to have much understanding of them (judging by the work I've seen).

@cian @david_chisnall @davidgerard
Rather than a lack of understanding, I suspect it is more the result of competitive pressures - i.e. people who publish flawed results will get ahead of people who do not.

There could be 90% of data scientists who understand statistics, the remaining 10% will be overwhelmingly represented in the set of “successful” researchers.

@jhominal @cian @david_chisnall the other problem is that the discussion happens in preprints and blog posts - and a lot of that is trash and marketing.

@cian @david_chisnall @davidgerard

IME, most(*) of the people working the coal face understand the limitations of what they've built - The information never seems to make it to the public or the sales people.

(* - CS in particular seems to have an unnaturally high percentage of people who refuse to admit they're ever wrong - even when confronted with evidence. Adjust accordingly.)

@cian my impression of “Data scientists” has so far been that they have as much in common with science as flat earthers.

@david_chisnall @davidgerard