Production AI systems are really hard
No, AGI isn't going to take over every social system when GPT5 comes out
NOTE: This post is not up to my normal writing standard, just felt compelled to get this down in some form. This is more like a blog post entry than a real newsletter addition.
I don't talk much about this - I obtained one of the first FDA approvals in ML + radiology and it informs much of how I think about AI systems and their impact on the world. If you're a pure technologist, you should read the following:
Geoffrey Hinton was one of the loudest voices decrying the decline of radiology 5 years, and now he’s crying fear for new AI systems.
There's a lot to unpack for both why Geof was wrong, and why his future predictions should not be taken seriously either. Geoff made a classic error that technologists often make, which is to observe a particular behavior (identifying some subset of radiology scans correctly) against some task (identifying hemorrhage on CT head scans correctly), and then to extrapolate based on that task alone.
The reality is that reducing any job, especially a wildly complex job that requires a decade of training, to a handful of tasks is quite absurd. Here's a bunch of stuff you wouldn't know about radiologists unless you built an AI company with them instead of opining about their job disappearing from an ivory tower.
Radiologists are not performing 2d pattern recognition - they have a 3d world model of the brain and its physical dynamics in their head. The motion and behavior of their brain to various traumas informs their prediction of hemorrhage determination.
Radiologists have a whole host of grounded models to make determinations, and actually, one of the most important first order determination they make is whether there is anything notably wrong with a brain structure that "feels" off. As a result, classifiers aren’t actually performing the same task even as radiologists.
Radiologists, because they have a grounded brain model, only need to see a single example of a rare and obscure condition to both remember it and identify it in the future. This long tail of rare conditions to avoid missing is a large part of their training, and no one has any clue how to make a model that acts similar in this way.
There’s so many ways to make Radiologist lives easier instead of just replacing them, it doesn’t even make sense to try. I interviewed and hired 25 radiologists, whose primary and chief complaint was that they had to reboot their computers several times a day..
A large part of the radiologist job is communicating their findings with physicians, so if you are thinking about automating them away you also need to understand the complex interactions between them and different clinics, which often are unique.
Every hospital is a snowflake, data is held under lock and key, so your algorithm might not work in a bunch of hospitals. Worse, the imagenet datasets have such wildly different feature sets they don’t do much for pretraining for you.
Have you ever tried to make anything in healthcare? The entire system is optimized to avoid introducing any harm to patients - explaining the ramifications of that would take an entire book, but suffice to say even if you had an algorithm that could automate away radiologists I don’t even know if you could create a viable adoption strategy in the US regulatory environment.
The reality is that for every application, the amount of specific and unknowlable domain knowledge is immense.
Long story short
Thinkers have a pattern where they are so divorced from implementation details that applications seem trivial, when in reality, the small details are exactly where value accrues.
Should you be worried about GPT5 being used to automate vulnerability detection on websites before they’re patched? Maybe.
Should you be worried GPT5 is going to interact with social systems and destroy our society single-handedly? No absolutely not.
So on appoint! Even the god father of AI (Geoffrey Hinton) is wrong about this. Yann LeCun, who is his student way back also isn’t afraid to point that out as well.
Exactly, there are a whole hosts of ground models that are not explicitly expressed in text or images that buried deep inside medial professionals brains. It’s a far stretched extrapolation from some image classification model to it will demolish the radiologist profession in 5 years. Eventually, yea, but we need much better learning techniques, and medical data collection and collaboration (somewhat even harder... due to PPI, regulations, and such)
That’s been said, there are some low hanging fruits that can improve efficiency and lives of the medical professionals in general not just radiologists. If you ever walked into any family doctors office or unfortunately enough to be hospitalized, you’d see doctors always carry a clipboard or sitting in front of computer typing away or they are followed by a flock of residency doctors with clipboards or tablets and they barely look at you, the patients. Because they have to keep records as much as they can for references, for patient records, for evidence purposes should anything happened. The existing LLMs with some enhancements (PSFT+VectorDB+Langchain,etc.) can create an easy to use app to help them record those data(audio, text, image, video) which can be used for further knowledge training, reduce their filling out form time, and protect them from malpractice or unreasonable complaints (which happens a lot in China...).