9 Comments

(1) Tech developers asking regulators to solve their Class 1 problems should set off alarm bells. In "The Talented Doctor Ripley (GPT): When Artificial Intelligence Lies about Medicine" (https://graboyes.substack.com/p/the-talented-doctor-ripley-gpt), I wrote about the dangers of AI insinuating incorrect information into medical journals and practice patterns, but also about the menace of injecting regulators into Class 1 problems--as in Sam Altman's plea to regulate ChatGPT. (2) My https://graboyes.substack.com/p/jambalaya-liztruss-vs-trussliz-and, includes two videos that highlight how daunting Class 2 problems can seem before social forces smooth things over. The videos--hilarious in retrospect--are aimed at senior citizens, terrified by the new technology of rotary telephones.

Expand full comment

I love this framing. This is a small nit to pick, but I'm curious about your opinion. It's my understanding that robotic surgery has become ubiquitous in performing prostatectomies. First, would you agree that's correct? Second, in that single use case, is it potentially in class 2?

Expand full comment

Great article, thank you. I'm fascinated by what those guys are doing with the phonograph in the photograph. It looks like a Far Side comic. If I could just think of a good caption for it...

Expand full comment

The problem I see after reading oyur excellent analysis is this, with the advent of Corporate Medicine run by CPA's, the Class 1 problems are selected by them with no feedback. Then when we have a class 2 solution, they ho no intention of listening, since they "run the show", and they regard providers as replaceable widgets.

Matthew Corey PA-C

Expand full comment