Automation Complacency, The Stepladder of AI in EHR’s, “Writing a note is how I think”. WHAT NOW?

A navel-gazing reflection on GPT, human cognitive effort, and the stepladder to the future. Where do YOU stand?

The image above generated by DALL-E embedded in the new BING, with the prompt “Doctors using a computer to treat patients, optimistic futuristic impressionistic image”. Wow. Not sure what the VR doctor coming out of the screen is doing.

Thanks to Dr. Brian Montague for prompting this post with his quote during a recent Large PIG meeting:

I find that I do a lot of my thinking when I write my progress note. If/when ChatGPT starts to write my note, when will I do that thinking?  — Brian Montague MD

That stopped me in my tracks.

We are so hell-bent on simplifying our work, reducing our EHR burden, we sometimes forget that this work is MORE than just pointing, clicking and typing.

It is also about THINKING. It is about assembling the data, carefully coaxing information and patterns out of our patients through skillful interview, parsimonious lab testing, and careful physical examination. It is how we, as physicians and APP’s, use our bodies and minds to craft an image of the syndrome, the disease: our hidden opponent.

Just like inserting a PC into the exam room changed dynamics, inserting GPT assistants into the EHR causes us to rethink … everything.

Pause to reflect

First, I think we should recall the technology adoption curve.

I fully acknowledge that I am currently dancing on the VERY PEAK of the peak of over-inflated expectations. Yes. That’s me right at the top.

Of concern, viewing the announcements this week from Google, Microsoft, and many others gives me chills (sometimes good, sometimes not) of what is coming: automated, deep-fake videos? Deep-fake images? Patients able to use GPT to write “more convincing” requests for … benzodiazepines? opiates? other controlled meds?

AND YET, think of the great things coming: GPT writing a first draft of the unending Patient Advice Requests coming to doctors. GPT writing a discharge summary based on events in a hospital stay. GPT gathering data relating to a particular disease process out of the terabytes of available data.

And where do we think physician/APP thinking might be impacted by excessive automation?

Automation Complacency

I refer you back to my book review of the book “The Glass Cage” by Nicholas Carr. As I said before, although this was written to critique the aircraft industry, I took it very personally as an attack on my whole career. I encourage you to read it.

In particular, I found the term “automation complacency” a fascinating and terrifying concept: that a user, who benefits from automation, will start to attribute MORE SKILL to the automation tool than it actually possesses, a COMPLACENCY that “don’t worry, I’m sure the automation will catch me if I make a mistake.”

We have already seen this among our clinicians, one of whom complained: “Why didn’t you warn me about the interaction between birth control pills and muscle relaxants? I expected the system to warn me of all relevant interactions. My patient had an adverse reaction because you did not warn me.”

Now, we have this problem. We have for years been turning off and reducing the number of interaction alerts we show to prescribers precisely because of alert fatigue. And now, we have complaints that “I want what I want when I want it. And you don’t have it right.” Seems like an impossible task. It IS an impossible task.

Thank you to all my fellow informaticists out there trying to make it right.

GPT and automation: helping or making worse?

Inserting a Large Language Model like GPT, that understands NOTHING, but just seems really fluent and sounding like an expert, could be helpful, but could also lull us into worse “automation complacency.” Even though we are supposed to (for now) read everything the GPT engine drafts, and we take full ownership of the output, how long will that last? Even today, I admit, as do most docs, that I use Dragon speech recognition and don’t read the output as carefully as I might.

Debating the steps in clinician thinking

So, here is where Dr. Montague and I had a discussion. We both believe it is true that a thoughtful, effective physician/APP will, after interviewing the patient and examining them, sit with the (formerly paper) chart, inhale all the relevant data, assemble it in their head. In the old days, we would suffer paper cuts and inky fingertips in this process of flipping pages. Now we just get carpal tunnel and dry eyes from the clicking, scrolling, scanning and typing.

Then when we’ve hunted and gathered the data, we slowly, carefully write an H/P or SOAP note (ok, an APSO-formatted SOAP note). It will include the Subjective (including a timeline of events), Objective (including relevant exam, lab findings), Assessment (assembly of symptoms into syndromes or diseases) and Plan (next steps to take).

During this laborious note-writing, we often come up with new ideas, new linkages, new insights. It is THIS PIECE we worry most about. If GPT can automate many of these pieces, WHERE WILL THE THINKING GO!?! I do not trust that GPT is truly thinking. I worry that the physician will instead STOP THINKING.

Then THERE IS NO THINKING.

Is this a race-to-the-bottom, or a competition to see who can speed us up so much that we are no longer healers, just fast documenters, since we are so burned out?

Who will we be?

Radio vs TV vs Internet

My optimistic thought is this. Instead of GPT coming to take our jobs, I’m hopeful GPT becomes a useful assistant, sorting through the chaff, sorting and highlighting the useful information in a data-rich, information-poor chart.

Just like the radio industry feared that TV would put them out of business (they didn’t), and TV feared that the Internet would put them out of business (they didn’t), the same, I think, goes for physicians, established healthcare teams, and GPT-automation tools.

Lines will be drawn (with luck, WE will draw them), and our jobs will change substantially. Just like emergent (unpredictable) properties like “GPT hallucinations” have arisen, we must re-invent our work as unexpected curves arise while deploying our new assistants.

Another Bing-Dall-E image of physicians at a computer. In the future, a doctor will apparently have more legs than before.

A possible step-ladder

I think physician thinking really occurs at the assembly of the Assessment and Plan. And that the early days of GPT assistance will begin in the Subjective and Objective sections of the note. GPT could for example:

SIMPLE
  • Subjective: Assemble a patient’s full chart on-demand for a new physician/APP meeting a patient in clinic, or on admission to hospital, focusing on previous events in can find in the local EHR or across a Health information exchange network, into an easily digestible timeline. Include a progression of symptoms, past history, past medications.
  • Objective: Filter a patient’s chart data to assemble a disease-specific timeline and summary: “show me all medications, test results, symptoms related to chest infection in the past year”
  • Then leave the assessment and planning to physician/APP assembly and un-assisted writing. This would leave clinician thinking largely untouched.
MODERATE
  • Subjective and Objective: GPT could take the entire chart and propose major diseases and syndromes it detects by pattern matching and assemble a brief page summary with supporting evidence and timeline, with citations.
  • Assessment and Plan: Suggest a prioritized list of Problems, severity, current state of treatment, suggested next treatments, based on a patient’s previous treatments and experience, as well as national best practices and guidelines. Leave the details, treatment adjustments and counseling to physicians/APPs interacting with the patient. Like Google Bard, GPT may suggest ‘top 3 suggestions with citations from literature or citations from EHR aggregate data’ and have the physician choose.
DREAMY/SCARY
  • Subjective and Objective: GPT could take the Moderate tools, add detection and surveillance for emerging diseases not yet described (the next Covid? the next Ebola? new-drug-associated-myocarditis? tryptophan eosinophilia-myalgia syndrome, not seen since 1989?) for public health monitoring. Step into the scanner for full body photography, CT, MRI, PET, with a comprehensive assessment in 1 simple step.
  • Assessment and Plan: GPT diagnoses common and also rare diseases via memorizing 1000’s clinical pathways and best-practice algorithms. GPT initiates treatment plans, needing just physician/APP cosignature.
  • A/P: Empowered by Eliza – like tools for empathy, takes on counseling the patient, discovering what conversational techniques engender the most patient behavior change. Recent studies already indicate that GPT can be considered more empathetic than doctors responding to online medical queries.

CMIO’s take? First things first. While we can wring our hands about “training our replacements”, there is lots yet to do and discover about our newest assistants. Shall we go on, eyes open?

Chatbot perspective from an insider (Rodney Brooks and Wired.com)

What Will Transformers Transform?

Thanks Rodney for a thoughtful discussion of

  • The Hype Cycle (peak of overinflated expectations)
  • The caution needed as our tools grow in skill exponentially
  • The ongoing risk of hallucination and unexpected errors in chatbots
  • The “grounding” problem with AI and robots

I particularly love the following quote:

Roy Amara, who died on the last day of 2007, was the president of a Palo Alto based think tank, the Institute for the future, and is credited with saying what is now known as Amara’s Law:

We tend to overestimate the effect of a technology in the short run and underestimate the effect in the long run.

I’m feeling an upward acceleration of AI skill with GPT applications, which could open up substantial risks as well as benefits.

For example, in the EHR space, there is discussion that GPT could take a patient and physician’s recorded conversation and automatically write a progress note, qualitatively more accurate and much faster than current commercial tools. Further, it could potentially summarize weeks of progress notes on hospitalized patients and write the discharge summary, a document that, when well-written, can take a human many hours of work. Or even, receive patients’ incoming MyChart messages and clinical questions and “reply to the patient in the voice of their clinician” based on the decade of writing by that clinician in the EHR.

Sure, these seem great. How about the potential deluge of GPT agents writing notes, requests ON BEHALF of patients, or other authors, that could junk up our systems? If 183,000 incoming patient messages is a lot (current monthly patient message volume at UCHealth), what if GPT somehow enabled 10x that number?

How much discussion will be GPT talking to GPT on behalf of employers/ patients/ discussants? I understand science fiction editors now have a 10x increase in sci-fi story submissions, a SUBSTANTIAL FRACTION now being written by GPT based on prior stories?

Too, wired.com has written about this:

https://www.wired.com/story/chatgpt-pause-ai-experiments-open-letter/

“I worry that we are very much in a ‘move fast and break things’ phase,” says Holstein, adding that the pace might be too quick for regulators to meaningfully keep up. “I like to think that we, in 2023, collectively, know better than this.”

CMIO’s take? Take a breath, everyone. It’s going to be a bumpy ride.

I Saw the Face of God in a TSMC Factory (Wired)

Nanometer scale production, Taiwan / Ukraine geopolitics, Huawei and spying in 5G, remote surgery, democracy at risk. What more can you ask for from a Wired Longread?

https://www.wired.com/story/i-saw-the-face-of-god-in-a-tsmc-factory/

I see myself in this article. Hints of:

  • My Asian background and the hunger for achievement
  • The constant threat of conflict
  • The importance of being comfortable with ambiguity
  • The unease of global tensions and political disagreement
  • The cultural clash between Asian and American habits
  • Moore’s law playing out in our lifetime
  • Lithography as art and the pinnacle of tech
  • The necessity of trust to drive innovation and growth
  • The knowledge that TSMC tech powers the vast majority of devices and servers without which American healthcare’s Electronic Health Records cannot exist.

Yes, it is a long read. This is what deeply researched, wide-ranging, thoughtful writing is about.

I’m a Taiwanese native American citizen. Taiwan Semiconductor Manufacturing Corporation (TSMC) probably isn’t a household name for most, and yet it produces the vast majority of the most advanced chips for the most advanced smartphones, laptops and computers. TSMC makes me proud to have been born on that tiny island.

CMIO’s take? I enjoyed this very much. This is a brilliant read. You may not agree with all of it, but it is a fascinating journey into the interconnectedness of our personal relationships, our technologies, our trust, our nations and leaders and nothing less than the future of our world.

AI will make human art more valuable (Wired.com)

What is it about human made art that makes us prefer that to AI art?

https://www.wired.com/story/art-artificial-intelligence-history/

Partly reassuring, partly cognitive puzzle. It turns out, if shown the SAME images labelled differently “made by human” “made by robot”, we will prefer the former. What does that say about us?

Infinite AI Interns for Everybody (Wired.com)

https://www.wired.com/story/artificial-intelligence-labor/

Here is our next AI challenge, as our jobs undergo gradual transformation. How will we as knowledge workers in informatics accommodate the growing sophistication of narrow AI assistants? Scheduling appointments, helping with spelling and grammar, now writing fluid manuscripts based on the library we point them to?

Teachers are starting to change “take home” assignments to in-class writing with no internet connection, to ensure students don’t outsource their thinking/writing to an AI chatbot.

What will we do? Can an AI replace us in writing the “one-pager” that summarizes thinking and succinctly and convincingly makes the case for change? Is Machiavelli vulnerable to being toppled? Do we no longer have need for governance and leadership if we can outsource thinking and fluency to an AI? Will my AI go up against your AI in the battle for mindshare?

Yes.

CMIO’s take? Yet again, time to re-imagine our jobs with the tools we have at hand.

META introduces AI-generated video

Another take on ‘AI is coming for you and your job’

https://www.technologyreview.com/2022/09/29/1060472/meta-text-to-video-ai/

Well, you thought AI-generated static images were both cool and concerning. Now, META (formerly Facebook) introduces AI-generated VIDEO based on simple user inputs. If distinguishing fake from real was hard before, where is this going?

DALL-E Mini Is the Internet’s Favorite AI Meme Machine (Wired.com)

https://www.wired.com/story/dalle-ai-meme-machine/

How can this be real? Read the story above at Wired.com.

I typed “Elephants breakdancing at midnight” into the prompt, and seriously, about a minute later I get this on my screen.

Let’s not go into why that sentence came out of my head thru my fingers, and instead focus on the technology. There is an AI, with the internet as infinite visual resource, that can now take brief text prompts and then render them for your viewing pleasure.

This is mind-blowing. Here’s “Frolicking Flying Cars”

Here’s “A family of dolphins using iPhones in the style of Picasso”

Here’s “Speed skating in the style of a Chinese landscape painting”

 

What is real? What is imaginary? Who drew this? Try it yourself at http://www.craiyon.com !

A.I. Is Mastering Language. Should We Trust What It Says? (nytimes)

GPT-3 can write text that astounds, based on its Large Language Model. Things are happening, people. Are we paying attention? #hcldr #hitsm #whyinformatics

I’m always on the looking for developments in computing outside of healthcare. This is a longer read, but so thought provoking:

  • What is a Large Language Model and why is it only recently important?
  • What is GPT-3 and what are all these magical things it supposedly does?
  • Can GPT-3 digest 1000 progress notes of a patient chart, say, and write a cogent 1-page summary for a clinician to digest rapidly? I’d pay for THAT.

‘‘The underlying idea of GPT-3 is a way of linking an intuitive notion of understanding to something that can be measured and understood mechanistically,’’ he finally said, ‘‘and that is the task of predicting the next word in text.’’ 

Prompt the algorithm with a sentence like ‘‘The writer has omitted the very last word of the first . . . ’’ and the guesses will be a kind of stream of nonsense: ‘‘satellite,’’ ‘‘puppy,’’ ‘‘Seattle,’’ ‘‘therefore.’’ But somewhere down the list — perhaps thousands of words down the list — the correct missing word appears: ‘‘paragraph.’’ The software then strengthens whatever random neural connections generated that particular suggestion and weakens all the connections that generated incorrect guesses. And then it moves on to the next prompt. Over time, with enough iterations, the software learns.

Ilya Sutskever

There is all this discussion of “is this a sophisticated parrot” or “truly an artificial intelligence capable of generating new ideas.” Well, in our Electronic Health Record world, just the first item would be transformative, if we can get an AI to digest a hyperobject large set of data into an executive brief. Just that.

CMIO’s take? This is an important article by Steven Johnson in the New York Times Magazine. Watch this space; the development of GPT-3 heralds a qualitative improvement in AI language models; so much so that we feel compelled to teach it values and culture lest it start spewing hatred it learns on the internet. This is a worthwhile long read.

%d